diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,53934 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 14.25925925925926, + "eval_steps": 500, + "global_step": 7700, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.001851851851851852, + "grad_norm": 16.01453742088244, + "learning_rate": 3.0864197530864196e-10, + "loss": 1.8896, + "step": 1 + }, + { + "epoch": 0.003703703703703704, + "grad_norm": 15.375592963986762, + "learning_rate": 6.172839506172839e-10, + "loss": 1.9114, + "step": 2 + }, + { + "epoch": 0.005555555555555556, + "grad_norm": 15.405659955056148, + "learning_rate": 9.259259259259258e-10, + "loss": 1.8899, + "step": 3 + }, + { + "epoch": 0.007407407407407408, + "grad_norm": 15.613867338083537, + "learning_rate": 1.2345679012345679e-09, + "loss": 1.8778, + "step": 4 + }, + { + "epoch": 0.009259259259259259, + "grad_norm": 16.24258898972724, + "learning_rate": 1.5432098765432097e-09, + "loss": 1.8796, + "step": 5 + }, + { + "epoch": 0.011111111111111112, + "grad_norm": 15.967516821922272, + "learning_rate": 1.8518518518518516e-09, + "loss": 1.8563, + "step": 6 + }, + { + "epoch": 0.012962962962962963, + "grad_norm": 16.330852281159242, + "learning_rate": 2.1604938271604934e-09, + "loss": 1.862, + "step": 7 + }, + { + "epoch": 0.014814814814814815, + "grad_norm": 15.376267659430235, + "learning_rate": 2.4691358024691357e-09, + "loss": 1.8683, + "step": 8 + }, + { + "epoch": 0.016666666666666666, + "grad_norm": 15.328960258881947, + "learning_rate": 2.7777777777777776e-09, + "loss": 1.8845, + "step": 9 + }, + { + "epoch": 0.018518518518518517, + "grad_norm": 15.39795148685497, + "learning_rate": 3.0864197530864194e-09, + "loss": 1.8821, + "step": 10 + }, + { + "epoch": 0.020370370370370372, + "grad_norm": 15.89711552333428, + "learning_rate": 3.3950617283950613e-09, + "loss": 1.8833, + "step": 11 + }, + { + "epoch": 0.022222222222222223, + "grad_norm": 15.436268793133458, + "learning_rate": 3.703703703703703e-09, + "loss": 1.886, + "step": 12 + }, + { + "epoch": 0.024074074074074074, + "grad_norm": 15.58921411331668, + "learning_rate": 4.012345679012345e-09, + "loss": 1.8843, + "step": 13 + }, + { + "epoch": 0.025925925925925925, + "grad_norm": 16.350936505218403, + "learning_rate": 4.320987654320987e-09, + "loss": 1.8674, + "step": 14 + }, + { + "epoch": 0.027777777777777776, + "grad_norm": 15.59724033876366, + "learning_rate": 4.6296296296296295e-09, + "loss": 1.8857, + "step": 15 + }, + { + "epoch": 0.02962962962962963, + "grad_norm": 15.806538261793046, + "learning_rate": 4.938271604938271e-09, + "loss": 1.8755, + "step": 16 + }, + { + "epoch": 0.03148148148148148, + "grad_norm": 15.902812899049493, + "learning_rate": 5.246913580246913e-09, + "loss": 1.8819, + "step": 17 + }, + { + "epoch": 0.03333333333333333, + "grad_norm": 16.448802476502646, + "learning_rate": 5.555555555555555e-09, + "loss": 1.8983, + "step": 18 + }, + { + "epoch": 0.03518518518518519, + "grad_norm": 15.990295108515035, + "learning_rate": 5.864197530864197e-09, + "loss": 1.8844, + "step": 19 + }, + { + "epoch": 0.037037037037037035, + "grad_norm": 15.174397301064433, + "learning_rate": 6.172839506172839e-09, + "loss": 1.916, + "step": 20 + }, + { + "epoch": 0.03888888888888889, + "grad_norm": 15.769790859011602, + "learning_rate": 6.481481481481481e-09, + "loss": 1.8974, + "step": 21 + }, + { + "epoch": 0.040740740740740744, + "grad_norm": 15.612263076501023, + "learning_rate": 6.7901234567901226e-09, + "loss": 1.8726, + "step": 22 + }, + { + "epoch": 0.04259259259259259, + "grad_norm": 15.051733205144405, + "learning_rate": 7.0987654320987644e-09, + "loss": 1.8778, + "step": 23 + }, + { + "epoch": 0.044444444444444446, + "grad_norm": 15.635262076445171, + "learning_rate": 7.407407407407406e-09, + "loss": 1.8593, + "step": 24 + }, + { + "epoch": 0.046296296296296294, + "grad_norm": 15.497782004107918, + "learning_rate": 7.716049382716048e-09, + "loss": 1.8676, + "step": 25 + }, + { + "epoch": 0.04814814814814815, + "grad_norm": 16.193103961933446, + "learning_rate": 8.02469135802469e-09, + "loss": 1.8722, + "step": 26 + }, + { + "epoch": 0.05, + "grad_norm": 15.309127038724816, + "learning_rate": 8.333333333333332e-09, + "loss": 1.9183, + "step": 27 + }, + { + "epoch": 0.05185185185185185, + "grad_norm": 15.65547088924166, + "learning_rate": 8.641975308641974e-09, + "loss": 1.8873, + "step": 28 + }, + { + "epoch": 0.053703703703703705, + "grad_norm": 16.00856397449651, + "learning_rate": 8.950617283950616e-09, + "loss": 1.8726, + "step": 29 + }, + { + "epoch": 0.05555555555555555, + "grad_norm": 15.60440723586894, + "learning_rate": 9.259259259259259e-09, + "loss": 1.8856, + "step": 30 + }, + { + "epoch": 0.05740740740740741, + "grad_norm": 15.371658124387729, + "learning_rate": 9.567901234567901e-09, + "loss": 1.8822, + "step": 31 + }, + { + "epoch": 0.05925925925925926, + "grad_norm": 16.084314624459886, + "learning_rate": 9.876543209876543e-09, + "loss": 1.8913, + "step": 32 + }, + { + "epoch": 0.06111111111111111, + "grad_norm": 15.580766097420982, + "learning_rate": 1.0185185185185185e-08, + "loss": 1.856, + "step": 33 + }, + { + "epoch": 0.06296296296296296, + "grad_norm": 16.17880487954311, + "learning_rate": 1.0493827160493827e-08, + "loss": 1.888, + "step": 34 + }, + { + "epoch": 0.06481481481481481, + "grad_norm": 15.936455325209984, + "learning_rate": 1.0802469135802468e-08, + "loss": 1.8952, + "step": 35 + }, + { + "epoch": 0.06666666666666667, + "grad_norm": 15.298388504567042, + "learning_rate": 1.111111111111111e-08, + "loss": 1.8959, + "step": 36 + }, + { + "epoch": 0.06851851851851852, + "grad_norm": 15.144670267344019, + "learning_rate": 1.1419753086419752e-08, + "loss": 1.8926, + "step": 37 + }, + { + "epoch": 0.07037037037037037, + "grad_norm": 15.960938559417837, + "learning_rate": 1.1728395061728394e-08, + "loss": 1.8889, + "step": 38 + }, + { + "epoch": 0.07222222222222222, + "grad_norm": 15.777294203296226, + "learning_rate": 1.2037037037037036e-08, + "loss": 1.8995, + "step": 39 + }, + { + "epoch": 0.07407407407407407, + "grad_norm": 15.497973406924203, + "learning_rate": 1.2345679012345678e-08, + "loss": 1.8804, + "step": 40 + }, + { + "epoch": 0.07592592592592592, + "grad_norm": 16.141837983841427, + "learning_rate": 1.2654320987654321e-08, + "loss": 1.8894, + "step": 41 + }, + { + "epoch": 0.07777777777777778, + "grad_norm": 14.684290707896258, + "learning_rate": 1.2962962962962961e-08, + "loss": 1.8917, + "step": 42 + }, + { + "epoch": 0.07962962962962963, + "grad_norm": 15.634733612368853, + "learning_rate": 1.3271604938271605e-08, + "loss": 1.9091, + "step": 43 + }, + { + "epoch": 0.08148148148148149, + "grad_norm": 15.961623048553635, + "learning_rate": 1.3580246913580245e-08, + "loss": 1.8841, + "step": 44 + }, + { + "epoch": 0.08333333333333333, + "grad_norm": 15.7446567956114, + "learning_rate": 1.3888888888888889e-08, + "loss": 1.8975, + "step": 45 + }, + { + "epoch": 0.08518518518518518, + "grad_norm": 15.949740281110246, + "learning_rate": 1.4197530864197529e-08, + "loss": 1.892, + "step": 46 + }, + { + "epoch": 0.08703703703703704, + "grad_norm": 15.43925634774645, + "learning_rate": 1.4506172839506172e-08, + "loss": 1.8864, + "step": 47 + }, + { + "epoch": 0.08888888888888889, + "grad_norm": 15.396877731257465, + "learning_rate": 1.4814814814814813e-08, + "loss": 1.8773, + "step": 48 + }, + { + "epoch": 0.09074074074074075, + "grad_norm": 15.230593923513473, + "learning_rate": 1.5123456790123456e-08, + "loss": 1.9074, + "step": 49 + }, + { + "epoch": 0.09259259259259259, + "grad_norm": 15.758712834103486, + "learning_rate": 1.5432098765432096e-08, + "loss": 1.8853, + "step": 50 + }, + { + "epoch": 0.09444444444444444, + "grad_norm": 15.207502212719666, + "learning_rate": 1.574074074074074e-08, + "loss": 1.8784, + "step": 51 + }, + { + "epoch": 0.0962962962962963, + "grad_norm": 15.123948116692935, + "learning_rate": 1.604938271604938e-08, + "loss": 1.8772, + "step": 52 + }, + { + "epoch": 0.09814814814814815, + "grad_norm": 15.278280361005264, + "learning_rate": 1.6358024691358024e-08, + "loss": 1.8965, + "step": 53 + }, + { + "epoch": 0.1, + "grad_norm": 15.839011867588935, + "learning_rate": 1.6666666666666664e-08, + "loss": 1.8967, + "step": 54 + }, + { + "epoch": 0.10185185185185185, + "grad_norm": 16.65603378069207, + "learning_rate": 1.6975308641975307e-08, + "loss": 1.8939, + "step": 55 + }, + { + "epoch": 0.1037037037037037, + "grad_norm": 15.816691700766857, + "learning_rate": 1.7283950617283947e-08, + "loss": 1.8847, + "step": 56 + }, + { + "epoch": 0.10555555555555556, + "grad_norm": 15.798316407389207, + "learning_rate": 1.759259259259259e-08, + "loss": 1.887, + "step": 57 + }, + { + "epoch": 0.10740740740740741, + "grad_norm": 15.47639797900752, + "learning_rate": 1.790123456790123e-08, + "loss": 1.8748, + "step": 58 + }, + { + "epoch": 0.10925925925925926, + "grad_norm": 15.27598829120491, + "learning_rate": 1.8209876543209875e-08, + "loss": 1.8878, + "step": 59 + }, + { + "epoch": 0.1111111111111111, + "grad_norm": 15.563651931865994, + "learning_rate": 1.8518518518518518e-08, + "loss": 1.8734, + "step": 60 + }, + { + "epoch": 0.11296296296296296, + "grad_norm": 15.480857981942341, + "learning_rate": 1.882716049382716e-08, + "loss": 1.8833, + "step": 61 + }, + { + "epoch": 0.11481481481481481, + "grad_norm": 15.464224881293541, + "learning_rate": 1.9135802469135802e-08, + "loss": 1.8972, + "step": 62 + }, + { + "epoch": 0.11666666666666667, + "grad_norm": 15.90152313570589, + "learning_rate": 1.9444444444444445e-08, + "loss": 1.87, + "step": 63 + }, + { + "epoch": 0.11851851851851852, + "grad_norm": 14.983814185373543, + "learning_rate": 1.9753086419753086e-08, + "loss": 1.8687, + "step": 64 + }, + { + "epoch": 0.12037037037037036, + "grad_norm": 15.356093556702819, + "learning_rate": 2.006172839506173e-08, + "loss": 1.8886, + "step": 65 + }, + { + "epoch": 0.12222222222222222, + "grad_norm": 15.982885648456714, + "learning_rate": 2.037037037037037e-08, + "loss": 1.8817, + "step": 66 + }, + { + "epoch": 0.12407407407407407, + "grad_norm": 15.130325195138646, + "learning_rate": 2.0679012345679013e-08, + "loss": 1.8789, + "step": 67 + }, + { + "epoch": 0.1259259259259259, + "grad_norm": 15.614819205673005, + "learning_rate": 2.0987654320987653e-08, + "loss": 1.8949, + "step": 68 + }, + { + "epoch": 0.12777777777777777, + "grad_norm": 15.860795548772547, + "learning_rate": 2.1296296296296297e-08, + "loss": 1.9031, + "step": 69 + }, + { + "epoch": 0.12962962962962962, + "grad_norm": 15.463036496331627, + "learning_rate": 2.1604938271604937e-08, + "loss": 1.9112, + "step": 70 + }, + { + "epoch": 0.13148148148148148, + "grad_norm": 16.012861099884283, + "learning_rate": 2.191358024691358e-08, + "loss": 1.8721, + "step": 71 + }, + { + "epoch": 0.13333333333333333, + "grad_norm": 14.842117485417482, + "learning_rate": 2.222222222222222e-08, + "loss": 1.8864, + "step": 72 + }, + { + "epoch": 0.13518518518518519, + "grad_norm": 15.066224179737386, + "learning_rate": 2.2530864197530864e-08, + "loss": 1.8667, + "step": 73 + }, + { + "epoch": 0.13703703703703704, + "grad_norm": 15.373961458497948, + "learning_rate": 2.2839506172839504e-08, + "loss": 1.8816, + "step": 74 + }, + { + "epoch": 0.1388888888888889, + "grad_norm": 15.784324151509434, + "learning_rate": 2.3148148148148148e-08, + "loss": 1.8444, + "step": 75 + }, + { + "epoch": 0.14074074074074075, + "grad_norm": 15.196481277715948, + "learning_rate": 2.3456790123456788e-08, + "loss": 1.8624, + "step": 76 + }, + { + "epoch": 0.1425925925925926, + "grad_norm": 15.761145080537604, + "learning_rate": 2.376543209876543e-08, + "loss": 1.8882, + "step": 77 + }, + { + "epoch": 0.14444444444444443, + "grad_norm": 15.184276001289748, + "learning_rate": 2.407407407407407e-08, + "loss": 1.8815, + "step": 78 + }, + { + "epoch": 0.14629629629629629, + "grad_norm": 15.867557039261923, + "learning_rate": 2.4382716049382715e-08, + "loss": 1.8617, + "step": 79 + }, + { + "epoch": 0.14814814814814814, + "grad_norm": 15.528652016760974, + "learning_rate": 2.4691358024691355e-08, + "loss": 1.8606, + "step": 80 + }, + { + "epoch": 0.15, + "grad_norm": 16.54859029739373, + "learning_rate": 2.5e-08, + "loss": 1.8925, + "step": 81 + }, + { + "epoch": 0.15185185185185185, + "grad_norm": 14.920939542797587, + "learning_rate": 2.5308641975308642e-08, + "loss": 1.8635, + "step": 82 + }, + { + "epoch": 0.1537037037037037, + "grad_norm": 15.386169707489259, + "learning_rate": 2.5617283950617286e-08, + "loss": 1.8643, + "step": 83 + }, + { + "epoch": 0.15555555555555556, + "grad_norm": 14.923801795083323, + "learning_rate": 2.5925925925925923e-08, + "loss": 1.8584, + "step": 84 + }, + { + "epoch": 0.1574074074074074, + "grad_norm": 16.573869792629477, + "learning_rate": 2.6234567901234566e-08, + "loss": 1.8974, + "step": 85 + }, + { + "epoch": 0.15925925925925927, + "grad_norm": 15.467925489897883, + "learning_rate": 2.654320987654321e-08, + "loss": 1.8619, + "step": 86 + }, + { + "epoch": 0.16111111111111112, + "grad_norm": 15.300386666285116, + "learning_rate": 2.6851851851851853e-08, + "loss": 1.8603, + "step": 87 + }, + { + "epoch": 0.16296296296296298, + "grad_norm": 15.605978677185178, + "learning_rate": 2.716049382716049e-08, + "loss": 1.8666, + "step": 88 + }, + { + "epoch": 0.1648148148148148, + "grad_norm": 15.459323029806637, + "learning_rate": 2.7469135802469134e-08, + "loss": 1.877, + "step": 89 + }, + { + "epoch": 0.16666666666666666, + "grad_norm": 15.37123605504257, + "learning_rate": 2.7777777777777777e-08, + "loss": 1.8859, + "step": 90 + }, + { + "epoch": 0.1685185185185185, + "grad_norm": 15.062010266255461, + "learning_rate": 2.808641975308642e-08, + "loss": 1.8555, + "step": 91 + }, + { + "epoch": 0.17037037037037037, + "grad_norm": 15.243190194498046, + "learning_rate": 2.8395061728395058e-08, + "loss": 1.8615, + "step": 92 + }, + { + "epoch": 0.17222222222222222, + "grad_norm": 15.299465623497353, + "learning_rate": 2.87037037037037e-08, + "loss": 1.8746, + "step": 93 + }, + { + "epoch": 0.17407407407407408, + "grad_norm": 16.017789825984863, + "learning_rate": 2.9012345679012345e-08, + "loss": 1.8422, + "step": 94 + }, + { + "epoch": 0.17592592592592593, + "grad_norm": 15.316176955655463, + "learning_rate": 2.9320987654320988e-08, + "loss": 1.8897, + "step": 95 + }, + { + "epoch": 0.17777777777777778, + "grad_norm": 15.120821230346957, + "learning_rate": 2.9629629629629625e-08, + "loss": 1.8606, + "step": 96 + }, + { + "epoch": 0.17962962962962964, + "grad_norm": 15.038634971379201, + "learning_rate": 2.993827160493827e-08, + "loss": 1.8772, + "step": 97 + }, + { + "epoch": 0.1814814814814815, + "grad_norm": 15.6967389857466, + "learning_rate": 3.024691358024691e-08, + "loss": 1.8485, + "step": 98 + }, + { + "epoch": 0.18333333333333332, + "grad_norm": 15.451950300728415, + "learning_rate": 3.0555555555555556e-08, + "loss": 1.8665, + "step": 99 + }, + { + "epoch": 0.18518518518518517, + "grad_norm": 15.212053142047973, + "learning_rate": 3.086419753086419e-08, + "loss": 1.8867, + "step": 100 + }, + { + "epoch": 0.18703703703703703, + "grad_norm": 15.702086924496298, + "learning_rate": 3.1172839506172836e-08, + "loss": 1.8727, + "step": 101 + }, + { + "epoch": 0.18888888888888888, + "grad_norm": 15.96363973798426, + "learning_rate": 3.148148148148148e-08, + "loss": 1.8869, + "step": 102 + }, + { + "epoch": 0.19074074074074074, + "grad_norm": 15.144738909008389, + "learning_rate": 3.179012345679012e-08, + "loss": 1.8564, + "step": 103 + }, + { + "epoch": 0.1925925925925926, + "grad_norm": 15.327783569924641, + "learning_rate": 3.209876543209876e-08, + "loss": 1.8837, + "step": 104 + }, + { + "epoch": 0.19444444444444445, + "grad_norm": 15.015824132699844, + "learning_rate": 3.2407407407407403e-08, + "loss": 1.8836, + "step": 105 + }, + { + "epoch": 0.1962962962962963, + "grad_norm": 16.079740454552617, + "learning_rate": 3.271604938271605e-08, + "loss": 1.8739, + "step": 106 + }, + { + "epoch": 0.19814814814814816, + "grad_norm": 15.469412631203738, + "learning_rate": 3.302469135802469e-08, + "loss": 1.8541, + "step": 107 + }, + { + "epoch": 0.2, + "grad_norm": 15.148696319429204, + "learning_rate": 3.333333333333333e-08, + "loss": 1.8431, + "step": 108 + }, + { + "epoch": 0.20185185185185187, + "grad_norm": 15.167549682839613, + "learning_rate": 3.364197530864197e-08, + "loss": 1.8696, + "step": 109 + }, + { + "epoch": 0.2037037037037037, + "grad_norm": 15.555184045371803, + "learning_rate": 3.3950617283950614e-08, + "loss": 1.861, + "step": 110 + }, + { + "epoch": 0.20555555555555555, + "grad_norm": 15.649091725683697, + "learning_rate": 3.425925925925926e-08, + "loss": 1.8514, + "step": 111 + }, + { + "epoch": 0.2074074074074074, + "grad_norm": 15.162869233515563, + "learning_rate": 3.4567901234567895e-08, + "loss": 1.8784, + "step": 112 + }, + { + "epoch": 0.20925925925925926, + "grad_norm": 15.73018581787736, + "learning_rate": 3.487654320987654e-08, + "loss": 1.8587, + "step": 113 + }, + { + "epoch": 0.2111111111111111, + "grad_norm": 14.870954626807174, + "learning_rate": 3.518518518518518e-08, + "loss": 1.8494, + "step": 114 + }, + { + "epoch": 0.21296296296296297, + "grad_norm": 15.143299331711805, + "learning_rate": 3.5493827160493825e-08, + "loss": 1.8211, + "step": 115 + }, + { + "epoch": 0.21481481481481482, + "grad_norm": 14.94353813265894, + "learning_rate": 3.580246913580246e-08, + "loss": 1.8483, + "step": 116 + }, + { + "epoch": 0.21666666666666667, + "grad_norm": 15.097702504420209, + "learning_rate": 3.6111111111111106e-08, + "loss": 1.836, + "step": 117 + }, + { + "epoch": 0.21851851851851853, + "grad_norm": 14.714407171183284, + "learning_rate": 3.641975308641975e-08, + "loss": 1.8255, + "step": 118 + }, + { + "epoch": 0.22037037037037038, + "grad_norm": 14.923597793071034, + "learning_rate": 3.672839506172839e-08, + "loss": 1.84, + "step": 119 + }, + { + "epoch": 0.2222222222222222, + "grad_norm": 15.348980889915634, + "learning_rate": 3.7037037037037036e-08, + "loss": 1.8453, + "step": 120 + }, + { + "epoch": 0.22407407407407406, + "grad_norm": 15.227747932391384, + "learning_rate": 3.734567901234568e-08, + "loss": 1.8052, + "step": 121 + }, + { + "epoch": 0.22592592592592592, + "grad_norm": 14.746296661183148, + "learning_rate": 3.765432098765432e-08, + "loss": 1.8264, + "step": 122 + }, + { + "epoch": 0.22777777777777777, + "grad_norm": 16.131628947607084, + "learning_rate": 3.796296296296296e-08, + "loss": 1.8502, + "step": 123 + }, + { + "epoch": 0.22962962962962963, + "grad_norm": 15.319469480247056, + "learning_rate": 3.8271604938271604e-08, + "loss": 1.8288, + "step": 124 + }, + { + "epoch": 0.23148148148148148, + "grad_norm": 15.177552741329016, + "learning_rate": 3.858024691358025e-08, + "loss": 1.83, + "step": 125 + }, + { + "epoch": 0.23333333333333334, + "grad_norm": 14.83299348919801, + "learning_rate": 3.888888888888889e-08, + "loss": 1.8164, + "step": 126 + }, + { + "epoch": 0.2351851851851852, + "grad_norm": 14.75159373435331, + "learning_rate": 3.9197530864197534e-08, + "loss": 1.8466, + "step": 127 + }, + { + "epoch": 0.23703703703703705, + "grad_norm": 14.804098646037755, + "learning_rate": 3.950617283950617e-08, + "loss": 1.8236, + "step": 128 + }, + { + "epoch": 0.2388888888888889, + "grad_norm": 14.709991449008557, + "learning_rate": 3.9814814814814815e-08, + "loss": 1.8148, + "step": 129 + }, + { + "epoch": 0.24074074074074073, + "grad_norm": 14.853414636656769, + "learning_rate": 4.012345679012346e-08, + "loss": 1.831, + "step": 130 + }, + { + "epoch": 0.24259259259259258, + "grad_norm": 15.651472651182399, + "learning_rate": 4.04320987654321e-08, + "loss": 1.8288, + "step": 131 + }, + { + "epoch": 0.24444444444444444, + "grad_norm": 15.965201353715011, + "learning_rate": 4.074074074074074e-08, + "loss": 1.8194, + "step": 132 + }, + { + "epoch": 0.2462962962962963, + "grad_norm": 15.18102590434218, + "learning_rate": 4.104938271604938e-08, + "loss": 1.8254, + "step": 133 + }, + { + "epoch": 0.24814814814814815, + "grad_norm": 14.932468681124538, + "learning_rate": 4.1358024691358026e-08, + "loss": 1.8061, + "step": 134 + }, + { + "epoch": 0.25, + "grad_norm": 15.397229031460373, + "learning_rate": 4.166666666666667e-08, + "loss": 1.8223, + "step": 135 + }, + { + "epoch": 0.2518518518518518, + "grad_norm": 14.98960680651719, + "learning_rate": 4.1975308641975306e-08, + "loss": 1.8094, + "step": 136 + }, + { + "epoch": 0.2537037037037037, + "grad_norm": 14.814808212375743, + "learning_rate": 4.228395061728395e-08, + "loss": 1.8249, + "step": 137 + }, + { + "epoch": 0.25555555555555554, + "grad_norm": 15.486470872876474, + "learning_rate": 4.259259259259259e-08, + "loss": 1.8442, + "step": 138 + }, + { + "epoch": 0.2574074074074074, + "grad_norm": 15.689354082865808, + "learning_rate": 4.2901234567901237e-08, + "loss": 1.8146, + "step": 139 + }, + { + "epoch": 0.25925925925925924, + "grad_norm": 14.83414161988492, + "learning_rate": 4.3209876543209874e-08, + "loss": 1.8466, + "step": 140 + }, + { + "epoch": 0.2611111111111111, + "grad_norm": 17.77942177387046, + "learning_rate": 4.351851851851852e-08, + "loss": 1.8304, + "step": 141 + }, + { + "epoch": 0.26296296296296295, + "grad_norm": 14.836118142473255, + "learning_rate": 4.382716049382716e-08, + "loss": 1.8189, + "step": 142 + }, + { + "epoch": 0.26481481481481484, + "grad_norm": 15.30036179717283, + "learning_rate": 4.4135802469135804e-08, + "loss": 1.8343, + "step": 143 + }, + { + "epoch": 0.26666666666666666, + "grad_norm": 14.738970792481036, + "learning_rate": 4.444444444444444e-08, + "loss": 1.8143, + "step": 144 + }, + { + "epoch": 0.26851851851851855, + "grad_norm": 14.542012053556816, + "learning_rate": 4.4753086419753084e-08, + "loss": 1.8194, + "step": 145 + }, + { + "epoch": 0.27037037037037037, + "grad_norm": 14.484645134463825, + "learning_rate": 4.506172839506173e-08, + "loss": 1.8177, + "step": 146 + }, + { + "epoch": 0.2722222222222222, + "grad_norm": 15.685240605752165, + "learning_rate": 4.537037037037037e-08, + "loss": 1.8179, + "step": 147 + }, + { + "epoch": 0.2740740740740741, + "grad_norm": 14.990947152614467, + "learning_rate": 4.567901234567901e-08, + "loss": 1.8319, + "step": 148 + }, + { + "epoch": 0.2759259259259259, + "grad_norm": 15.451965202760704, + "learning_rate": 4.598765432098765e-08, + "loss": 1.8357, + "step": 149 + }, + { + "epoch": 0.2777777777777778, + "grad_norm": 15.204525042476222, + "learning_rate": 4.6296296296296295e-08, + "loss": 1.8303, + "step": 150 + }, + { + "epoch": 0.2796296296296296, + "grad_norm": 15.104514218766667, + "learning_rate": 4.660493827160494e-08, + "loss": 1.7974, + "step": 151 + }, + { + "epoch": 0.2814814814814815, + "grad_norm": 15.105548365392572, + "learning_rate": 4.6913580246913576e-08, + "loss": 1.8249, + "step": 152 + }, + { + "epoch": 0.2833333333333333, + "grad_norm": 15.747532055963093, + "learning_rate": 4.722222222222222e-08, + "loss": 1.8008, + "step": 153 + }, + { + "epoch": 0.2851851851851852, + "grad_norm": 15.353991895527601, + "learning_rate": 4.753086419753086e-08, + "loss": 1.8251, + "step": 154 + }, + { + "epoch": 0.28703703703703703, + "grad_norm": 15.209788532634054, + "learning_rate": 4.7839506172839506e-08, + "loss": 1.809, + "step": 155 + }, + { + "epoch": 0.28888888888888886, + "grad_norm": 15.80173994023578, + "learning_rate": 4.814814814814814e-08, + "loss": 1.8013, + "step": 156 + }, + { + "epoch": 0.29074074074074074, + "grad_norm": 16.048047139021342, + "learning_rate": 4.845679012345679e-08, + "loss": 1.7916, + "step": 157 + }, + { + "epoch": 0.29259259259259257, + "grad_norm": 15.962081332069088, + "learning_rate": 4.876543209876543e-08, + "loss": 1.7817, + "step": 158 + }, + { + "epoch": 0.29444444444444445, + "grad_norm": 16.466556952911027, + "learning_rate": 4.9074074074074074e-08, + "loss": 1.7185, + "step": 159 + }, + { + "epoch": 0.2962962962962963, + "grad_norm": 16.223198762951395, + "learning_rate": 4.938271604938271e-08, + "loss": 1.7481, + "step": 160 + }, + { + "epoch": 0.29814814814814816, + "grad_norm": 15.887450875336, + "learning_rate": 4.9691358024691354e-08, + "loss": 1.7135, + "step": 161 + }, + { + "epoch": 0.3, + "grad_norm": 15.365124448323291, + "learning_rate": 5e-08, + "loss": 1.7141, + "step": 162 + }, + { + "epoch": 0.30185185185185187, + "grad_norm": 16.476719186450662, + "learning_rate": 5.030864197530864e-08, + "loss": 1.712, + "step": 163 + }, + { + "epoch": 0.3037037037037037, + "grad_norm": 15.080000212695525, + "learning_rate": 5.0617283950617285e-08, + "loss": 1.7059, + "step": 164 + }, + { + "epoch": 0.3055555555555556, + "grad_norm": 16.143726335461462, + "learning_rate": 5.092592592592593e-08, + "loss": 1.6906, + "step": 165 + }, + { + "epoch": 0.3074074074074074, + "grad_norm": 14.6141471939126, + "learning_rate": 5.123456790123457e-08, + "loss": 1.7233, + "step": 166 + }, + { + "epoch": 0.30925925925925923, + "grad_norm": 15.057721002857242, + "learning_rate": 5.15432098765432e-08, + "loss": 1.6839, + "step": 167 + }, + { + "epoch": 0.3111111111111111, + "grad_norm": 14.889380937788225, + "learning_rate": 5.1851851851851846e-08, + "loss": 1.6885, + "step": 168 + }, + { + "epoch": 0.31296296296296294, + "grad_norm": 14.851576464414487, + "learning_rate": 5.216049382716049e-08, + "loss": 1.6857, + "step": 169 + }, + { + "epoch": 0.3148148148148148, + "grad_norm": 15.295223573712184, + "learning_rate": 5.246913580246913e-08, + "loss": 1.7006, + "step": 170 + }, + { + "epoch": 0.31666666666666665, + "grad_norm": 14.803728888120778, + "learning_rate": 5.2777777777777776e-08, + "loss": 1.6766, + "step": 171 + }, + { + "epoch": 0.31851851851851853, + "grad_norm": 15.086524537977429, + "learning_rate": 5.308641975308642e-08, + "loss": 1.6815, + "step": 172 + }, + { + "epoch": 0.32037037037037036, + "grad_norm": 15.802758931254427, + "learning_rate": 5.339506172839506e-08, + "loss": 1.6949, + "step": 173 + }, + { + "epoch": 0.32222222222222224, + "grad_norm": 15.247930974320342, + "learning_rate": 5.3703703703703707e-08, + "loss": 1.6797, + "step": 174 + }, + { + "epoch": 0.32407407407407407, + "grad_norm": 15.27233097586875, + "learning_rate": 5.401234567901234e-08, + "loss": 1.6789, + "step": 175 + }, + { + "epoch": 0.32592592592592595, + "grad_norm": 15.380931053883348, + "learning_rate": 5.432098765432098e-08, + "loss": 1.6681, + "step": 176 + }, + { + "epoch": 0.3277777777777778, + "grad_norm": 14.755118321502803, + "learning_rate": 5.4629629629629624e-08, + "loss": 1.6586, + "step": 177 + }, + { + "epoch": 0.3296296296296296, + "grad_norm": 15.535643296681924, + "learning_rate": 5.493827160493827e-08, + "loss": 1.6645, + "step": 178 + }, + { + "epoch": 0.3314814814814815, + "grad_norm": 15.16851840528597, + "learning_rate": 5.524691358024691e-08, + "loss": 1.6719, + "step": 179 + }, + { + "epoch": 0.3333333333333333, + "grad_norm": 15.071491861044388, + "learning_rate": 5.5555555555555555e-08, + "loss": 1.6874, + "step": 180 + }, + { + "epoch": 0.3351851851851852, + "grad_norm": 15.0838502603752, + "learning_rate": 5.58641975308642e-08, + "loss": 1.6684, + "step": 181 + }, + { + "epoch": 0.337037037037037, + "grad_norm": 15.035761965661147, + "learning_rate": 5.617283950617284e-08, + "loss": 1.6646, + "step": 182 + }, + { + "epoch": 0.3388888888888889, + "grad_norm": 15.3909017297271, + "learning_rate": 5.648148148148147e-08, + "loss": 1.6647, + "step": 183 + }, + { + "epoch": 0.34074074074074073, + "grad_norm": 15.389123026027889, + "learning_rate": 5.6790123456790115e-08, + "loss": 1.6439, + "step": 184 + }, + { + "epoch": 0.3425925925925926, + "grad_norm": 15.538843832757056, + "learning_rate": 5.709876543209876e-08, + "loss": 1.6501, + "step": 185 + }, + { + "epoch": 0.34444444444444444, + "grad_norm": 16.121666139011364, + "learning_rate": 5.74074074074074e-08, + "loss": 1.6504, + "step": 186 + }, + { + "epoch": 0.34629629629629627, + "grad_norm": 15.597442964448913, + "learning_rate": 5.7716049382716046e-08, + "loss": 1.6622, + "step": 187 + }, + { + "epoch": 0.34814814814814815, + "grad_norm": 16.13349804618981, + "learning_rate": 5.802469135802469e-08, + "loss": 1.6236, + "step": 188 + }, + { + "epoch": 0.35, + "grad_norm": 15.863696027424748, + "learning_rate": 5.833333333333333e-08, + "loss": 1.642, + "step": 189 + }, + { + "epoch": 0.35185185185185186, + "grad_norm": 16.4252638158174, + "learning_rate": 5.8641975308641976e-08, + "loss": 1.6307, + "step": 190 + }, + { + "epoch": 0.3537037037037037, + "grad_norm": 15.841001838915014, + "learning_rate": 5.895061728395061e-08, + "loss": 1.6103, + "step": 191 + }, + { + "epoch": 0.35555555555555557, + "grad_norm": 15.673573923427517, + "learning_rate": 5.925925925925925e-08, + "loss": 1.602, + "step": 192 + }, + { + "epoch": 0.3574074074074074, + "grad_norm": 16.441184669156055, + "learning_rate": 5.9567901234567894e-08, + "loss": 1.602, + "step": 193 + }, + { + "epoch": 0.3592592592592593, + "grad_norm": 16.36704864970962, + "learning_rate": 5.987654320987654e-08, + "loss": 1.6094, + "step": 194 + }, + { + "epoch": 0.3611111111111111, + "grad_norm": 16.143726546373465, + "learning_rate": 6.018518518518519e-08, + "loss": 1.6203, + "step": 195 + }, + { + "epoch": 0.362962962962963, + "grad_norm": 15.767462442066334, + "learning_rate": 6.049382716049382e-08, + "loss": 1.5872, + "step": 196 + }, + { + "epoch": 0.3648148148148148, + "grad_norm": 16.183360356848596, + "learning_rate": 6.080246913580247e-08, + "loss": 1.6067, + "step": 197 + }, + { + "epoch": 0.36666666666666664, + "grad_norm": 15.6568073944227, + "learning_rate": 6.111111111111111e-08, + "loss": 1.5832, + "step": 198 + }, + { + "epoch": 0.3685185185185185, + "grad_norm": 15.031914020275904, + "learning_rate": 6.141975308641975e-08, + "loss": 1.5804, + "step": 199 + }, + { + "epoch": 0.37037037037037035, + "grad_norm": 13.901553800473613, + "learning_rate": 6.172839506172839e-08, + "loss": 1.5736, + "step": 200 + }, + { + "epoch": 0.37222222222222223, + "grad_norm": 13.529862363036859, + "learning_rate": 6.203703703703704e-08, + "loss": 1.5735, + "step": 201 + }, + { + "epoch": 0.37407407407407406, + "grad_norm": 13.769584697382156, + "learning_rate": 6.234567901234567e-08, + "loss": 1.5502, + "step": 202 + }, + { + "epoch": 0.37592592592592594, + "grad_norm": 13.283164609686715, + "learning_rate": 6.265432098765432e-08, + "loss": 1.5369, + "step": 203 + }, + { + "epoch": 0.37777777777777777, + "grad_norm": 13.219835975939468, + "learning_rate": 6.296296296296296e-08, + "loss": 1.5405, + "step": 204 + }, + { + "epoch": 0.37962962962962965, + "grad_norm": 12.823211358025947, + "learning_rate": 6.327160493827161e-08, + "loss": 1.5308, + "step": 205 + }, + { + "epoch": 0.3814814814814815, + "grad_norm": 12.77571724349778, + "learning_rate": 6.358024691358025e-08, + "loss": 1.51, + "step": 206 + }, + { + "epoch": 0.38333333333333336, + "grad_norm": 12.684182910096494, + "learning_rate": 6.388888888888888e-08, + "loss": 1.5063, + "step": 207 + }, + { + "epoch": 0.3851851851851852, + "grad_norm": 11.306595481582432, + "learning_rate": 6.419753086419752e-08, + "loss": 1.5226, + "step": 208 + }, + { + "epoch": 0.387037037037037, + "grad_norm": 11.70198984938417, + "learning_rate": 6.450617283950617e-08, + "loss": 1.5003, + "step": 209 + }, + { + "epoch": 0.3888888888888889, + "grad_norm": 12.39623859805833, + "learning_rate": 6.481481481481481e-08, + "loss": 1.506, + "step": 210 + }, + { + "epoch": 0.3907407407407407, + "grad_norm": 11.180344111163487, + "learning_rate": 6.512345679012346e-08, + "loss": 1.4859, + "step": 211 + }, + { + "epoch": 0.3925925925925926, + "grad_norm": 10.71706262975008, + "learning_rate": 6.54320987654321e-08, + "loss": 1.4878, + "step": 212 + }, + { + "epoch": 0.39444444444444443, + "grad_norm": 10.475139159343728, + "learning_rate": 6.574074074074074e-08, + "loss": 1.477, + "step": 213 + }, + { + "epoch": 0.3962962962962963, + "grad_norm": 10.935818149952677, + "learning_rate": 6.604938271604938e-08, + "loss": 1.4568, + "step": 214 + }, + { + "epoch": 0.39814814814814814, + "grad_norm": 10.847247788731964, + "learning_rate": 6.635802469135802e-08, + "loss": 1.4365, + "step": 215 + }, + { + "epoch": 0.4, + "grad_norm": 10.499892118117987, + "learning_rate": 6.666666666666665e-08, + "loss": 1.4678, + "step": 216 + }, + { + "epoch": 0.40185185185185185, + "grad_norm": 9.807924411811511, + "learning_rate": 6.69753086419753e-08, + "loss": 1.4558, + "step": 217 + }, + { + "epoch": 0.40370370370370373, + "grad_norm": 10.627392105783526, + "learning_rate": 6.728395061728394e-08, + "loss": 1.4156, + "step": 218 + }, + { + "epoch": 0.40555555555555556, + "grad_norm": 10.453639635708425, + "learning_rate": 6.759259259259259e-08, + "loss": 1.4222, + "step": 219 + }, + { + "epoch": 0.4074074074074074, + "grad_norm": 10.436967114672623, + "learning_rate": 6.790123456790123e-08, + "loss": 1.4234, + "step": 220 + }, + { + "epoch": 0.40925925925925927, + "grad_norm": 11.502093278036753, + "learning_rate": 6.820987654320988e-08, + "loss": 1.4168, + "step": 221 + }, + { + "epoch": 0.4111111111111111, + "grad_norm": 12.109663297773258, + "learning_rate": 6.851851851851852e-08, + "loss": 1.364, + "step": 222 + }, + { + "epoch": 0.412962962962963, + "grad_norm": 10.288026845785605, + "learning_rate": 6.882716049382715e-08, + "loss": 1.4177, + "step": 223 + }, + { + "epoch": 0.4148148148148148, + "grad_norm": 10.420630646475221, + "learning_rate": 6.913580246913579e-08, + "loss": 1.4085, + "step": 224 + }, + { + "epoch": 0.4166666666666667, + "grad_norm": 9.784391847322885, + "learning_rate": 6.944444444444444e-08, + "loss": 1.405, + "step": 225 + }, + { + "epoch": 0.4185185185185185, + "grad_norm": 9.671341865243887, + "learning_rate": 6.975308641975308e-08, + "loss": 1.3689, + "step": 226 + }, + { + "epoch": 0.4203703703703704, + "grad_norm": 10.416339402031978, + "learning_rate": 7.006172839506173e-08, + "loss": 1.3843, + "step": 227 + }, + { + "epoch": 0.4222222222222222, + "grad_norm": 9.67646153902481, + "learning_rate": 7.037037037037036e-08, + "loss": 1.3711, + "step": 228 + }, + { + "epoch": 0.42407407407407405, + "grad_norm": 9.224987510519476, + "learning_rate": 7.067901234567901e-08, + "loss": 1.3723, + "step": 229 + }, + { + "epoch": 0.42592592592592593, + "grad_norm": 9.468803750953535, + "learning_rate": 7.098765432098765e-08, + "loss": 1.3412, + "step": 230 + }, + { + "epoch": 0.42777777777777776, + "grad_norm": 9.347082531477147, + "learning_rate": 7.129629629629629e-08, + "loss": 1.3381, + "step": 231 + }, + { + "epoch": 0.42962962962962964, + "grad_norm": 8.898812306680487, + "learning_rate": 7.160493827160492e-08, + "loss": 1.3566, + "step": 232 + }, + { + "epoch": 0.43148148148148147, + "grad_norm": 9.416233114849978, + "learning_rate": 7.191358024691357e-08, + "loss": 1.3309, + "step": 233 + }, + { + "epoch": 0.43333333333333335, + "grad_norm": 9.378481244087665, + "learning_rate": 7.222222222222221e-08, + "loss": 1.2955, + "step": 234 + }, + { + "epoch": 0.4351851851851852, + "grad_norm": 9.590341847604815, + "learning_rate": 7.253086419753086e-08, + "loss": 1.2931, + "step": 235 + }, + { + "epoch": 0.43703703703703706, + "grad_norm": 9.204696036060932, + "learning_rate": 7.28395061728395e-08, + "loss": 1.3017, + "step": 236 + }, + { + "epoch": 0.4388888888888889, + "grad_norm": 9.107752164095162, + "learning_rate": 7.314814814814815e-08, + "loss": 1.2897, + "step": 237 + }, + { + "epoch": 0.44074074074074077, + "grad_norm": 9.003981478035788, + "learning_rate": 7.345679012345679e-08, + "loss": 1.3058, + "step": 238 + }, + { + "epoch": 0.4425925925925926, + "grad_norm": 9.058466378046083, + "learning_rate": 7.376543209876542e-08, + "loss": 1.2834, + "step": 239 + }, + { + "epoch": 0.4444444444444444, + "grad_norm": 8.705122307100883, + "learning_rate": 7.407407407407407e-08, + "loss": 1.2818, + "step": 240 + }, + { + "epoch": 0.4462962962962963, + "grad_norm": 9.205049637998936, + "learning_rate": 7.438271604938271e-08, + "loss": 1.2662, + "step": 241 + }, + { + "epoch": 0.44814814814814813, + "grad_norm": 8.785421682988352, + "learning_rate": 7.469135802469136e-08, + "loss": 1.2836, + "step": 242 + }, + { + "epoch": 0.45, + "grad_norm": 9.191977042098788, + "learning_rate": 7.5e-08, + "loss": 1.2499, + "step": 243 + }, + { + "epoch": 0.45185185185185184, + "grad_norm": 8.598255115042189, + "learning_rate": 7.530864197530863e-08, + "loss": 1.2633, + "step": 244 + }, + { + "epoch": 0.4537037037037037, + "grad_norm": 9.512153148550793, + "learning_rate": 7.561728395061728e-08, + "loss": 1.2379, + "step": 245 + }, + { + "epoch": 0.45555555555555555, + "grad_norm": 8.724604329013456, + "learning_rate": 7.592592592592592e-08, + "loss": 1.2412, + "step": 246 + }, + { + "epoch": 0.45740740740740743, + "grad_norm": 7.934232264621604, + "learning_rate": 7.623456790123457e-08, + "loss": 1.2831, + "step": 247 + }, + { + "epoch": 0.45925925925925926, + "grad_norm": 7.935473144855036, + "learning_rate": 7.654320987654321e-08, + "loss": 1.2201, + "step": 248 + }, + { + "epoch": 0.46111111111111114, + "grad_norm": 8.099377891087682, + "learning_rate": 7.685185185185184e-08, + "loss": 1.2542, + "step": 249 + }, + { + "epoch": 0.46296296296296297, + "grad_norm": 8.406755975217708, + "learning_rate": 7.71604938271605e-08, + "loss": 1.1994, + "step": 250 + }, + { + "epoch": 0.4648148148148148, + "grad_norm": 8.201107335726807, + "learning_rate": 7.746913580246913e-08, + "loss": 1.2031, + "step": 251 + }, + { + "epoch": 0.4666666666666667, + "grad_norm": 8.196934615673412, + "learning_rate": 7.777777777777778e-08, + "loss": 1.1864, + "step": 252 + }, + { + "epoch": 0.4685185185185185, + "grad_norm": 7.230676191162654, + "learning_rate": 7.808641975308642e-08, + "loss": 1.2167, + "step": 253 + }, + { + "epoch": 0.4703703703703704, + "grad_norm": 7.426414216347639, + "learning_rate": 7.839506172839507e-08, + "loss": 1.1935, + "step": 254 + }, + { + "epoch": 0.4722222222222222, + "grad_norm": 7.243389599418745, + "learning_rate": 7.87037037037037e-08, + "loss": 1.235, + "step": 255 + }, + { + "epoch": 0.4740740740740741, + "grad_norm": 7.634374622663417, + "learning_rate": 7.901234567901234e-08, + "loss": 1.1815, + "step": 256 + }, + { + "epoch": 0.4759259259259259, + "grad_norm": 6.941994252703506, + "learning_rate": 7.932098765432098e-08, + "loss": 1.2203, + "step": 257 + }, + { + "epoch": 0.4777777777777778, + "grad_norm": 6.666200938968774, + "learning_rate": 7.962962962962963e-08, + "loss": 1.2104, + "step": 258 + }, + { + "epoch": 0.47962962962962963, + "grad_norm": 7.164672262499252, + "learning_rate": 7.993827160493827e-08, + "loss": 1.1653, + "step": 259 + }, + { + "epoch": 0.48148148148148145, + "grad_norm": 7.227692093630713, + "learning_rate": 8.024691358024692e-08, + "loss": 1.1453, + "step": 260 + }, + { + "epoch": 0.48333333333333334, + "grad_norm": 6.7590989140523465, + "learning_rate": 8.055555555555555e-08, + "loss": 1.1861, + "step": 261 + }, + { + "epoch": 0.48518518518518516, + "grad_norm": 6.595710379020176, + "learning_rate": 8.08641975308642e-08, + "loss": 1.137, + "step": 262 + }, + { + "epoch": 0.48703703703703705, + "grad_norm": 6.367587250699385, + "learning_rate": 8.117283950617284e-08, + "loss": 1.1767, + "step": 263 + }, + { + "epoch": 0.4888888888888889, + "grad_norm": 6.2427486232307094, + "learning_rate": 8.148148148148148e-08, + "loss": 1.1324, + "step": 264 + }, + { + "epoch": 0.49074074074074076, + "grad_norm": 6.079506123745785, + "learning_rate": 8.179012345679011e-08, + "loss": 1.1635, + "step": 265 + }, + { + "epoch": 0.4925925925925926, + "grad_norm": 6.5345425826060195, + "learning_rate": 8.209876543209876e-08, + "loss": 1.1588, + "step": 266 + }, + { + "epoch": 0.49444444444444446, + "grad_norm": 5.767447481234529, + "learning_rate": 8.24074074074074e-08, + "loss": 1.1006, + "step": 267 + }, + { + "epoch": 0.4962962962962963, + "grad_norm": 5.503571960558135, + "learning_rate": 8.271604938271605e-08, + "loss": 1.1806, + "step": 268 + }, + { + "epoch": 0.4981481481481482, + "grad_norm": 5.413380853058772, + "learning_rate": 8.302469135802469e-08, + "loss": 1.1344, + "step": 269 + }, + { + "epoch": 0.5, + "grad_norm": 5.1633535564884285, + "learning_rate": 8.333333333333334e-08, + "loss": 1.1065, + "step": 270 + }, + { + "epoch": 0.5018518518518519, + "grad_norm": 5.096040714664006, + "learning_rate": 8.364197530864198e-08, + "loss": 1.1796, + "step": 271 + }, + { + "epoch": 0.5037037037037037, + "grad_norm": 5.148802562497922, + "learning_rate": 8.395061728395061e-08, + "loss": 1.0927, + "step": 272 + }, + { + "epoch": 0.5055555555555555, + "grad_norm": 4.819744826698795, + "learning_rate": 8.425925925925925e-08, + "loss": 1.1802, + "step": 273 + }, + { + "epoch": 0.5074074074074074, + "grad_norm": 6.940152223659339, + "learning_rate": 8.45679012345679e-08, + "loss": 1.1431, + "step": 274 + }, + { + "epoch": 0.5092592592592593, + "grad_norm": 4.8899229426671935, + "learning_rate": 8.487654320987654e-08, + "loss": 1.1424, + "step": 275 + }, + { + "epoch": 0.5111111111111111, + "grad_norm": 4.633461899062297, + "learning_rate": 8.518518518518519e-08, + "loss": 1.1732, + "step": 276 + }, + { + "epoch": 0.512962962962963, + "grad_norm": 5.366341922696669, + "learning_rate": 8.549382716049382e-08, + "loss": 1.1262, + "step": 277 + }, + { + "epoch": 0.5148148148148148, + "grad_norm": 5.407296170169389, + "learning_rate": 8.580246913580247e-08, + "loss": 1.1421, + "step": 278 + }, + { + "epoch": 0.5166666666666667, + "grad_norm": 4.64055400633339, + "learning_rate": 8.611111111111111e-08, + "loss": 1.1244, + "step": 279 + }, + { + "epoch": 0.5185185185185185, + "grad_norm": 4.790759190356316, + "learning_rate": 8.641975308641975e-08, + "loss": 1.0795, + "step": 280 + }, + { + "epoch": 0.5203703703703704, + "grad_norm": 4.508466333612143, + "learning_rate": 8.672839506172838e-08, + "loss": 1.1168, + "step": 281 + }, + { + "epoch": 0.5222222222222223, + "grad_norm": 4.601535680356002, + "learning_rate": 8.703703703703703e-08, + "loss": 1.0751, + "step": 282 + }, + { + "epoch": 0.524074074074074, + "grad_norm": 4.356061978717215, + "learning_rate": 8.734567901234567e-08, + "loss": 1.1383, + "step": 283 + }, + { + "epoch": 0.5259259259259259, + "grad_norm": 4.739645434419619, + "learning_rate": 8.765432098765432e-08, + "loss": 1.0895, + "step": 284 + }, + { + "epoch": 0.5277777777777778, + "grad_norm": 4.557295878083224, + "learning_rate": 8.796296296296296e-08, + "loss": 1.0586, + "step": 285 + }, + { + "epoch": 0.5296296296296297, + "grad_norm": 4.983263941075996, + "learning_rate": 8.827160493827161e-08, + "loss": 1.1093, + "step": 286 + }, + { + "epoch": 0.5314814814814814, + "grad_norm": 4.322416983003166, + "learning_rate": 8.858024691358025e-08, + "loss": 1.0785, + "step": 287 + }, + { + "epoch": 0.5333333333333333, + "grad_norm": 4.421637263273455, + "learning_rate": 8.888888888888888e-08, + "loss": 1.1444, + "step": 288 + }, + { + "epoch": 0.5351851851851852, + "grad_norm": 4.222520171046782, + "learning_rate": 8.919753086419752e-08, + "loss": 1.1139, + "step": 289 + }, + { + "epoch": 0.5370370370370371, + "grad_norm": 4.283460517774991, + "learning_rate": 8.950617283950617e-08, + "loss": 1.1247, + "step": 290 + }, + { + "epoch": 0.5388888888888889, + "grad_norm": 4.182758065374512, + "learning_rate": 8.98148148148148e-08, + "loss": 1.0799, + "step": 291 + }, + { + "epoch": 0.5407407407407407, + "grad_norm": 4.475415379775168, + "learning_rate": 9.012345679012346e-08, + "loss": 1.0955, + "step": 292 + }, + { + "epoch": 0.5425925925925926, + "grad_norm": 4.568332544127409, + "learning_rate": 9.043209876543209e-08, + "loss": 1.0592, + "step": 293 + }, + { + "epoch": 0.5444444444444444, + "grad_norm": 4.117536951160185, + "learning_rate": 9.074074074074074e-08, + "loss": 1.1324, + "step": 294 + }, + { + "epoch": 0.5462962962962963, + "grad_norm": 4.316537569625379, + "learning_rate": 9.104938271604938e-08, + "loss": 1.107, + "step": 295 + }, + { + "epoch": 0.5481481481481482, + "grad_norm": 4.1795506768589155, + "learning_rate": 9.135802469135802e-08, + "loss": 1.0782, + "step": 296 + }, + { + "epoch": 0.55, + "grad_norm": 4.359117337896192, + "learning_rate": 9.166666666666665e-08, + "loss": 1.0395, + "step": 297 + }, + { + "epoch": 0.5518518518518518, + "grad_norm": 4.11711181502351, + "learning_rate": 9.19753086419753e-08, + "loss": 1.0698, + "step": 298 + }, + { + "epoch": 0.5537037037037037, + "grad_norm": 6.256660607973502, + "learning_rate": 9.228395061728394e-08, + "loss": 1.0683, + "step": 299 + }, + { + "epoch": 0.5555555555555556, + "grad_norm": 4.040085947010745, + "learning_rate": 9.259259259259259e-08, + "loss": 1.055, + "step": 300 + }, + { + "epoch": 0.5574074074074075, + "grad_norm": 4.0753999882391385, + "learning_rate": 9.290123456790123e-08, + "loss": 1.0595, + "step": 301 + }, + { + "epoch": 0.5592592592592592, + "grad_norm": 3.8715528015682525, + "learning_rate": 9.320987654320988e-08, + "loss": 1.0637, + "step": 302 + }, + { + "epoch": 0.5611111111111111, + "grad_norm": 4.147891280688802, + "learning_rate": 9.351851851851851e-08, + "loss": 1.0428, + "step": 303 + }, + { + "epoch": 0.562962962962963, + "grad_norm": 4.475004281259303, + "learning_rate": 9.382716049382715e-08, + "loss": 1.0994, + "step": 304 + }, + { + "epoch": 0.5648148148148148, + "grad_norm": 3.804844933042943, + "learning_rate": 9.413580246913579e-08, + "loss": 1.0936, + "step": 305 + }, + { + "epoch": 0.5666666666666667, + "grad_norm": 3.93283281444395, + "learning_rate": 9.444444444444444e-08, + "loss": 1.0874, + "step": 306 + }, + { + "epoch": 0.5685185185185185, + "grad_norm": 3.8545783040385437, + "learning_rate": 9.475308641975308e-08, + "loss": 1.0828, + "step": 307 + }, + { + "epoch": 0.5703703703703704, + "grad_norm": 3.879617339937958, + "learning_rate": 9.506172839506173e-08, + "loss": 1.0993, + "step": 308 + }, + { + "epoch": 0.5722222222222222, + "grad_norm": 3.853722588826157, + "learning_rate": 9.537037037037036e-08, + "loss": 1.0596, + "step": 309 + }, + { + "epoch": 0.5740740740740741, + "grad_norm": 4.4180795472700485, + "learning_rate": 9.567901234567901e-08, + "loss": 1.056, + "step": 310 + }, + { + "epoch": 0.575925925925926, + "grad_norm": 4.13332656152739, + "learning_rate": 9.598765432098765e-08, + "loss": 1.0385, + "step": 311 + }, + { + "epoch": 0.5777777777777777, + "grad_norm": 3.8698195381240224, + "learning_rate": 9.629629629629629e-08, + "loss": 1.0905, + "step": 312 + }, + { + "epoch": 0.5796296296296296, + "grad_norm": 3.7649247371434913, + "learning_rate": 9.660493827160492e-08, + "loss": 1.0244, + "step": 313 + }, + { + "epoch": 0.5814814814814815, + "grad_norm": 3.886218344399538, + "learning_rate": 9.691358024691357e-08, + "loss": 1.0756, + "step": 314 + }, + { + "epoch": 0.5833333333333334, + "grad_norm": 3.832552287760325, + "learning_rate": 9.722222222222221e-08, + "loss": 1.0114, + "step": 315 + }, + { + "epoch": 0.5851851851851851, + "grad_norm": 3.6146573722704267, + "learning_rate": 9.753086419753086e-08, + "loss": 1.0214, + "step": 316 + }, + { + "epoch": 0.587037037037037, + "grad_norm": 3.7899711682732704, + "learning_rate": 9.78395061728395e-08, + "loss": 1.0346, + "step": 317 + }, + { + "epoch": 0.5888888888888889, + "grad_norm": 3.8241333661862056, + "learning_rate": 9.814814814814815e-08, + "loss": 1.0269, + "step": 318 + }, + { + "epoch": 0.5907407407407408, + "grad_norm": 3.9222450614025743, + "learning_rate": 9.845679012345678e-08, + "loss": 1.036, + "step": 319 + }, + { + "epoch": 0.5925925925925926, + "grad_norm": 3.6406348375553756, + "learning_rate": 9.876543209876542e-08, + "loss": 1.0167, + "step": 320 + }, + { + "epoch": 0.5944444444444444, + "grad_norm": 3.6373113117333564, + "learning_rate": 9.907407407407407e-08, + "loss": 1.0365, + "step": 321 + }, + { + "epoch": 0.5962962962962963, + "grad_norm": 3.841999292920067, + "learning_rate": 9.938271604938271e-08, + "loss": 0.9952, + "step": 322 + }, + { + "epoch": 0.5981481481481481, + "grad_norm": 3.5190549270282756, + "learning_rate": 9.969135802469136e-08, + "loss": 1.0176, + "step": 323 + }, + { + "epoch": 0.6, + "grad_norm": 3.9277792826794435, + "learning_rate": 1e-07, + "loss": 1.0132, + "step": 324 + }, + { + "epoch": 0.6018518518518519, + "grad_norm": 3.4991407608885914, + "learning_rate": 1.0030864197530863e-07, + "loss": 1.068, + "step": 325 + }, + { + "epoch": 0.6037037037037037, + "grad_norm": 3.500378688791865, + "learning_rate": 1.0061728395061728e-07, + "loss": 1.0703, + "step": 326 + }, + { + "epoch": 0.6055555555555555, + "grad_norm": 4.193052022493057, + "learning_rate": 1.0092592592592592e-07, + "loss": 0.9992, + "step": 327 + }, + { + "epoch": 0.6074074074074074, + "grad_norm": 4.091013817038656, + "learning_rate": 1.0123456790123457e-07, + "loss": 1.0896, + "step": 328 + }, + { + "epoch": 0.6092592592592593, + "grad_norm": 3.7112977635689233, + "learning_rate": 1.015432098765432e-07, + "loss": 0.984, + "step": 329 + }, + { + "epoch": 0.6111111111111112, + "grad_norm": 3.360409883844207, + "learning_rate": 1.0185185185185186e-07, + "loss": 1.035, + "step": 330 + }, + { + "epoch": 0.6129629629629629, + "grad_norm": 3.433642670718037, + "learning_rate": 1.021604938271605e-07, + "loss": 1.035, + "step": 331 + }, + { + "epoch": 0.6148148148148148, + "grad_norm": 3.212656245186043, + "learning_rate": 1.0246913580246914e-07, + "loss": 1.0085, + "step": 332 + }, + { + "epoch": 0.6166666666666667, + "grad_norm": 3.254282617277017, + "learning_rate": 1.0277777777777777e-07, + "loss": 1.0448, + "step": 333 + }, + { + "epoch": 0.6185185185185185, + "grad_norm": 3.518638459412198, + "learning_rate": 1.030864197530864e-07, + "loss": 0.9822, + "step": 334 + }, + { + "epoch": 0.6203703703703703, + "grad_norm": 3.203534842703008, + "learning_rate": 1.0339506172839505e-07, + "loss": 1.0235, + "step": 335 + }, + { + "epoch": 0.6222222222222222, + "grad_norm": 3.2226059542304553, + "learning_rate": 1.0370370370370369e-07, + "loss": 1.0092, + "step": 336 + }, + { + "epoch": 0.6240740740740741, + "grad_norm": 3.1951991177522103, + "learning_rate": 1.0401234567901234e-07, + "loss": 1.0156, + "step": 337 + }, + { + "epoch": 0.6259259259259259, + "grad_norm": 3.1456877871563607, + "learning_rate": 1.0432098765432098e-07, + "loss": 0.9946, + "step": 338 + }, + { + "epoch": 0.6277777777777778, + "grad_norm": 3.754516469146756, + "learning_rate": 1.0462962962962963e-07, + "loss": 0.9868, + "step": 339 + }, + { + "epoch": 0.6296296296296297, + "grad_norm": 3.7534332564859816, + "learning_rate": 1.0493827160493827e-07, + "loss": 1.0737, + "step": 340 + }, + { + "epoch": 0.6314814814814815, + "grad_norm": 3.3612600856994677, + "learning_rate": 1.0524691358024692e-07, + "loss": 0.9916, + "step": 341 + }, + { + "epoch": 0.6333333333333333, + "grad_norm": 3.0886146650751676, + "learning_rate": 1.0555555555555555e-07, + "loss": 1.0262, + "step": 342 + }, + { + "epoch": 0.6351851851851852, + "grad_norm": 10.782510450941533, + "learning_rate": 1.058641975308642e-07, + "loss": 0.9866, + "step": 343 + }, + { + "epoch": 0.6370370370370371, + "grad_norm": 3.142360207609036, + "learning_rate": 1.0617283950617284e-07, + "loss": 0.9486, + "step": 344 + }, + { + "epoch": 0.6388888888888888, + "grad_norm": 3.0910722749301893, + "learning_rate": 1.0648148148148149e-07, + "loss": 0.9855, + "step": 345 + }, + { + "epoch": 0.6407407407407407, + "grad_norm": 3.0858788934690446, + "learning_rate": 1.0679012345679013e-07, + "loss": 1.0133, + "step": 346 + }, + { + "epoch": 0.6425925925925926, + "grad_norm": 3.0911926600551696, + "learning_rate": 1.0709876543209878e-07, + "loss": 0.9406, + "step": 347 + }, + { + "epoch": 0.6444444444444445, + "grad_norm": 3.0597800798170414, + "learning_rate": 1.0740740740740741e-07, + "loss": 0.9874, + "step": 348 + }, + { + "epoch": 0.6462962962962963, + "grad_norm": 2.981092097330975, + "learning_rate": 1.0771604938271604e-07, + "loss": 0.9965, + "step": 349 + }, + { + "epoch": 0.6481481481481481, + "grad_norm": 2.9987104420125354, + "learning_rate": 1.0802469135802467e-07, + "loss": 0.9332, + "step": 350 + }, + { + "epoch": 0.65, + "grad_norm": 3.04519271249131, + "learning_rate": 1.0833333333333332e-07, + "loss": 1.0534, + "step": 351 + }, + { + "epoch": 0.6518518518518519, + "grad_norm": 2.849398460345178, + "learning_rate": 1.0864197530864196e-07, + "loss": 0.9925, + "step": 352 + }, + { + "epoch": 0.6537037037037037, + "grad_norm": 2.8476217984678374, + "learning_rate": 1.0895061728395061e-07, + "loss": 1.0026, + "step": 353 + }, + { + "epoch": 0.6555555555555556, + "grad_norm": 3.2860186349924936, + "learning_rate": 1.0925925925925925e-07, + "loss": 0.9693, + "step": 354 + }, + { + "epoch": 0.6574074074074074, + "grad_norm": 2.9055661435857685, + "learning_rate": 1.095679012345679e-07, + "loss": 0.9959, + "step": 355 + }, + { + "epoch": 0.6592592592592592, + "grad_norm": 2.8698105783205876, + "learning_rate": 1.0987654320987653e-07, + "loss": 1.0042, + "step": 356 + }, + { + "epoch": 0.6611111111111111, + "grad_norm": 2.8994083090789986, + "learning_rate": 1.1018518518518519e-07, + "loss": 0.9425, + "step": 357 + }, + { + "epoch": 0.662962962962963, + "grad_norm": 3.0711753606889958, + "learning_rate": 1.1049382716049382e-07, + "loss": 0.9382, + "step": 358 + }, + { + "epoch": 0.6648148148148149, + "grad_norm": 3.055805679959846, + "learning_rate": 1.1080246913580247e-07, + "loss": 0.9751, + "step": 359 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 2.7437744731379006, + "learning_rate": 1.1111111111111111e-07, + "loss": 0.9771, + "step": 360 + }, + { + "epoch": 0.6685185185185185, + "grad_norm": 3.2083309317148307, + "learning_rate": 1.1141975308641976e-07, + "loss": 0.9216, + "step": 361 + }, + { + "epoch": 0.6703703703703704, + "grad_norm": 2.9211725325299085, + "learning_rate": 1.117283950617284e-07, + "loss": 0.992, + "step": 362 + }, + { + "epoch": 0.6722222222222223, + "grad_norm": 3.3679385542846547, + "learning_rate": 1.1203703703703705e-07, + "loss": 0.9802, + "step": 363 + }, + { + "epoch": 0.674074074074074, + "grad_norm": 2.6680671454091813, + "learning_rate": 1.1234567901234568e-07, + "loss": 0.9565, + "step": 364 + }, + { + "epoch": 0.6759259259259259, + "grad_norm": 2.9327345491165775, + "learning_rate": 1.1265432098765431e-07, + "loss": 0.9714, + "step": 365 + }, + { + "epoch": 0.6777777777777778, + "grad_norm": 2.914073529883788, + "learning_rate": 1.1296296296296294e-07, + "loss": 0.9827, + "step": 366 + }, + { + "epoch": 0.6796296296296296, + "grad_norm": 2.886811903293114, + "learning_rate": 1.132716049382716e-07, + "loss": 0.911, + "step": 367 + }, + { + "epoch": 0.6814814814814815, + "grad_norm": 3.0308042131753994, + "learning_rate": 1.1358024691358023e-07, + "loss": 1.0101, + "step": 368 + }, + { + "epoch": 0.6833333333333333, + "grad_norm": 2.7958369287491625, + "learning_rate": 1.1388888888888888e-07, + "loss": 0.8902, + "step": 369 + }, + { + "epoch": 0.6851851851851852, + "grad_norm": 2.6122766390774563, + "learning_rate": 1.1419753086419752e-07, + "loss": 0.9456, + "step": 370 + }, + { + "epoch": 0.687037037037037, + "grad_norm": 2.662744553113895, + "learning_rate": 1.1450617283950617e-07, + "loss": 0.9444, + "step": 371 + }, + { + "epoch": 0.6888888888888889, + "grad_norm": 2.819124116450683, + "learning_rate": 1.148148148148148e-07, + "loss": 0.9955, + "step": 372 + }, + { + "epoch": 0.6907407407407408, + "grad_norm": 2.900655599234632, + "learning_rate": 1.1512345679012345e-07, + "loss": 1.0019, + "step": 373 + }, + { + "epoch": 0.6925925925925925, + "grad_norm": 2.706102039336673, + "learning_rate": 1.1543209876543209e-07, + "loss": 1.0713, + "step": 374 + }, + { + "epoch": 0.6944444444444444, + "grad_norm": 2.7726424374425425, + "learning_rate": 1.1574074074074074e-07, + "loss": 1.0292, + "step": 375 + }, + { + "epoch": 0.6962962962962963, + "grad_norm": 2.7566159571889886, + "learning_rate": 1.1604938271604938e-07, + "loss": 0.9611, + "step": 376 + }, + { + "epoch": 0.6981481481481482, + "grad_norm": 3.1510749199541443, + "learning_rate": 1.1635802469135803e-07, + "loss": 0.9945, + "step": 377 + }, + { + "epoch": 0.7, + "grad_norm": 2.7425882172844007, + "learning_rate": 1.1666666666666667e-07, + "loss": 0.9356, + "step": 378 + }, + { + "epoch": 0.7018518518518518, + "grad_norm": 2.4352316404435252, + "learning_rate": 1.1697530864197532e-07, + "loss": 0.9164, + "step": 379 + }, + { + "epoch": 0.7037037037037037, + "grad_norm": 2.3459367599905896, + "learning_rate": 1.1728395061728395e-07, + "loss": 0.9506, + "step": 380 + }, + { + "epoch": 0.7055555555555556, + "grad_norm": 3.0041608313424173, + "learning_rate": 1.1759259259259258e-07, + "loss": 0.9725, + "step": 381 + }, + { + "epoch": 0.7074074074074074, + "grad_norm": 2.742575853590362, + "learning_rate": 1.1790123456790121e-07, + "loss": 0.9795, + "step": 382 + }, + { + "epoch": 0.7092592592592593, + "grad_norm": 4.840488192646905, + "learning_rate": 1.1820987654320986e-07, + "loss": 0.9137, + "step": 383 + }, + { + "epoch": 0.7111111111111111, + "grad_norm": 2.479164235734708, + "learning_rate": 1.185185185185185e-07, + "loss": 0.9409, + "step": 384 + }, + { + "epoch": 0.7129629629629629, + "grad_norm": 2.4522212362192892, + "learning_rate": 1.1882716049382715e-07, + "loss": 0.9407, + "step": 385 + }, + { + "epoch": 0.7148148148148148, + "grad_norm": 2.3959924395157777, + "learning_rate": 1.1913580246913579e-07, + "loss": 0.9627, + "step": 386 + }, + { + "epoch": 0.7166666666666667, + "grad_norm": 2.403811670042839, + "learning_rate": 1.1944444444444442e-07, + "loss": 0.9516, + "step": 387 + }, + { + "epoch": 0.7185185185185186, + "grad_norm": 2.3419572615787816, + "learning_rate": 1.1975308641975307e-07, + "loss": 0.9268, + "step": 388 + }, + { + "epoch": 0.7203703703703703, + "grad_norm": 2.497470338646794, + "learning_rate": 1.2006172839506172e-07, + "loss": 0.9109, + "step": 389 + }, + { + "epoch": 0.7222222222222222, + "grad_norm": 2.324528256770774, + "learning_rate": 1.2037037037037037e-07, + "loss": 0.9467, + "step": 390 + }, + { + "epoch": 0.7240740740740741, + "grad_norm": 2.3451110299913736, + "learning_rate": 1.20679012345679e-07, + "loss": 0.9169, + "step": 391 + }, + { + "epoch": 0.725925925925926, + "grad_norm": 3.316057861621452, + "learning_rate": 1.2098765432098765e-07, + "loss": 0.9395, + "step": 392 + }, + { + "epoch": 0.7277777777777777, + "grad_norm": 2.9385499228836838, + "learning_rate": 1.212962962962963e-07, + "loss": 0.9744, + "step": 393 + }, + { + "epoch": 0.7296296296296296, + "grad_norm": 2.2885592797725165, + "learning_rate": 1.2160493827160495e-07, + "loss": 0.9238, + "step": 394 + }, + { + "epoch": 0.7314814814814815, + "grad_norm": 2.2054451654866707, + "learning_rate": 1.2191358024691357e-07, + "loss": 0.9436, + "step": 395 + }, + { + "epoch": 0.7333333333333333, + "grad_norm": 2.5514270934723067, + "learning_rate": 1.2222222222222222e-07, + "loss": 0.9315, + "step": 396 + }, + { + "epoch": 0.7351851851851852, + "grad_norm": 2.2368117556543456, + "learning_rate": 1.2253086419753085e-07, + "loss": 0.9638, + "step": 397 + }, + { + "epoch": 0.737037037037037, + "grad_norm": 2.266035335465804, + "learning_rate": 1.228395061728395e-07, + "loss": 0.8805, + "step": 398 + }, + { + "epoch": 0.7388888888888889, + "grad_norm": 2.2358935123241, + "learning_rate": 1.2314814814814815e-07, + "loss": 0.9705, + "step": 399 + }, + { + "epoch": 0.7407407407407407, + "grad_norm": 2.31380824764338, + "learning_rate": 1.2345679012345677e-07, + "loss": 0.9071, + "step": 400 + }, + { + "epoch": 0.7425925925925926, + "grad_norm": 2.5111441323147985, + "learning_rate": 1.2376543209876542e-07, + "loss": 0.9094, + "step": 401 + }, + { + "epoch": 0.7444444444444445, + "grad_norm": 2.3304081052081282, + "learning_rate": 1.2407407407407407e-07, + "loss": 0.93, + "step": 402 + }, + { + "epoch": 0.7462962962962963, + "grad_norm": 2.37840030150777, + "learning_rate": 1.2438271604938272e-07, + "loss": 0.9203, + "step": 403 + }, + { + "epoch": 0.7481481481481481, + "grad_norm": 2.432117775628416, + "learning_rate": 1.2469135802469134e-07, + "loss": 0.904, + "step": 404 + }, + { + "epoch": 0.75, + "grad_norm": 3.0092139480408537, + "learning_rate": 1.25e-07, + "loss": 0.9309, + "step": 405 + }, + { + "epoch": 0.7518518518518519, + "grad_norm": 3.672761262710248, + "learning_rate": 1.2530864197530864e-07, + "loss": 0.9027, + "step": 406 + }, + { + "epoch": 0.7537037037037037, + "grad_norm": 2.1249398432733253, + "learning_rate": 1.2561728395061727e-07, + "loss": 0.9258, + "step": 407 + }, + { + "epoch": 0.7555555555555555, + "grad_norm": 2.361832315773115, + "learning_rate": 1.2592592592592592e-07, + "loss": 0.8982, + "step": 408 + }, + { + "epoch": 0.7574074074074074, + "grad_norm": 2.2076222382736828, + "learning_rate": 1.2623456790123457e-07, + "loss": 0.9559, + "step": 409 + }, + { + "epoch": 0.7592592592592593, + "grad_norm": 2.319232858121409, + "learning_rate": 1.2654320987654322e-07, + "loss": 0.9057, + "step": 410 + }, + { + "epoch": 0.7611111111111111, + "grad_norm": 2.227871917953055, + "learning_rate": 1.2685185185185184e-07, + "loss": 0.9421, + "step": 411 + }, + { + "epoch": 0.762962962962963, + "grad_norm": 2.223934637384949, + "learning_rate": 1.271604938271605e-07, + "loss": 0.9403, + "step": 412 + }, + { + "epoch": 0.7648148148148148, + "grad_norm": 2.142718818730283, + "learning_rate": 1.2746913580246914e-07, + "loss": 0.91, + "step": 413 + }, + { + "epoch": 0.7666666666666667, + "grad_norm": 2.0779692106131264, + "learning_rate": 1.2777777777777777e-07, + "loss": 0.9485, + "step": 414 + }, + { + "epoch": 0.7685185185185185, + "grad_norm": 2.342474251086406, + "learning_rate": 1.2808641975308642e-07, + "loss": 0.9255, + "step": 415 + }, + { + "epoch": 0.7703703703703704, + "grad_norm": 2.329358870521356, + "learning_rate": 1.2839506172839504e-07, + "loss": 0.9434, + "step": 416 + }, + { + "epoch": 0.7722222222222223, + "grad_norm": 2.21863092723048, + "learning_rate": 1.287037037037037e-07, + "loss": 0.9363, + "step": 417 + }, + { + "epoch": 0.774074074074074, + "grad_norm": 2.033704886221468, + "learning_rate": 1.2901234567901234e-07, + "loss": 0.8838, + "step": 418 + }, + { + "epoch": 0.7759259259259259, + "grad_norm": 1.9977248258436775, + "learning_rate": 1.29320987654321e-07, + "loss": 0.8925, + "step": 419 + }, + { + "epoch": 0.7777777777777778, + "grad_norm": 2.04238549943764, + "learning_rate": 1.2962962962962961e-07, + "loss": 0.9343, + "step": 420 + }, + { + "epoch": 0.7796296296296297, + "grad_norm": 2.2302422155062276, + "learning_rate": 1.2993827160493826e-07, + "loss": 0.9192, + "step": 421 + }, + { + "epoch": 0.7814814814814814, + "grad_norm": 1.9950899526602213, + "learning_rate": 1.3024691358024691e-07, + "loss": 0.8803, + "step": 422 + }, + { + "epoch": 0.7833333333333333, + "grad_norm": 2.2831484969299587, + "learning_rate": 1.3055555555555556e-07, + "loss": 0.8651, + "step": 423 + }, + { + "epoch": 0.7851851851851852, + "grad_norm": 2.344415429236907, + "learning_rate": 1.308641975308642e-07, + "loss": 0.8864, + "step": 424 + }, + { + "epoch": 0.7870370370370371, + "grad_norm": 2.222725365830128, + "learning_rate": 1.3117283950617284e-07, + "loss": 0.9135, + "step": 425 + }, + { + "epoch": 0.7888888888888889, + "grad_norm": 2.2566021695397906, + "learning_rate": 1.314814814814815e-07, + "loss": 0.9597, + "step": 426 + }, + { + "epoch": 0.7907407407407407, + "grad_norm": 2.237807988628356, + "learning_rate": 1.3179012345679014e-07, + "loss": 0.8831, + "step": 427 + }, + { + "epoch": 0.7925925925925926, + "grad_norm": 2.297155195766305, + "learning_rate": 1.3209876543209876e-07, + "loss": 0.9147, + "step": 428 + }, + { + "epoch": 0.7944444444444444, + "grad_norm": 2.809965128994331, + "learning_rate": 1.324074074074074e-07, + "loss": 0.9325, + "step": 429 + }, + { + "epoch": 0.7962962962962963, + "grad_norm": 2.025045575533884, + "learning_rate": 1.3271604938271604e-07, + "loss": 0.9587, + "step": 430 + }, + { + "epoch": 0.7981481481481482, + "grad_norm": 2.0159452459757143, + "learning_rate": 1.3302469135802469e-07, + "loss": 0.9072, + "step": 431 + }, + { + "epoch": 0.8, + "grad_norm": 2.029223716341751, + "learning_rate": 1.333333333333333e-07, + "loss": 0.8628, + "step": 432 + }, + { + "epoch": 0.8018518518518518, + "grad_norm": 1.921147791939492, + "learning_rate": 1.3364197530864196e-07, + "loss": 0.8962, + "step": 433 + }, + { + "epoch": 0.8037037037037037, + "grad_norm": 1.998351573317739, + "learning_rate": 1.339506172839506e-07, + "loss": 0.9269, + "step": 434 + }, + { + "epoch": 0.8055555555555556, + "grad_norm": 2.202907873704923, + "learning_rate": 1.3425925925925926e-07, + "loss": 0.9064, + "step": 435 + }, + { + "epoch": 0.8074074074074075, + "grad_norm": 2.070570348046798, + "learning_rate": 1.3456790123456788e-07, + "loss": 0.9168, + "step": 436 + }, + { + "epoch": 0.8092592592592592, + "grad_norm": 2.0174231832876517, + "learning_rate": 1.3487654320987653e-07, + "loss": 0.8804, + "step": 437 + }, + { + "epoch": 0.8111111111111111, + "grad_norm": 2.246933453381705, + "learning_rate": 1.3518518518518518e-07, + "loss": 0.9697, + "step": 438 + }, + { + "epoch": 0.812962962962963, + "grad_norm": 2.199080672800522, + "learning_rate": 1.3549382716049383e-07, + "loss": 0.9057, + "step": 439 + }, + { + "epoch": 0.8148148148148148, + "grad_norm": 2.2364283143669286, + "learning_rate": 1.3580246913580246e-07, + "loss": 0.8524, + "step": 440 + }, + { + "epoch": 0.8166666666666667, + "grad_norm": 4.024158243823565, + "learning_rate": 1.361111111111111e-07, + "loss": 0.9327, + "step": 441 + }, + { + "epoch": 0.8185185185185185, + "grad_norm": 2.044294128111021, + "learning_rate": 1.3641975308641976e-07, + "loss": 0.8726, + "step": 442 + }, + { + "epoch": 0.8203703703703704, + "grad_norm": 1.8158630448309119, + "learning_rate": 1.367283950617284e-07, + "loss": 0.8908, + "step": 443 + }, + { + "epoch": 0.8222222222222222, + "grad_norm": 2.4331116688216783, + "learning_rate": 1.3703703703703703e-07, + "loss": 0.8874, + "step": 444 + }, + { + "epoch": 0.8240740740740741, + "grad_norm": 1.9981430466466217, + "learning_rate": 1.3734567901234568e-07, + "loss": 0.869, + "step": 445 + }, + { + "epoch": 0.825925925925926, + "grad_norm": 1.9795801077027648, + "learning_rate": 1.376543209876543e-07, + "loss": 0.8912, + "step": 446 + }, + { + "epoch": 0.8277777777777777, + "grad_norm": 1.9397925528326283, + "learning_rate": 1.3796296296296296e-07, + "loss": 0.9536, + "step": 447 + }, + { + "epoch": 0.8296296296296296, + "grad_norm": 2.2321691004910305, + "learning_rate": 1.3827160493827158e-07, + "loss": 0.9704, + "step": 448 + }, + { + "epoch": 0.8314814814814815, + "grad_norm": 1.876055348972833, + "learning_rate": 1.3858024691358023e-07, + "loss": 0.8835, + "step": 449 + }, + { + "epoch": 0.8333333333333334, + "grad_norm": 1.9096879514198053, + "learning_rate": 1.3888888888888888e-07, + "loss": 0.8971, + "step": 450 + }, + { + "epoch": 0.8351851851851851, + "grad_norm": 1.8105515638937926, + "learning_rate": 1.3919753086419753e-07, + "loss": 0.9556, + "step": 451 + }, + { + "epoch": 0.837037037037037, + "grad_norm": 2.0514395378516252, + "learning_rate": 1.3950617283950615e-07, + "loss": 0.8561, + "step": 452 + }, + { + "epoch": 0.8388888888888889, + "grad_norm": 1.8642844364032094, + "learning_rate": 1.398148148148148e-07, + "loss": 0.8433, + "step": 453 + }, + { + "epoch": 0.8407407407407408, + "grad_norm": 1.804313908556319, + "learning_rate": 1.4012345679012345e-07, + "loss": 0.8654, + "step": 454 + }, + { + "epoch": 0.8425925925925926, + "grad_norm": 1.8953109256785983, + "learning_rate": 1.404320987654321e-07, + "loss": 0.9814, + "step": 455 + }, + { + "epoch": 0.8444444444444444, + "grad_norm": 2.1225876682364273, + "learning_rate": 1.4074074074074073e-07, + "loss": 0.9057, + "step": 456 + }, + { + "epoch": 0.8462962962962963, + "grad_norm": 2.2016444974916323, + "learning_rate": 1.4104938271604938e-07, + "loss": 0.8201, + "step": 457 + }, + { + "epoch": 0.8481481481481481, + "grad_norm": 2.0432606819951524, + "learning_rate": 1.4135802469135803e-07, + "loss": 0.8842, + "step": 458 + }, + { + "epoch": 0.85, + "grad_norm": 2.1280328243286752, + "learning_rate": 1.4166666666666668e-07, + "loss": 0.8402, + "step": 459 + }, + { + "epoch": 0.8518518518518519, + "grad_norm": 2.000472025214132, + "learning_rate": 1.419753086419753e-07, + "loss": 0.8819, + "step": 460 + }, + { + "epoch": 0.8537037037037037, + "grad_norm": 1.8512466548066744, + "learning_rate": 1.4228395061728395e-07, + "loss": 0.9432, + "step": 461 + }, + { + "epoch": 0.8555555555555555, + "grad_norm": 2.0106659004010137, + "learning_rate": 1.4259259259259258e-07, + "loss": 0.8719, + "step": 462 + }, + { + "epoch": 0.8574074074074074, + "grad_norm": 1.8848461393858795, + "learning_rate": 1.4290123456790123e-07, + "loss": 0.9284, + "step": 463 + }, + { + "epoch": 0.8592592592592593, + "grad_norm": 2.530928394391292, + "learning_rate": 1.4320987654320985e-07, + "loss": 0.9414, + "step": 464 + }, + { + "epoch": 0.8611111111111112, + "grad_norm": 1.7563339526281114, + "learning_rate": 1.435185185185185e-07, + "loss": 0.839, + "step": 465 + }, + { + "epoch": 0.8629629629629629, + "grad_norm": 1.7548405700159353, + "learning_rate": 1.4382716049382715e-07, + "loss": 0.8824, + "step": 466 + }, + { + "epoch": 0.8648148148148148, + "grad_norm": 5.659133892657317, + "learning_rate": 1.441358024691358e-07, + "loss": 0.8713, + "step": 467 + }, + { + "epoch": 0.8666666666666667, + "grad_norm": 3.3047969228291567, + "learning_rate": 1.4444444444444442e-07, + "loss": 0.9101, + "step": 468 + }, + { + "epoch": 0.8685185185185185, + "grad_norm": 1.8286881417606347, + "learning_rate": 1.4475308641975307e-07, + "loss": 0.8855, + "step": 469 + }, + { + "epoch": 0.8703703703703703, + "grad_norm": 1.713564429231544, + "learning_rate": 1.4506172839506172e-07, + "loss": 0.8755, + "step": 470 + }, + { + "epoch": 0.8722222222222222, + "grad_norm": 1.7522479788067298, + "learning_rate": 1.4537037037037037e-07, + "loss": 0.8801, + "step": 471 + }, + { + "epoch": 0.8740740740740741, + "grad_norm": 1.7913855849127627, + "learning_rate": 1.45679012345679e-07, + "loss": 0.8893, + "step": 472 + }, + { + "epoch": 0.8759259259259259, + "grad_norm": 1.8590698519635387, + "learning_rate": 1.4598765432098765e-07, + "loss": 0.8477, + "step": 473 + }, + { + "epoch": 0.8777777777777778, + "grad_norm": 1.9168882323220686, + "learning_rate": 1.462962962962963e-07, + "loss": 0.8839, + "step": 474 + }, + { + "epoch": 0.8796296296296297, + "grad_norm": 1.8287509487852251, + "learning_rate": 1.4660493827160495e-07, + "loss": 0.934, + "step": 475 + }, + { + "epoch": 0.8814814814814815, + "grad_norm": 1.7318355238854954, + "learning_rate": 1.4691358024691357e-07, + "loss": 0.8592, + "step": 476 + }, + { + "epoch": 0.8833333333333333, + "grad_norm": 1.8155966159699268, + "learning_rate": 1.4722222222222222e-07, + "loss": 0.9373, + "step": 477 + }, + { + "epoch": 0.8851851851851852, + "grad_norm": 1.8778776325197515, + "learning_rate": 1.4753086419753085e-07, + "loss": 0.8686, + "step": 478 + }, + { + "epoch": 0.8870370370370371, + "grad_norm": 2.1537192152388185, + "learning_rate": 1.478395061728395e-07, + "loss": 0.8472, + "step": 479 + }, + { + "epoch": 0.8888888888888888, + "grad_norm": 1.8763174725335232, + "learning_rate": 1.4814814814814815e-07, + "loss": 0.809, + "step": 480 + }, + { + "epoch": 0.8907407407407407, + "grad_norm": 1.8861836499017712, + "learning_rate": 1.4845679012345677e-07, + "loss": 0.8237, + "step": 481 + }, + { + "epoch": 0.8925925925925926, + "grad_norm": 1.8513811134477731, + "learning_rate": 1.4876543209876542e-07, + "loss": 0.9212, + "step": 482 + }, + { + "epoch": 0.8944444444444445, + "grad_norm": 1.7768570889746993, + "learning_rate": 1.4907407407407407e-07, + "loss": 0.8951, + "step": 483 + }, + { + "epoch": 0.8962962962962963, + "grad_norm": 1.7852128710682684, + "learning_rate": 1.4938271604938272e-07, + "loss": 0.9029, + "step": 484 + }, + { + "epoch": 0.8981481481481481, + "grad_norm": 3.4913964921006713, + "learning_rate": 1.4969135802469134e-07, + "loss": 0.9091, + "step": 485 + }, + { + "epoch": 0.9, + "grad_norm": 2.3549257572774924, + "learning_rate": 1.5e-07, + "loss": 0.8579, + "step": 486 + }, + { + "epoch": 0.9018518518518519, + "grad_norm": 1.7330307330370882, + "learning_rate": 1.5030864197530864e-07, + "loss": 0.9276, + "step": 487 + }, + { + "epoch": 0.9037037037037037, + "grad_norm": 1.8259607839651515, + "learning_rate": 1.5061728395061727e-07, + "loss": 0.891, + "step": 488 + }, + { + "epoch": 0.9055555555555556, + "grad_norm": 1.7047001942211497, + "learning_rate": 1.5092592592592592e-07, + "loss": 0.8804, + "step": 489 + }, + { + "epoch": 0.9074074074074074, + "grad_norm": 1.9490305282427514, + "learning_rate": 1.5123456790123457e-07, + "loss": 0.8768, + "step": 490 + }, + { + "epoch": 0.9092592592592592, + "grad_norm": 1.658719915297525, + "learning_rate": 1.5154320987654322e-07, + "loss": 0.858, + "step": 491 + }, + { + "epoch": 0.9111111111111111, + "grad_norm": 1.9402965103571035, + "learning_rate": 1.5185185185185184e-07, + "loss": 0.8649, + "step": 492 + }, + { + "epoch": 0.912962962962963, + "grad_norm": 1.718571810022627, + "learning_rate": 1.521604938271605e-07, + "loss": 0.8937, + "step": 493 + }, + { + "epoch": 0.9148148148148149, + "grad_norm": 1.6297690955124837, + "learning_rate": 1.5246913580246914e-07, + "loss": 0.828, + "step": 494 + }, + { + "epoch": 0.9166666666666666, + "grad_norm": 1.8414582414718768, + "learning_rate": 1.5277777777777776e-07, + "loss": 0.9126, + "step": 495 + }, + { + "epoch": 0.9185185185185185, + "grad_norm": 1.7802538734709807, + "learning_rate": 1.5308641975308642e-07, + "loss": 0.9056, + "step": 496 + }, + { + "epoch": 0.9203703703703704, + "grad_norm": 5.228694141231115, + "learning_rate": 1.5339506172839504e-07, + "loss": 0.8687, + "step": 497 + }, + { + "epoch": 0.9222222222222223, + "grad_norm": 1.8464622991838917, + "learning_rate": 1.537037037037037e-07, + "loss": 0.9016, + "step": 498 + }, + { + "epoch": 0.924074074074074, + "grad_norm": 2.4297371592733694, + "learning_rate": 1.5401234567901234e-07, + "loss": 0.8035, + "step": 499 + }, + { + "epoch": 0.9259259259259259, + "grad_norm": 2.0737176515541664, + "learning_rate": 1.54320987654321e-07, + "loss": 0.898, + "step": 500 + }, + { + "epoch": 0.9277777777777778, + "grad_norm": 1.8183579167972654, + "learning_rate": 1.546296296296296e-07, + "loss": 0.8512, + "step": 501 + }, + { + "epoch": 0.9296296296296296, + "grad_norm": 1.816892174063631, + "learning_rate": 1.5493827160493826e-07, + "loss": 0.8708, + "step": 502 + }, + { + "epoch": 0.9314814814814815, + "grad_norm": 1.6465826736383782, + "learning_rate": 1.552469135802469e-07, + "loss": 0.8607, + "step": 503 + }, + { + "epoch": 0.9333333333333333, + "grad_norm": 2.2704556478446714, + "learning_rate": 1.5555555555555556e-07, + "loss": 0.8925, + "step": 504 + }, + { + "epoch": 0.9351851851851852, + "grad_norm": 1.7467114879223669, + "learning_rate": 1.558641975308642e-07, + "loss": 0.8675, + "step": 505 + }, + { + "epoch": 0.937037037037037, + "grad_norm": 1.6062240697663543, + "learning_rate": 1.5617283950617284e-07, + "loss": 0.7859, + "step": 506 + }, + { + "epoch": 0.9388888888888889, + "grad_norm": 1.640348857593773, + "learning_rate": 1.564814814814815e-07, + "loss": 0.8543, + "step": 507 + }, + { + "epoch": 0.9407407407407408, + "grad_norm": 1.550816377843016, + "learning_rate": 1.5679012345679014e-07, + "loss": 0.8109, + "step": 508 + }, + { + "epoch": 0.9425925925925925, + "grad_norm": 2.0305085131556058, + "learning_rate": 1.5709876543209876e-07, + "loss": 0.8309, + "step": 509 + }, + { + "epoch": 0.9444444444444444, + "grad_norm": 1.568766105297353, + "learning_rate": 1.574074074074074e-07, + "loss": 0.8161, + "step": 510 + }, + { + "epoch": 0.9462962962962963, + "grad_norm": 1.7340153277737653, + "learning_rate": 1.5771604938271603e-07, + "loss": 0.9177, + "step": 511 + }, + { + "epoch": 0.9481481481481482, + "grad_norm": 1.885988879145682, + "learning_rate": 1.5802469135802468e-07, + "loss": 0.8394, + "step": 512 + }, + { + "epoch": 0.95, + "grad_norm": 2.29569659414859, + "learning_rate": 1.583333333333333e-07, + "loss": 0.8561, + "step": 513 + }, + { + "epoch": 0.9518518518518518, + "grad_norm": 1.7335115853805168, + "learning_rate": 1.5864197530864196e-07, + "loss": 0.9011, + "step": 514 + }, + { + "epoch": 0.9537037037037037, + "grad_norm": 1.675001315928241, + "learning_rate": 1.589506172839506e-07, + "loss": 0.858, + "step": 515 + }, + { + "epoch": 0.9555555555555556, + "grad_norm": 1.9824008234504722, + "learning_rate": 1.5925925925925926e-07, + "loss": 0.8752, + "step": 516 + }, + { + "epoch": 0.9574074074074074, + "grad_norm": 1.5903658251420252, + "learning_rate": 1.5956790123456788e-07, + "loss": 0.8516, + "step": 517 + }, + { + "epoch": 0.9592592592592593, + "grad_norm": 3.004633637313183, + "learning_rate": 1.5987654320987653e-07, + "loss": 0.8521, + "step": 518 + }, + { + "epoch": 0.9611111111111111, + "grad_norm": 2.5953683287541898, + "learning_rate": 1.6018518518518518e-07, + "loss": 0.8415, + "step": 519 + }, + { + "epoch": 0.9629629629629629, + "grad_norm": 1.6725515794786023, + "learning_rate": 1.6049382716049383e-07, + "loss": 0.8566, + "step": 520 + }, + { + "epoch": 0.9648148148148148, + "grad_norm": 3.2869740744171803, + "learning_rate": 1.6080246913580246e-07, + "loss": 0.8579, + "step": 521 + }, + { + "epoch": 0.9666666666666667, + "grad_norm": 1.6233912867628748, + "learning_rate": 1.611111111111111e-07, + "loss": 0.8437, + "step": 522 + }, + { + "epoch": 0.9685185185185186, + "grad_norm": 1.6642946681133515, + "learning_rate": 1.6141975308641976e-07, + "loss": 0.8328, + "step": 523 + }, + { + "epoch": 0.9703703703703703, + "grad_norm": 1.6643806957389826, + "learning_rate": 1.617283950617284e-07, + "loss": 0.8804, + "step": 524 + }, + { + "epoch": 0.9722222222222222, + "grad_norm": 1.6615537227064379, + "learning_rate": 1.6203703703703703e-07, + "loss": 0.8534, + "step": 525 + }, + { + "epoch": 0.9740740740740741, + "grad_norm": 1.574154420468767, + "learning_rate": 1.6234567901234568e-07, + "loss": 0.8414, + "step": 526 + }, + { + "epoch": 0.975925925925926, + "grad_norm": 2.2283760433888857, + "learning_rate": 1.626543209876543e-07, + "loss": 0.8231, + "step": 527 + }, + { + "epoch": 0.9777777777777777, + "grad_norm": 1.7037218290467242, + "learning_rate": 1.6296296296296295e-07, + "loss": 0.8769, + "step": 528 + }, + { + "epoch": 0.9796296296296296, + "grad_norm": 2.754981651435544, + "learning_rate": 1.6327160493827158e-07, + "loss": 0.8094, + "step": 529 + }, + { + "epoch": 0.9814814814814815, + "grad_norm": 3.3500242013818915, + "learning_rate": 1.6358024691358023e-07, + "loss": 0.8842, + "step": 530 + }, + { + "epoch": 0.9833333333333333, + "grad_norm": 1.6897837295849563, + "learning_rate": 1.6388888888888888e-07, + "loss": 0.8687, + "step": 531 + }, + { + "epoch": 0.9851851851851852, + "grad_norm": 1.541496447171207, + "learning_rate": 1.6419753086419753e-07, + "loss": 0.8503, + "step": 532 + }, + { + "epoch": 0.987037037037037, + "grad_norm": 1.7026082175637889, + "learning_rate": 1.6450617283950615e-07, + "loss": 0.8877, + "step": 533 + }, + { + "epoch": 0.9888888888888889, + "grad_norm": 1.6431224642792341, + "learning_rate": 1.648148148148148e-07, + "loss": 0.8548, + "step": 534 + }, + { + "epoch": 0.9907407407407407, + "grad_norm": 2.1178964677350605, + "learning_rate": 1.6512345679012345e-07, + "loss": 0.8484, + "step": 535 + }, + { + "epoch": 0.9925925925925926, + "grad_norm": 3.250841692000507, + "learning_rate": 1.654320987654321e-07, + "loss": 0.8962, + "step": 536 + }, + { + "epoch": 0.9944444444444445, + "grad_norm": 1.66266098194129, + "learning_rate": 1.6574074074074073e-07, + "loss": 0.8529, + "step": 537 + }, + { + "epoch": 0.9962962962962963, + "grad_norm": 1.5252470823405428, + "learning_rate": 1.6604938271604938e-07, + "loss": 0.8466, + "step": 538 + }, + { + "epoch": 0.9981481481481481, + "grad_norm": 2.9922623261242984, + "learning_rate": 1.6635802469135803e-07, + "loss": 0.758, + "step": 539 + }, + { + "epoch": 1.0, + "grad_norm": 1.934430205759249, + "learning_rate": 1.6666666666666668e-07, + "loss": 0.8866, + "step": 540 + }, + { + "epoch": 1.0018518518518518, + "grad_norm": 1.8883347084868174, + "learning_rate": 1.669753086419753e-07, + "loss": 0.8547, + "step": 541 + }, + { + "epoch": 1.0037037037037038, + "grad_norm": 1.5748500720605405, + "learning_rate": 1.6728395061728395e-07, + "loss": 0.87, + "step": 542 + }, + { + "epoch": 1.0055555555555555, + "grad_norm": 1.560708327560299, + "learning_rate": 1.6759259259259257e-07, + "loss": 0.8182, + "step": 543 + }, + { + "epoch": 1.0074074074074073, + "grad_norm": 2.0051977214364407, + "learning_rate": 1.6790123456790122e-07, + "loss": 0.8718, + "step": 544 + }, + { + "epoch": 1.0092592592592593, + "grad_norm": 1.6191423023001053, + "learning_rate": 1.6820987654320985e-07, + "loss": 0.8248, + "step": 545 + }, + { + "epoch": 1.011111111111111, + "grad_norm": 1.5051175720206331, + "learning_rate": 1.685185185185185e-07, + "loss": 0.8916, + "step": 546 + }, + { + "epoch": 1.012962962962963, + "grad_norm": 2.9158430296966045, + "learning_rate": 1.6882716049382715e-07, + "loss": 0.8395, + "step": 547 + }, + { + "epoch": 1.0148148148148148, + "grad_norm": 1.4991342659178246, + "learning_rate": 1.691358024691358e-07, + "loss": 0.8426, + "step": 548 + }, + { + "epoch": 1.0166666666666666, + "grad_norm": 1.6995730111540799, + "learning_rate": 1.6944444444444442e-07, + "loss": 0.8478, + "step": 549 + }, + { + "epoch": 1.0185185185185186, + "grad_norm": 1.4665838725695703, + "learning_rate": 1.6975308641975307e-07, + "loss": 0.8478, + "step": 550 + }, + { + "epoch": 1.0203703703703704, + "grad_norm": 1.5934504149694464, + "learning_rate": 1.7006172839506172e-07, + "loss": 0.8472, + "step": 551 + }, + { + "epoch": 1.0222222222222221, + "grad_norm": 1.5776774920082102, + "learning_rate": 1.7037037037037037e-07, + "loss": 0.8501, + "step": 552 + }, + { + "epoch": 1.0240740740740741, + "grad_norm": 1.8108062171862294, + "learning_rate": 1.70679012345679e-07, + "loss": 0.7989, + "step": 553 + }, + { + "epoch": 1.025925925925926, + "grad_norm": 1.5609766897230202, + "learning_rate": 1.7098765432098765e-07, + "loss": 0.8796, + "step": 554 + }, + { + "epoch": 1.0277777777777777, + "grad_norm": 1.5421333063522007, + "learning_rate": 1.712962962962963e-07, + "loss": 0.8144, + "step": 555 + }, + { + "epoch": 1.0296296296296297, + "grad_norm": 1.5906872588547103, + "learning_rate": 1.7160493827160495e-07, + "loss": 0.8231, + "step": 556 + }, + { + "epoch": 1.0314814814814814, + "grad_norm": 1.55563369873077, + "learning_rate": 1.7191358024691357e-07, + "loss": 0.8441, + "step": 557 + }, + { + "epoch": 1.0333333333333334, + "grad_norm": 2.034442281969645, + "learning_rate": 1.7222222222222222e-07, + "loss": 0.8555, + "step": 558 + }, + { + "epoch": 1.0351851851851852, + "grad_norm": 1.6802229628138714, + "learning_rate": 1.7253086419753084e-07, + "loss": 0.8632, + "step": 559 + }, + { + "epoch": 1.037037037037037, + "grad_norm": 1.6434079268881443, + "learning_rate": 1.728395061728395e-07, + "loss": 0.9062, + "step": 560 + }, + { + "epoch": 1.038888888888889, + "grad_norm": 10.170009592221364, + "learning_rate": 1.7314814814814814e-07, + "loss": 0.8269, + "step": 561 + }, + { + "epoch": 1.0407407407407407, + "grad_norm": 1.5129567429071522, + "learning_rate": 1.7345679012345677e-07, + "loss": 0.7992, + "step": 562 + }, + { + "epoch": 1.0425925925925925, + "grad_norm": 1.4693189752768574, + "learning_rate": 1.7376543209876542e-07, + "loss": 0.8625, + "step": 563 + }, + { + "epoch": 1.0444444444444445, + "grad_norm": 1.4837637561015875, + "learning_rate": 1.7407407407407407e-07, + "loss": 0.8337, + "step": 564 + }, + { + "epoch": 1.0462962962962963, + "grad_norm": 1.4800586250040941, + "learning_rate": 1.7438271604938272e-07, + "loss": 0.8073, + "step": 565 + }, + { + "epoch": 1.048148148148148, + "grad_norm": 1.5010334413639592, + "learning_rate": 1.7469135802469134e-07, + "loss": 0.8723, + "step": 566 + }, + { + "epoch": 1.05, + "grad_norm": 1.5246163461303288, + "learning_rate": 1.75e-07, + "loss": 0.9139, + "step": 567 + }, + { + "epoch": 1.0518518518518518, + "grad_norm": 1.5096409785175586, + "learning_rate": 1.7530864197530864e-07, + "loss": 0.7871, + "step": 568 + }, + { + "epoch": 1.0537037037037038, + "grad_norm": 1.7033015989867157, + "learning_rate": 1.7561728395061727e-07, + "loss": 0.8503, + "step": 569 + }, + { + "epoch": 1.0555555555555556, + "grad_norm": 1.8331304797783068, + "learning_rate": 1.7592592592592592e-07, + "loss": 0.8303, + "step": 570 + }, + { + "epoch": 1.0574074074074074, + "grad_norm": 2.657686596984486, + "learning_rate": 1.7623456790123457e-07, + "loss": 0.8945, + "step": 571 + }, + { + "epoch": 1.0592592592592593, + "grad_norm": 1.4015798649304902, + "learning_rate": 1.7654320987654322e-07, + "loss": 0.7581, + "step": 572 + }, + { + "epoch": 1.0611111111111111, + "grad_norm": 2.6653805849934242, + "learning_rate": 1.7685185185185184e-07, + "loss": 0.8767, + "step": 573 + }, + { + "epoch": 1.0629629629629629, + "grad_norm": 1.5964624363478677, + "learning_rate": 1.771604938271605e-07, + "loss": 0.8748, + "step": 574 + }, + { + "epoch": 1.0648148148148149, + "grad_norm": 1.656722307672164, + "learning_rate": 1.7746913580246914e-07, + "loss": 0.8251, + "step": 575 + }, + { + "epoch": 1.0666666666666667, + "grad_norm": 3.2261514557816664, + "learning_rate": 1.7777777777777776e-07, + "loss": 0.7857, + "step": 576 + }, + { + "epoch": 1.0685185185185184, + "grad_norm": 1.4007949119888583, + "learning_rate": 1.7808641975308641e-07, + "loss": 0.7776, + "step": 577 + }, + { + "epoch": 1.0703703703703704, + "grad_norm": 1.6083368995055551, + "learning_rate": 1.7839506172839504e-07, + "loss": 0.915, + "step": 578 + }, + { + "epoch": 1.0722222222222222, + "grad_norm": 1.6301023545495026, + "learning_rate": 1.787037037037037e-07, + "loss": 0.8663, + "step": 579 + }, + { + "epoch": 1.074074074074074, + "grad_norm": 1.5182789492763844, + "learning_rate": 1.7901234567901234e-07, + "loss": 0.832, + "step": 580 + }, + { + "epoch": 1.075925925925926, + "grad_norm": 1.7588871797491485, + "learning_rate": 1.79320987654321e-07, + "loss": 0.8315, + "step": 581 + }, + { + "epoch": 1.0777777777777777, + "grad_norm": 1.529148150776432, + "learning_rate": 1.796296296296296e-07, + "loss": 0.8357, + "step": 582 + }, + { + "epoch": 1.0796296296296297, + "grad_norm": 1.4254554207617063, + "learning_rate": 1.7993827160493826e-07, + "loss": 0.8107, + "step": 583 + }, + { + "epoch": 1.0814814814814815, + "grad_norm": 1.5027242983018325, + "learning_rate": 1.802469135802469e-07, + "loss": 0.8475, + "step": 584 + }, + { + "epoch": 1.0833333333333333, + "grad_norm": 1.4657323461202916, + "learning_rate": 1.8055555555555556e-07, + "loss": 0.8119, + "step": 585 + }, + { + "epoch": 1.0851851851851853, + "grad_norm": 1.4079806350705697, + "learning_rate": 1.8086419753086419e-07, + "loss": 0.8334, + "step": 586 + }, + { + "epoch": 1.087037037037037, + "grad_norm": 1.6242605991898924, + "learning_rate": 1.8117283950617284e-07, + "loss": 0.7631, + "step": 587 + }, + { + "epoch": 1.0888888888888888, + "grad_norm": 1.8769687310635081, + "learning_rate": 1.8148148148148149e-07, + "loss": 0.7859, + "step": 588 + }, + { + "epoch": 1.0907407407407408, + "grad_norm": 1.934757438979972, + "learning_rate": 1.8179012345679014e-07, + "loss": 0.8405, + "step": 589 + }, + { + "epoch": 1.0925925925925926, + "grad_norm": 1.5275657148498663, + "learning_rate": 1.8209876543209876e-07, + "loss": 0.8969, + "step": 590 + }, + { + "epoch": 1.0944444444444446, + "grad_norm": 1.535143711770453, + "learning_rate": 1.824074074074074e-07, + "loss": 0.7807, + "step": 591 + }, + { + "epoch": 1.0962962962962963, + "grad_norm": 1.7247098676492747, + "learning_rate": 1.8271604938271603e-07, + "loss": 0.8063, + "step": 592 + }, + { + "epoch": 1.098148148148148, + "grad_norm": 1.8582360291923699, + "learning_rate": 1.8302469135802468e-07, + "loss": 0.8432, + "step": 593 + }, + { + "epoch": 1.1, + "grad_norm": 1.4459341873170923, + "learning_rate": 1.833333333333333e-07, + "loss": 0.8294, + "step": 594 + }, + { + "epoch": 1.1018518518518519, + "grad_norm": 1.6402085048812947, + "learning_rate": 1.8364197530864196e-07, + "loss": 0.8303, + "step": 595 + }, + { + "epoch": 1.1037037037037036, + "grad_norm": 1.5383980751937993, + "learning_rate": 1.839506172839506e-07, + "loss": 0.843, + "step": 596 + }, + { + "epoch": 1.1055555555555556, + "grad_norm": 1.816409667660388, + "learning_rate": 1.8425925925925926e-07, + "loss": 0.8793, + "step": 597 + }, + { + "epoch": 1.1074074074074074, + "grad_norm": 1.6956364381623301, + "learning_rate": 1.8456790123456788e-07, + "loss": 0.9443, + "step": 598 + }, + { + "epoch": 1.1092592592592592, + "grad_norm": 2.328299425083654, + "learning_rate": 1.8487654320987653e-07, + "loss": 0.8211, + "step": 599 + }, + { + "epoch": 1.1111111111111112, + "grad_norm": 1.6009889909653843, + "learning_rate": 1.8518518518518518e-07, + "loss": 0.8219, + "step": 600 + }, + { + "epoch": 1.112962962962963, + "grad_norm": 1.7262702252944861, + "learning_rate": 1.8549382716049383e-07, + "loss": 0.8141, + "step": 601 + }, + { + "epoch": 1.1148148148148147, + "grad_norm": 2.175237774059329, + "learning_rate": 1.8580246913580246e-07, + "loss": 0.8183, + "step": 602 + }, + { + "epoch": 1.1166666666666667, + "grad_norm": 1.6705397958700674, + "learning_rate": 1.861111111111111e-07, + "loss": 0.8491, + "step": 603 + }, + { + "epoch": 1.1185185185185185, + "grad_norm": 1.4166684807479268, + "learning_rate": 1.8641975308641976e-07, + "loss": 0.7582, + "step": 604 + }, + { + "epoch": 1.1203703703703705, + "grad_norm": 4.789621688237432, + "learning_rate": 1.867283950617284e-07, + "loss": 0.8323, + "step": 605 + }, + { + "epoch": 1.1222222222222222, + "grad_norm": 1.542870095454017, + "learning_rate": 1.8703703703703703e-07, + "loss": 0.82, + "step": 606 + }, + { + "epoch": 1.124074074074074, + "grad_norm": 1.4251956480393917, + "learning_rate": 1.8734567901234568e-07, + "loss": 0.8415, + "step": 607 + }, + { + "epoch": 1.125925925925926, + "grad_norm": 1.6725642490505486, + "learning_rate": 1.876543209876543e-07, + "loss": 0.8563, + "step": 608 + }, + { + "epoch": 1.1277777777777778, + "grad_norm": 1.8001724023147403, + "learning_rate": 1.8796296296296295e-07, + "loss": 0.7856, + "step": 609 + }, + { + "epoch": 1.1296296296296295, + "grad_norm": 1.5777165127139179, + "learning_rate": 1.8827160493827158e-07, + "loss": 0.8428, + "step": 610 + }, + { + "epoch": 1.1314814814814815, + "grad_norm": 1.5761579111337034, + "learning_rate": 1.8858024691358023e-07, + "loss": 0.8513, + "step": 611 + }, + { + "epoch": 1.1333333333333333, + "grad_norm": 1.8242570335542432, + "learning_rate": 1.8888888888888888e-07, + "loss": 0.8156, + "step": 612 + }, + { + "epoch": 1.1351851851851853, + "grad_norm": 1.530691819395211, + "learning_rate": 1.8919753086419753e-07, + "loss": 0.7881, + "step": 613 + }, + { + "epoch": 1.137037037037037, + "grad_norm": 1.9164515154073438, + "learning_rate": 1.8950617283950615e-07, + "loss": 0.8257, + "step": 614 + }, + { + "epoch": 1.1388888888888888, + "grad_norm": 1.4929066752543092, + "learning_rate": 1.898148148148148e-07, + "loss": 0.8359, + "step": 615 + }, + { + "epoch": 1.1407407407407408, + "grad_norm": 1.4081027554450505, + "learning_rate": 1.9012345679012345e-07, + "loss": 0.8767, + "step": 616 + }, + { + "epoch": 1.1425925925925926, + "grad_norm": 1.5770738488789227, + "learning_rate": 1.904320987654321e-07, + "loss": 0.8134, + "step": 617 + }, + { + "epoch": 1.1444444444444444, + "grad_norm": 1.7184457454114574, + "learning_rate": 1.9074074074074073e-07, + "loss": 0.8596, + "step": 618 + }, + { + "epoch": 1.1462962962962964, + "grad_norm": 1.4055190812277196, + "learning_rate": 1.9104938271604938e-07, + "loss": 0.823, + "step": 619 + }, + { + "epoch": 1.1481481481481481, + "grad_norm": 1.348126776083916, + "learning_rate": 1.9135802469135803e-07, + "loss": 0.8183, + "step": 620 + }, + { + "epoch": 1.15, + "grad_norm": 1.5088465269448812, + "learning_rate": 1.9166666666666668e-07, + "loss": 0.7821, + "step": 621 + }, + { + "epoch": 1.151851851851852, + "grad_norm": 1.7632370443667849, + "learning_rate": 1.919753086419753e-07, + "loss": 0.84, + "step": 622 + }, + { + "epoch": 1.1537037037037037, + "grad_norm": 1.9255514334517234, + "learning_rate": 1.9228395061728395e-07, + "loss": 0.7746, + "step": 623 + }, + { + "epoch": 1.1555555555555554, + "grad_norm": 1.6286763962754056, + "learning_rate": 1.9259259259259257e-07, + "loss": 0.8576, + "step": 624 + }, + { + "epoch": 1.1574074074074074, + "grad_norm": 1.6633722663137955, + "learning_rate": 1.9290123456790122e-07, + "loss": 0.8155, + "step": 625 + }, + { + "epoch": 1.1592592592592592, + "grad_norm": 1.4192426771058053, + "learning_rate": 1.9320987654320985e-07, + "loss": 0.8083, + "step": 626 + }, + { + "epoch": 1.1611111111111112, + "grad_norm": 1.4440661199091922, + "learning_rate": 1.935185185185185e-07, + "loss": 0.8555, + "step": 627 + }, + { + "epoch": 1.162962962962963, + "grad_norm": 1.6233104448677302, + "learning_rate": 1.9382716049382715e-07, + "loss": 0.8894, + "step": 628 + }, + { + "epoch": 1.1648148148148147, + "grad_norm": 1.4165073974240723, + "learning_rate": 1.941358024691358e-07, + "loss": 0.8253, + "step": 629 + }, + { + "epoch": 1.1666666666666667, + "grad_norm": 1.628051655179612, + "learning_rate": 1.9444444444444442e-07, + "loss": 0.8778, + "step": 630 + }, + { + "epoch": 1.1685185185185185, + "grad_norm": 1.5499826920909507, + "learning_rate": 1.9475308641975307e-07, + "loss": 0.8489, + "step": 631 + }, + { + "epoch": 1.1703703703703703, + "grad_norm": 1.962979237347802, + "learning_rate": 1.9506172839506172e-07, + "loss": 0.7925, + "step": 632 + }, + { + "epoch": 1.1722222222222223, + "grad_norm": 1.7591379566308454, + "learning_rate": 1.9537037037037037e-07, + "loss": 0.8667, + "step": 633 + }, + { + "epoch": 1.174074074074074, + "grad_norm": 2.460868892243051, + "learning_rate": 1.95679012345679e-07, + "loss": 0.8494, + "step": 634 + }, + { + "epoch": 1.175925925925926, + "grad_norm": 1.8877059014847026, + "learning_rate": 1.9598765432098765e-07, + "loss": 0.8017, + "step": 635 + }, + { + "epoch": 1.1777777777777778, + "grad_norm": 1.7045657591829686, + "learning_rate": 1.962962962962963e-07, + "loss": 0.8192, + "step": 636 + }, + { + "epoch": 1.1796296296296296, + "grad_norm": 1.5452640381715912, + "learning_rate": 1.9660493827160495e-07, + "loss": 0.7966, + "step": 637 + }, + { + "epoch": 1.1814814814814816, + "grad_norm": 1.732657264409647, + "learning_rate": 1.9691358024691357e-07, + "loss": 0.8075, + "step": 638 + }, + { + "epoch": 1.1833333333333333, + "grad_norm": 1.4526087769238891, + "learning_rate": 1.9722222222222222e-07, + "loss": 0.8596, + "step": 639 + }, + { + "epoch": 1.1851851851851851, + "grad_norm": 1.601317179763928, + "learning_rate": 1.9753086419753084e-07, + "loss": 0.7948, + "step": 640 + }, + { + "epoch": 1.1870370370370371, + "grad_norm": 1.5937199109006792, + "learning_rate": 1.978395061728395e-07, + "loss": 0.8487, + "step": 641 + }, + { + "epoch": 1.1888888888888889, + "grad_norm": 1.418344467800995, + "learning_rate": 1.9814814814814814e-07, + "loss": 0.8215, + "step": 642 + }, + { + "epoch": 1.1907407407407407, + "grad_norm": 1.3326736296022426, + "learning_rate": 1.9845679012345677e-07, + "loss": 0.7929, + "step": 643 + }, + { + "epoch": 1.1925925925925926, + "grad_norm": 1.8151537792443233, + "learning_rate": 1.9876543209876542e-07, + "loss": 0.8282, + "step": 644 + }, + { + "epoch": 1.1944444444444444, + "grad_norm": 1.6758174444592453, + "learning_rate": 1.9907407407407407e-07, + "loss": 0.769, + "step": 645 + }, + { + "epoch": 1.1962962962962962, + "grad_norm": 1.4288624957309835, + "learning_rate": 1.9938271604938272e-07, + "loss": 0.812, + "step": 646 + }, + { + "epoch": 1.1981481481481482, + "grad_norm": 2.2318233017324633, + "learning_rate": 1.9969135802469134e-07, + "loss": 0.7757, + "step": 647 + }, + { + "epoch": 1.2, + "grad_norm": 1.673189714917265, + "learning_rate": 2e-07, + "loss": 0.8142, + "step": 648 + }, + { + "epoch": 1.201851851851852, + "grad_norm": 1.81829191197008, + "learning_rate": 1.9999999887586419e-07, + "loss": 0.7731, + "step": 649 + }, + { + "epoch": 1.2037037037037037, + "grad_norm": 1.676121940820495, + "learning_rate": 1.999999955034568e-07, + "loss": 0.827, + "step": 650 + }, + { + "epoch": 1.2055555555555555, + "grad_norm": 1.3725717260554386, + "learning_rate": 1.9999998988277795e-07, + "loss": 0.7714, + "step": 651 + }, + { + "epoch": 1.2074074074074075, + "grad_norm": 1.5384999249290476, + "learning_rate": 1.9999998201382774e-07, + "loss": 0.8222, + "step": 652 + }, + { + "epoch": 1.2092592592592593, + "grad_norm": 1.3696578068453862, + "learning_rate": 1.9999997189660628e-07, + "loss": 0.8165, + "step": 653 + }, + { + "epoch": 1.211111111111111, + "grad_norm": 1.855368937170554, + "learning_rate": 1.9999995953111393e-07, + "loss": 0.8613, + "step": 654 + }, + { + "epoch": 1.212962962962963, + "grad_norm": 1.2370747321595208, + "learning_rate": 1.9999994491735084e-07, + "loss": 0.768, + "step": 655 + }, + { + "epoch": 1.2148148148148148, + "grad_norm": 1.5050274168084548, + "learning_rate": 1.9999992805531741e-07, + "loss": 0.796, + "step": 656 + }, + { + "epoch": 1.2166666666666668, + "grad_norm": 1.3984355995590978, + "learning_rate": 1.99999908945014e-07, + "loss": 0.8355, + "step": 657 + }, + { + "epoch": 1.2185185185185186, + "grad_norm": 2.0841481858806574, + "learning_rate": 1.9999988758644104e-07, + "loss": 0.8257, + "step": 658 + }, + { + "epoch": 1.2203703703703703, + "grad_norm": 1.280201023265319, + "learning_rate": 1.9999986397959903e-07, + "loss": 0.8096, + "step": 659 + }, + { + "epoch": 1.2222222222222223, + "grad_norm": 1.4407652952391021, + "learning_rate": 1.9999983812448847e-07, + "loss": 0.8393, + "step": 660 + }, + { + "epoch": 1.224074074074074, + "grad_norm": 4.297211053403042, + "learning_rate": 1.9999981002110996e-07, + "loss": 0.8565, + "step": 661 + }, + { + "epoch": 1.2259259259259259, + "grad_norm": 1.3281484149102734, + "learning_rate": 1.999997796694641e-07, + "loss": 0.8095, + "step": 662 + }, + { + "epoch": 1.2277777777777779, + "grad_norm": 1.4225231105051765, + "learning_rate": 1.999997470695516e-07, + "loss": 0.8483, + "step": 663 + }, + { + "epoch": 1.2296296296296296, + "grad_norm": 2.041529020910693, + "learning_rate": 1.999997122213732e-07, + "loss": 0.8252, + "step": 664 + }, + { + "epoch": 1.2314814814814814, + "grad_norm": 1.3320722685618107, + "learning_rate": 1.9999967512492969e-07, + "loss": 0.8237, + "step": 665 + }, + { + "epoch": 1.2333333333333334, + "grad_norm": 1.6871437488594683, + "learning_rate": 1.9999963578022186e-07, + "loss": 0.851, + "step": 666 + }, + { + "epoch": 1.2351851851851852, + "grad_norm": 1.5737732007069218, + "learning_rate": 1.9999959418725066e-07, + "loss": 0.8535, + "step": 667 + }, + { + "epoch": 1.237037037037037, + "grad_norm": 2.0286472424418567, + "learning_rate": 1.9999955034601696e-07, + "loss": 0.8725, + "step": 668 + }, + { + "epoch": 1.238888888888889, + "grad_norm": 3.61800278877449, + "learning_rate": 1.9999950425652178e-07, + "loss": 0.7998, + "step": 669 + }, + { + "epoch": 1.2407407407407407, + "grad_norm": 2.8285737528671437, + "learning_rate": 1.9999945591876616e-07, + "loss": 0.834, + "step": 670 + }, + { + "epoch": 1.2425925925925925, + "grad_norm": 1.6329172129507732, + "learning_rate": 1.9999940533275115e-07, + "loss": 0.7917, + "step": 671 + }, + { + "epoch": 1.2444444444444445, + "grad_norm": 4.083465952931434, + "learning_rate": 1.9999935249847796e-07, + "loss": 0.8115, + "step": 672 + }, + { + "epoch": 1.2462962962962962, + "grad_norm": 1.5603826600388242, + "learning_rate": 1.999992974159477e-07, + "loss": 0.8261, + "step": 673 + }, + { + "epoch": 1.2481481481481482, + "grad_norm": 2.1610248517678112, + "learning_rate": 1.999992400851616e-07, + "loss": 0.8175, + "step": 674 + }, + { + "epoch": 1.25, + "grad_norm": 1.301137053762421, + "learning_rate": 1.9999918050612107e-07, + "loss": 0.7972, + "step": 675 + }, + { + "epoch": 1.2518518518518518, + "grad_norm": 1.4350718372391504, + "learning_rate": 1.999991186788273e-07, + "loss": 0.8086, + "step": 676 + }, + { + "epoch": 1.2537037037037038, + "grad_norm": 1.3996355983978452, + "learning_rate": 1.999990546032818e-07, + "loss": 0.8132, + "step": 677 + }, + { + "epoch": 1.2555555555555555, + "grad_norm": 2.046076241863195, + "learning_rate": 1.999989882794859e-07, + "loss": 0.825, + "step": 678 + }, + { + "epoch": 1.2574074074074075, + "grad_norm": 1.9174275837783201, + "learning_rate": 1.9999891970744122e-07, + "loss": 0.8017, + "step": 679 + }, + { + "epoch": 1.2592592592592593, + "grad_norm": 1.5740862949497096, + "learning_rate": 1.9999884888714916e-07, + "loss": 0.8454, + "step": 680 + }, + { + "epoch": 1.261111111111111, + "grad_norm": 1.573943685824295, + "learning_rate": 1.9999877581861147e-07, + "loss": 0.8092, + "step": 681 + }, + { + "epoch": 1.262962962962963, + "grad_norm": 1.858044964623011, + "learning_rate": 1.9999870050182963e-07, + "loss": 0.8688, + "step": 682 + }, + { + "epoch": 1.2648148148148148, + "grad_norm": 1.608219461922296, + "learning_rate": 1.9999862293680544e-07, + "loss": 0.8168, + "step": 683 + }, + { + "epoch": 1.2666666666666666, + "grad_norm": 1.3532272377834922, + "learning_rate": 1.9999854312354063e-07, + "loss": 0.825, + "step": 684 + }, + { + "epoch": 1.2685185185185186, + "grad_norm": 1.5895271404188862, + "learning_rate": 1.9999846106203698e-07, + "loss": 0.8429, + "step": 685 + }, + { + "epoch": 1.2703703703703704, + "grad_norm": 17.48759201834969, + "learning_rate": 1.999983767522963e-07, + "loss": 0.7783, + "step": 686 + }, + { + "epoch": 1.2722222222222221, + "grad_norm": 1.5828554984901237, + "learning_rate": 1.9999829019432056e-07, + "loss": 0.7948, + "step": 687 + }, + { + "epoch": 1.2740740740740741, + "grad_norm": 2.701786831731794, + "learning_rate": 1.9999820138811164e-07, + "loss": 0.767, + "step": 688 + }, + { + "epoch": 1.275925925925926, + "grad_norm": 1.5710332993008467, + "learning_rate": 1.9999811033367155e-07, + "loss": 0.8305, + "step": 689 + }, + { + "epoch": 1.2777777777777777, + "grad_norm": 1.6722669093862321, + "learning_rate": 1.9999801703100236e-07, + "loss": 0.8075, + "step": 690 + }, + { + "epoch": 1.2796296296296297, + "grad_norm": 1.3571600013004477, + "learning_rate": 1.9999792148010616e-07, + "loss": 0.785, + "step": 691 + }, + { + "epoch": 1.2814814814814814, + "grad_norm": 1.7584995967761679, + "learning_rate": 1.999978236809851e-07, + "loss": 0.8275, + "step": 692 + }, + { + "epoch": 1.2833333333333332, + "grad_norm": 1.6908576105463775, + "learning_rate": 1.9999772363364138e-07, + "loss": 0.7764, + "step": 693 + }, + { + "epoch": 1.2851851851851852, + "grad_norm": 1.8336220147272548, + "learning_rate": 1.9999762133807723e-07, + "loss": 0.7871, + "step": 694 + }, + { + "epoch": 1.287037037037037, + "grad_norm": 1.5197665661563555, + "learning_rate": 1.9999751679429496e-07, + "loss": 0.8317, + "step": 695 + }, + { + "epoch": 1.2888888888888888, + "grad_norm": 1.392921843686158, + "learning_rate": 1.9999741000229693e-07, + "loss": 0.8176, + "step": 696 + }, + { + "epoch": 1.2907407407407407, + "grad_norm": 1.7142334810776394, + "learning_rate": 1.9999730096208553e-07, + "loss": 0.7566, + "step": 697 + }, + { + "epoch": 1.2925925925925925, + "grad_norm": 1.33257801620074, + "learning_rate": 1.999971896736632e-07, + "loss": 0.8202, + "step": 698 + }, + { + "epoch": 1.2944444444444445, + "grad_norm": 1.357846537448651, + "learning_rate": 1.9999707613703246e-07, + "loss": 0.8475, + "step": 699 + }, + { + "epoch": 1.2962962962962963, + "grad_norm": 2.933054090341049, + "learning_rate": 1.999969603521959e-07, + "loss": 0.8341, + "step": 700 + }, + { + "epoch": 1.2981481481481483, + "grad_norm": 1.346188543492007, + "learning_rate": 1.9999684231915606e-07, + "loss": 0.763, + "step": 701 + }, + { + "epoch": 1.3, + "grad_norm": 1.3967016006284687, + "learning_rate": 1.9999672203791561e-07, + "loss": 0.8189, + "step": 702 + }, + { + "epoch": 1.3018518518518518, + "grad_norm": 1.5955327686122693, + "learning_rate": 1.9999659950847728e-07, + "loss": 0.8343, + "step": 703 + }, + { + "epoch": 1.3037037037037038, + "grad_norm": 1.5976141423974135, + "learning_rate": 1.999964747308438e-07, + "loss": 0.8329, + "step": 704 + }, + { + "epoch": 1.3055555555555556, + "grad_norm": 1.4392796070897462, + "learning_rate": 1.9999634770501798e-07, + "loss": 0.7379, + "step": 705 + }, + { + "epoch": 1.3074074074074074, + "grad_norm": 1.584308070732416, + "learning_rate": 1.9999621843100267e-07, + "loss": 0.8115, + "step": 706 + }, + { + "epoch": 1.3092592592592593, + "grad_norm": 2.702347813899695, + "learning_rate": 1.999960869088008e-07, + "loss": 0.7978, + "step": 707 + }, + { + "epoch": 1.3111111111111111, + "grad_norm": 1.4954508281545915, + "learning_rate": 1.999959531384153e-07, + "loss": 0.8064, + "step": 708 + }, + { + "epoch": 1.3129629629629629, + "grad_norm": 1.7314897367970046, + "learning_rate": 1.9999581711984916e-07, + "loss": 0.8648, + "step": 709 + }, + { + "epoch": 1.3148148148148149, + "grad_norm": 1.8234726133694985, + "learning_rate": 1.9999567885310552e-07, + "loss": 0.7563, + "step": 710 + }, + { + "epoch": 1.3166666666666667, + "grad_norm": 2.0815322421777607, + "learning_rate": 1.999955383381874e-07, + "loss": 0.8034, + "step": 711 + }, + { + "epoch": 1.3185185185185184, + "grad_norm": 1.452416477738848, + "learning_rate": 1.99995395575098e-07, + "loss": 0.8275, + "step": 712 + }, + { + "epoch": 1.3203703703703704, + "grad_norm": 1.385547178875855, + "learning_rate": 1.9999525056384052e-07, + "loss": 0.7986, + "step": 713 + }, + { + "epoch": 1.3222222222222222, + "grad_norm": 1.6018152115518913, + "learning_rate": 1.9999510330441826e-07, + "loss": 0.7761, + "step": 714 + }, + { + "epoch": 1.324074074074074, + "grad_norm": 2.613760755025956, + "learning_rate": 1.9999495379683444e-07, + "loss": 0.7934, + "step": 715 + }, + { + "epoch": 1.325925925925926, + "grad_norm": 1.4548932715290424, + "learning_rate": 1.9999480204109251e-07, + "loss": 0.7822, + "step": 716 + }, + { + "epoch": 1.3277777777777777, + "grad_norm": 1.3911296980557761, + "learning_rate": 1.9999464803719587e-07, + "loss": 0.789, + "step": 717 + }, + { + "epoch": 1.3296296296296295, + "grad_norm": 1.383933108423946, + "learning_rate": 1.9999449178514792e-07, + "loss": 0.7718, + "step": 718 + }, + { + "epoch": 1.3314814814814815, + "grad_norm": 3.1847143238699864, + "learning_rate": 1.999943332849522e-07, + "loss": 0.814, + "step": 719 + }, + { + "epoch": 1.3333333333333333, + "grad_norm": 1.594065175286285, + "learning_rate": 1.9999417253661234e-07, + "loss": 0.825, + "step": 720 + }, + { + "epoch": 1.3351851851851853, + "grad_norm": 1.5185822785528786, + "learning_rate": 1.9999400954013188e-07, + "loss": 0.8408, + "step": 721 + }, + { + "epoch": 1.337037037037037, + "grad_norm": 1.6331028949400026, + "learning_rate": 1.999938442955145e-07, + "loss": 0.7784, + "step": 722 + }, + { + "epoch": 1.338888888888889, + "grad_norm": 1.4885723201648686, + "learning_rate": 1.999936768027639e-07, + "loss": 0.8108, + "step": 723 + }, + { + "epoch": 1.3407407407407408, + "grad_norm": 1.6253958747394246, + "learning_rate": 1.999935070618839e-07, + "loss": 0.7606, + "step": 724 + }, + { + "epoch": 1.3425925925925926, + "grad_norm": 2.100753153108363, + "learning_rate": 1.9999333507287827e-07, + "loss": 0.7585, + "step": 725 + }, + { + "epoch": 1.3444444444444446, + "grad_norm": 1.711567579429311, + "learning_rate": 1.999931608357509e-07, + "loss": 0.8569, + "step": 726 + }, + { + "epoch": 1.3462962962962963, + "grad_norm": 2.731792389961523, + "learning_rate": 1.9999298435050568e-07, + "loss": 0.8126, + "step": 727 + }, + { + "epoch": 1.348148148148148, + "grad_norm": 1.6302884387160579, + "learning_rate": 1.999928056171466e-07, + "loss": 0.7948, + "step": 728 + }, + { + "epoch": 1.35, + "grad_norm": 1.765748181345404, + "learning_rate": 1.999926246356777e-07, + "loss": 0.8708, + "step": 729 + }, + { + "epoch": 1.3518518518518519, + "grad_norm": 1.9756705535899186, + "learning_rate": 1.9999244140610298e-07, + "loss": 0.8435, + "step": 730 + }, + { + "epoch": 1.3537037037037036, + "grad_norm": 1.4228246423515123, + "learning_rate": 1.9999225592842665e-07, + "loss": 0.7985, + "step": 731 + }, + { + "epoch": 1.3555555555555556, + "grad_norm": 1.744488402830874, + "learning_rate": 1.999920682026528e-07, + "loss": 0.8642, + "step": 732 + }, + { + "epoch": 1.3574074074074074, + "grad_norm": 1.4106173918746343, + "learning_rate": 1.9999187822878568e-07, + "loss": 0.7934, + "step": 733 + }, + { + "epoch": 1.3592592592592592, + "grad_norm": 1.5603012222840908, + "learning_rate": 1.9999168600682958e-07, + "loss": 0.8443, + "step": 734 + }, + { + "epoch": 1.3611111111111112, + "grad_norm": 1.9684145208270543, + "learning_rate": 1.999914915367888e-07, + "loss": 0.6867, + "step": 735 + }, + { + "epoch": 1.362962962962963, + "grad_norm": 1.9078147431897725, + "learning_rate": 1.999912948186677e-07, + "loss": 0.835, + "step": 736 + }, + { + "epoch": 1.3648148148148147, + "grad_norm": 1.3041471586294773, + "learning_rate": 1.9999109585247075e-07, + "loss": 0.7823, + "step": 737 + }, + { + "epoch": 1.3666666666666667, + "grad_norm": 1.4188184952644083, + "learning_rate": 1.9999089463820237e-07, + "loss": 0.7751, + "step": 738 + }, + { + "epoch": 1.3685185185185185, + "grad_norm": 2.444346601690097, + "learning_rate": 1.9999069117586712e-07, + "loss": 0.8326, + "step": 739 + }, + { + "epoch": 1.3703703703703702, + "grad_norm": 1.3501257826270001, + "learning_rate": 1.9999048546546953e-07, + "loss": 0.8196, + "step": 740 + }, + { + "epoch": 1.3722222222222222, + "grad_norm": 2.3372204727965236, + "learning_rate": 1.9999027750701429e-07, + "loss": 0.8256, + "step": 741 + }, + { + "epoch": 1.374074074074074, + "grad_norm": 2.025133239373389, + "learning_rate": 1.9999006730050602e-07, + "loss": 0.7684, + "step": 742 + }, + { + "epoch": 1.375925925925926, + "grad_norm": 1.4560990570388062, + "learning_rate": 1.999898548459495e-07, + "loss": 0.8069, + "step": 743 + }, + { + "epoch": 1.3777777777777778, + "grad_norm": 1.587066567252244, + "learning_rate": 1.9998964014334944e-07, + "loss": 0.7644, + "step": 744 + }, + { + "epoch": 1.3796296296296298, + "grad_norm": 1.5048099316236232, + "learning_rate": 1.9998942319271075e-07, + "loss": 0.7822, + "step": 745 + }, + { + "epoch": 1.3814814814814815, + "grad_norm": 1.5456315308946729, + "learning_rate": 1.9998920399403822e-07, + "loss": 0.8821, + "step": 746 + }, + { + "epoch": 1.3833333333333333, + "grad_norm": 2.0098488030872033, + "learning_rate": 1.9998898254733683e-07, + "loss": 0.8151, + "step": 747 + }, + { + "epoch": 1.3851851851851853, + "grad_norm": 1.5106633218657948, + "learning_rate": 1.9998875885261156e-07, + "loss": 0.8001, + "step": 748 + }, + { + "epoch": 1.387037037037037, + "grad_norm": 1.4314783784907348, + "learning_rate": 1.999885329098674e-07, + "loss": 0.826, + "step": 749 + }, + { + "epoch": 1.3888888888888888, + "grad_norm": 1.7604892244147972, + "learning_rate": 1.9998830471910945e-07, + "loss": 0.8427, + "step": 750 + }, + { + "epoch": 1.3907407407407408, + "grad_norm": 1.4427224781548, + "learning_rate": 1.9998807428034285e-07, + "loss": 0.812, + "step": 751 + }, + { + "epoch": 1.3925925925925926, + "grad_norm": 1.3919650093002605, + "learning_rate": 1.999878415935728e-07, + "loss": 0.8232, + "step": 752 + }, + { + "epoch": 1.3944444444444444, + "grad_norm": 1.5123998153492724, + "learning_rate": 1.999876066588045e-07, + "loss": 0.7973, + "step": 753 + }, + { + "epoch": 1.3962962962962964, + "grad_norm": 2.023370613360696, + "learning_rate": 1.9998736947604323e-07, + "loss": 0.8064, + "step": 754 + }, + { + "epoch": 1.3981481481481481, + "grad_norm": 2.571057131225494, + "learning_rate": 1.9998713004529434e-07, + "loss": 0.7671, + "step": 755 + }, + { + "epoch": 1.4, + "grad_norm": 2.2689081255007357, + "learning_rate": 1.9998688836656322e-07, + "loss": 0.761, + "step": 756 + }, + { + "epoch": 1.401851851851852, + "grad_norm": 1.4377393351999066, + "learning_rate": 1.9998664443985524e-07, + "loss": 0.745, + "step": 757 + }, + { + "epoch": 1.4037037037037037, + "grad_norm": 2.1939721261948253, + "learning_rate": 1.99986398265176e-07, + "loss": 0.7851, + "step": 758 + }, + { + "epoch": 1.4055555555555554, + "grad_norm": 1.4503461755590286, + "learning_rate": 1.999861498425309e-07, + "loss": 0.7936, + "step": 759 + }, + { + "epoch": 1.4074074074074074, + "grad_norm": 1.8394565547533492, + "learning_rate": 1.9998589917192564e-07, + "loss": 0.7988, + "step": 760 + }, + { + "epoch": 1.4092592592592592, + "grad_norm": 1.468303491569509, + "learning_rate": 1.999856462533658e-07, + "loss": 0.7708, + "step": 761 + }, + { + "epoch": 1.411111111111111, + "grad_norm": 2.0549287530631526, + "learning_rate": 1.9998539108685708e-07, + "loss": 0.7606, + "step": 762 + }, + { + "epoch": 1.412962962962963, + "grad_norm": 1.4655881735425278, + "learning_rate": 1.9998513367240523e-07, + "loss": 0.8103, + "step": 763 + }, + { + "epoch": 1.4148148148148147, + "grad_norm": 1.3635001556298394, + "learning_rate": 1.9998487401001597e-07, + "loss": 0.761, + "step": 764 + }, + { + "epoch": 1.4166666666666667, + "grad_norm": 1.3484868745303848, + "learning_rate": 1.9998461209969522e-07, + "loss": 0.7891, + "step": 765 + }, + { + "epoch": 1.4185185185185185, + "grad_norm": 2.381878464718892, + "learning_rate": 1.9998434794144884e-07, + "loss": 0.7819, + "step": 766 + }, + { + "epoch": 1.4203703703703705, + "grad_norm": 1.7386202947061842, + "learning_rate": 1.999840815352827e-07, + "loss": 0.8041, + "step": 767 + }, + { + "epoch": 1.4222222222222223, + "grad_norm": 1.3368181347470822, + "learning_rate": 1.9998381288120295e-07, + "loss": 0.8021, + "step": 768 + }, + { + "epoch": 1.424074074074074, + "grad_norm": 3.1464029983683677, + "learning_rate": 1.9998354197921548e-07, + "loss": 0.7841, + "step": 769 + }, + { + "epoch": 1.425925925925926, + "grad_norm": 1.8305186362597654, + "learning_rate": 1.9998326882932643e-07, + "loss": 0.7639, + "step": 770 + }, + { + "epoch": 1.4277777777777778, + "grad_norm": 1.2428686113792757, + "learning_rate": 1.9998299343154198e-07, + "loss": 0.7986, + "step": 771 + }, + { + "epoch": 1.4296296296296296, + "grad_norm": 1.392108608033091, + "learning_rate": 1.9998271578586827e-07, + "loss": 0.8157, + "step": 772 + }, + { + "epoch": 1.4314814814814816, + "grad_norm": 1.5540628226099635, + "learning_rate": 1.9998243589231153e-07, + "loss": 0.7705, + "step": 773 + }, + { + "epoch": 1.4333333333333333, + "grad_norm": 1.543105889651524, + "learning_rate": 1.9998215375087813e-07, + "loss": 0.7952, + "step": 774 + }, + { + "epoch": 1.4351851851851851, + "grad_norm": 1.3917823867151182, + "learning_rate": 1.9998186936157436e-07, + "loss": 0.8254, + "step": 775 + }, + { + "epoch": 1.4370370370370371, + "grad_norm": 1.2977219318503417, + "learning_rate": 1.999815827244066e-07, + "loss": 0.811, + "step": 776 + }, + { + "epoch": 1.4388888888888889, + "grad_norm": 1.4195135393158569, + "learning_rate": 1.9998129383938133e-07, + "loss": 0.7993, + "step": 777 + }, + { + "epoch": 1.4407407407407407, + "grad_norm": 1.332976070867669, + "learning_rate": 1.9998100270650502e-07, + "loss": 0.7471, + "step": 778 + }, + { + "epoch": 1.4425925925925926, + "grad_norm": 1.5295871560467964, + "learning_rate": 1.9998070932578422e-07, + "loss": 0.7723, + "step": 779 + }, + { + "epoch": 1.4444444444444444, + "grad_norm": 1.3320993191659307, + "learning_rate": 1.9998041369722555e-07, + "loss": 0.7905, + "step": 780 + }, + { + "epoch": 1.4462962962962962, + "grad_norm": 2.5443449059167342, + "learning_rate": 1.9998011582083564e-07, + "loss": 0.8171, + "step": 781 + }, + { + "epoch": 1.4481481481481482, + "grad_norm": 1.4307395946540447, + "learning_rate": 1.9997981569662116e-07, + "loss": 0.7868, + "step": 782 + }, + { + "epoch": 1.45, + "grad_norm": 1.458921852302464, + "learning_rate": 1.9997951332458888e-07, + "loss": 0.8283, + "step": 783 + }, + { + "epoch": 1.4518518518518517, + "grad_norm": 1.3580288423255023, + "learning_rate": 1.9997920870474563e-07, + "loss": 0.862, + "step": 784 + }, + { + "epoch": 1.4537037037037037, + "grad_norm": 1.5810332384189063, + "learning_rate": 1.9997890183709824e-07, + "loss": 0.8137, + "step": 785 + }, + { + "epoch": 1.4555555555555555, + "grad_norm": 1.3458762592152296, + "learning_rate": 1.9997859272165354e-07, + "loss": 0.7453, + "step": 786 + }, + { + "epoch": 1.4574074074074075, + "grad_norm": 4.059567578444875, + "learning_rate": 1.9997828135841856e-07, + "loss": 0.7874, + "step": 787 + }, + { + "epoch": 1.4592592592592593, + "grad_norm": 1.9165797879044952, + "learning_rate": 1.999779677474003e-07, + "loss": 0.7749, + "step": 788 + }, + { + "epoch": 1.4611111111111112, + "grad_norm": 1.3348032667290637, + "learning_rate": 1.9997765188860577e-07, + "loss": 0.7775, + "step": 789 + }, + { + "epoch": 1.462962962962963, + "grad_norm": 1.2852606649162384, + "learning_rate": 1.999773337820421e-07, + "loss": 0.7656, + "step": 790 + }, + { + "epoch": 1.4648148148148148, + "grad_norm": 1.3187710472389504, + "learning_rate": 1.9997701342771642e-07, + "loss": 0.8221, + "step": 791 + }, + { + "epoch": 1.4666666666666668, + "grad_norm": 1.3164764119936931, + "learning_rate": 1.9997669082563595e-07, + "loss": 0.8143, + "step": 792 + }, + { + "epoch": 1.4685185185185186, + "grad_norm": 1.5709538616240728, + "learning_rate": 1.9997636597580794e-07, + "loss": 0.7649, + "step": 793 + }, + { + "epoch": 1.4703703703703703, + "grad_norm": 1.5347942294432204, + "learning_rate": 1.999760388782397e-07, + "loss": 0.7905, + "step": 794 + }, + { + "epoch": 1.4722222222222223, + "grad_norm": 6.655542854853261, + "learning_rate": 1.999757095329386e-07, + "loss": 0.7781, + "step": 795 + }, + { + "epoch": 1.474074074074074, + "grad_norm": 1.5316136147155046, + "learning_rate": 1.9997537793991194e-07, + "loss": 0.8179, + "step": 796 + }, + { + "epoch": 1.4759259259259259, + "grad_norm": 1.7833710510859966, + "learning_rate": 1.9997504409916733e-07, + "loss": 0.7802, + "step": 797 + }, + { + "epoch": 1.4777777777777779, + "grad_norm": 5.27321204738106, + "learning_rate": 1.9997470801071218e-07, + "loss": 0.7445, + "step": 798 + }, + { + "epoch": 1.4796296296296296, + "grad_norm": 1.6615755275028636, + "learning_rate": 1.9997436967455404e-07, + "loss": 0.7525, + "step": 799 + }, + { + "epoch": 1.4814814814814814, + "grad_norm": 1.3827060496974195, + "learning_rate": 1.9997402909070057e-07, + "loss": 0.8076, + "step": 800 + }, + { + "epoch": 1.4833333333333334, + "grad_norm": 1.515486984875506, + "learning_rate": 1.999736862591594e-07, + "loss": 0.7582, + "step": 801 + }, + { + "epoch": 1.4851851851851852, + "grad_norm": 2.1911955532337553, + "learning_rate": 1.9997334117993823e-07, + "loss": 0.7468, + "step": 802 + }, + { + "epoch": 1.487037037037037, + "grad_norm": 1.3137153001241144, + "learning_rate": 1.9997299385304483e-07, + "loss": 0.849, + "step": 803 + }, + { + "epoch": 1.488888888888889, + "grad_norm": 1.4033594381191727, + "learning_rate": 1.9997264427848698e-07, + "loss": 0.7699, + "step": 804 + }, + { + "epoch": 1.4907407407407407, + "grad_norm": 1.616486331551188, + "learning_rate": 1.999722924562726e-07, + "loss": 0.8119, + "step": 805 + }, + { + "epoch": 1.4925925925925925, + "grad_norm": 1.611378589097031, + "learning_rate": 1.9997193838640955e-07, + "loss": 0.7491, + "step": 806 + }, + { + "epoch": 1.4944444444444445, + "grad_norm": 1.8885748958730557, + "learning_rate": 1.9997158206890582e-07, + "loss": 0.7771, + "step": 807 + }, + { + "epoch": 1.4962962962962962, + "grad_norm": 1.2975278134562398, + "learning_rate": 1.9997122350376938e-07, + "loss": 0.8015, + "step": 808 + }, + { + "epoch": 1.4981481481481482, + "grad_norm": 1.5747918569426438, + "learning_rate": 1.9997086269100833e-07, + "loss": 0.8423, + "step": 809 + }, + { + "epoch": 1.5, + "grad_norm": 1.8186690943705288, + "learning_rate": 1.999704996306308e-07, + "loss": 0.8081, + "step": 810 + }, + { + "epoch": 1.501851851851852, + "grad_norm": 1.3900499893368639, + "learning_rate": 1.999701343226449e-07, + "loss": 0.7831, + "step": 811 + }, + { + "epoch": 1.5037037037037035, + "grad_norm": 1.414262466865652, + "learning_rate": 1.9996976676705886e-07, + "loss": 0.7735, + "step": 812 + }, + { + "epoch": 1.5055555555555555, + "grad_norm": 1.3854954652081493, + "learning_rate": 1.9996939696388092e-07, + "loss": 0.8117, + "step": 813 + }, + { + "epoch": 1.5074074074074075, + "grad_norm": 1.3352991112748513, + "learning_rate": 1.9996902491311948e-07, + "loss": 0.759, + "step": 814 + }, + { + "epoch": 1.5092592592592593, + "grad_norm": 1.4606364588246132, + "learning_rate": 1.999686506147828e-07, + "loss": 0.7859, + "step": 815 + }, + { + "epoch": 1.511111111111111, + "grad_norm": 1.3072406803222494, + "learning_rate": 1.999682740688794e-07, + "loss": 0.7727, + "step": 816 + }, + { + "epoch": 1.512962962962963, + "grad_norm": 1.577789969006411, + "learning_rate": 1.9996789527541766e-07, + "loss": 0.7997, + "step": 817 + }, + { + "epoch": 1.5148148148148148, + "grad_norm": 1.3342314218130178, + "learning_rate": 1.9996751423440612e-07, + "loss": 0.7569, + "step": 818 + }, + { + "epoch": 1.5166666666666666, + "grad_norm": 1.4337330612761081, + "learning_rate": 1.9996713094585334e-07, + "loss": 0.8181, + "step": 819 + }, + { + "epoch": 1.5185185185185186, + "grad_norm": 1.4425600988022205, + "learning_rate": 1.9996674540976797e-07, + "loss": 0.8009, + "step": 820 + }, + { + "epoch": 1.5203703703703704, + "grad_norm": 1.3742195597196458, + "learning_rate": 1.9996635762615863e-07, + "loss": 0.8167, + "step": 821 + }, + { + "epoch": 1.5222222222222221, + "grad_norm": 1.5195254091486206, + "learning_rate": 1.999659675950341e-07, + "loss": 0.7807, + "step": 822 + }, + { + "epoch": 1.5240740740740741, + "grad_norm": 1.3234522103246977, + "learning_rate": 1.9996557531640312e-07, + "loss": 0.8112, + "step": 823 + }, + { + "epoch": 1.525925925925926, + "grad_norm": 1.4747702928793858, + "learning_rate": 1.9996518079027446e-07, + "loss": 0.7531, + "step": 824 + }, + { + "epoch": 1.5277777777777777, + "grad_norm": 1.4713620896887833, + "learning_rate": 1.9996478401665708e-07, + "loss": 0.7493, + "step": 825 + }, + { + "epoch": 1.5296296296296297, + "grad_norm": 1.332434865024635, + "learning_rate": 1.9996438499555982e-07, + "loss": 0.786, + "step": 826 + }, + { + "epoch": 1.5314814814814814, + "grad_norm": 1.3230372557402514, + "learning_rate": 1.999639837269917e-07, + "loss": 0.7977, + "step": 827 + }, + { + "epoch": 1.5333333333333332, + "grad_norm": 1.667462527734816, + "learning_rate": 1.9996358021096172e-07, + "loss": 0.7757, + "step": 828 + }, + { + "epoch": 1.5351851851851852, + "grad_norm": 1.4809443789626107, + "learning_rate": 1.99963174447479e-07, + "loss": 0.8128, + "step": 829 + }, + { + "epoch": 1.5370370370370372, + "grad_norm": 1.4015098922549134, + "learning_rate": 1.9996276643655258e-07, + "loss": 0.7927, + "step": 830 + }, + { + "epoch": 1.5388888888888888, + "grad_norm": 6.474043305591081, + "learning_rate": 1.999623561781917e-07, + "loss": 0.7773, + "step": 831 + }, + { + "epoch": 1.5407407407407407, + "grad_norm": 1.3965801879089659, + "learning_rate": 1.9996194367240554e-07, + "loss": 0.7864, + "step": 832 + }, + { + "epoch": 1.5425925925925927, + "grad_norm": 1.6855553465095316, + "learning_rate": 1.9996152891920342e-07, + "loss": 0.7943, + "step": 833 + }, + { + "epoch": 1.5444444444444443, + "grad_norm": 1.3397043658891168, + "learning_rate": 1.999611119185946e-07, + "loss": 0.7993, + "step": 834 + }, + { + "epoch": 1.5462962962962963, + "grad_norm": 1.6467530903268044, + "learning_rate": 1.9996069267058851e-07, + "loss": 0.8165, + "step": 835 + }, + { + "epoch": 1.5481481481481483, + "grad_norm": 1.379256390059123, + "learning_rate": 1.9996027117519458e-07, + "loss": 0.7634, + "step": 836 + }, + { + "epoch": 1.55, + "grad_norm": 1.4035934399022019, + "learning_rate": 1.9995984743242226e-07, + "loss": 0.8012, + "step": 837 + }, + { + "epoch": 1.5518518518518518, + "grad_norm": 1.3337453757666426, + "learning_rate": 1.9995942144228108e-07, + "loss": 0.8022, + "step": 838 + }, + { + "epoch": 1.5537037037037038, + "grad_norm": 1.7528515777046711, + "learning_rate": 1.999589932047806e-07, + "loss": 0.76, + "step": 839 + }, + { + "epoch": 1.5555555555555556, + "grad_norm": 1.9684089615083693, + "learning_rate": 1.999585627199305e-07, + "loss": 0.8091, + "step": 840 + }, + { + "epoch": 1.5574074074074074, + "grad_norm": 1.427520378401312, + "learning_rate": 1.9995812998774038e-07, + "loss": 0.833, + "step": 841 + }, + { + "epoch": 1.5592592592592593, + "grad_norm": 1.3881160996675828, + "learning_rate": 1.9995769500822006e-07, + "loss": 0.7653, + "step": 842 + }, + { + "epoch": 1.5611111111111111, + "grad_norm": 2.2195578266559184, + "learning_rate": 1.9995725778137924e-07, + "loss": 0.8062, + "step": 843 + }, + { + "epoch": 1.5629629629629629, + "grad_norm": 1.2614323844099824, + "learning_rate": 1.9995681830722782e-07, + "loss": 0.786, + "step": 844 + }, + { + "epoch": 1.5648148148148149, + "grad_norm": 1.5655955039081937, + "learning_rate": 1.9995637658577562e-07, + "loss": 0.7901, + "step": 845 + }, + { + "epoch": 1.5666666666666667, + "grad_norm": 1.3021620993914482, + "learning_rate": 1.999559326170326e-07, + "loss": 0.8106, + "step": 846 + }, + { + "epoch": 1.5685185185185184, + "grad_norm": 1.442966557857159, + "learning_rate": 1.9995548640100878e-07, + "loss": 0.7262, + "step": 847 + }, + { + "epoch": 1.5703703703703704, + "grad_norm": 1.8039328844248892, + "learning_rate": 1.999550379377141e-07, + "loss": 0.7614, + "step": 848 + }, + { + "epoch": 1.5722222222222222, + "grad_norm": 1.5717220513889545, + "learning_rate": 1.999545872271587e-07, + "loss": 0.7749, + "step": 849 + }, + { + "epoch": 1.574074074074074, + "grad_norm": 1.4449487934438003, + "learning_rate": 1.999541342693527e-07, + "loss": 0.7997, + "step": 850 + }, + { + "epoch": 1.575925925925926, + "grad_norm": 2.3902711381662884, + "learning_rate": 1.9995367906430634e-07, + "loss": 0.7744, + "step": 851 + }, + { + "epoch": 1.5777777777777777, + "grad_norm": 1.5521337852002064, + "learning_rate": 1.9995322161202974e-07, + "loss": 0.7766, + "step": 852 + }, + { + "epoch": 1.5796296296296295, + "grad_norm": 1.4810952028164313, + "learning_rate": 1.999527619125333e-07, + "loss": 0.792, + "step": 853 + }, + { + "epoch": 1.5814814814814815, + "grad_norm": 1.873965974787256, + "learning_rate": 1.9995229996582727e-07, + "loss": 0.7844, + "step": 854 + }, + { + "epoch": 1.5833333333333335, + "grad_norm": 1.5975769776683897, + "learning_rate": 1.999518357719221e-07, + "loss": 0.7781, + "step": 855 + }, + { + "epoch": 1.585185185185185, + "grad_norm": 1.3433087398314394, + "learning_rate": 1.9995136933082818e-07, + "loss": 0.7814, + "step": 856 + }, + { + "epoch": 1.587037037037037, + "grad_norm": 1.4733235690859368, + "learning_rate": 1.9995090064255601e-07, + "loss": 0.8272, + "step": 857 + }, + { + "epoch": 1.588888888888889, + "grad_norm": 1.538597131002794, + "learning_rate": 1.9995042970711614e-07, + "loss": 0.7559, + "step": 858 + }, + { + "epoch": 1.5907407407407408, + "grad_norm": 1.8744026521188375, + "learning_rate": 1.9994995652451916e-07, + "loss": 0.735, + "step": 859 + }, + { + "epoch": 1.5925925925925926, + "grad_norm": 1.3854505744333534, + "learning_rate": 1.9994948109477567e-07, + "loss": 0.8418, + "step": 860 + }, + { + "epoch": 1.5944444444444446, + "grad_norm": 1.4661803675537535, + "learning_rate": 1.9994900341789643e-07, + "loss": 0.7467, + "step": 861 + }, + { + "epoch": 1.5962962962962963, + "grad_norm": 1.3205550748055843, + "learning_rate": 1.9994852349389208e-07, + "loss": 0.812, + "step": 862 + }, + { + "epoch": 1.598148148148148, + "grad_norm": 1.453642466089374, + "learning_rate": 1.999480413227735e-07, + "loss": 0.762, + "step": 863 + }, + { + "epoch": 1.6, + "grad_norm": 1.5143196961255514, + "learning_rate": 1.999475569045515e-07, + "loss": 0.7903, + "step": 864 + }, + { + "epoch": 1.6018518518518519, + "grad_norm": 1.3319493284980677, + "learning_rate": 1.9994707023923694e-07, + "loss": 0.797, + "step": 865 + }, + { + "epoch": 1.6037037037037036, + "grad_norm": 1.8022799954583217, + "learning_rate": 1.999465813268408e-07, + "loss": 0.8108, + "step": 866 + }, + { + "epoch": 1.6055555555555556, + "grad_norm": 1.3501851646630176, + "learning_rate": 1.9994609016737407e-07, + "loss": 0.7945, + "step": 867 + }, + { + "epoch": 1.6074074074074074, + "grad_norm": 1.524640301680115, + "learning_rate": 1.999455967608478e-07, + "loss": 0.7771, + "step": 868 + }, + { + "epoch": 1.6092592592592592, + "grad_norm": 2.3873074893284825, + "learning_rate": 1.9994510110727302e-07, + "loss": 0.7603, + "step": 869 + }, + { + "epoch": 1.6111111111111112, + "grad_norm": 3.7068813222005312, + "learning_rate": 1.9994460320666095e-07, + "loss": 0.7635, + "step": 870 + }, + { + "epoch": 1.612962962962963, + "grad_norm": 1.4231771848430872, + "learning_rate": 1.9994410305902277e-07, + "loss": 0.7609, + "step": 871 + }, + { + "epoch": 1.6148148148148147, + "grad_norm": 1.531615697390392, + "learning_rate": 1.9994360066436967e-07, + "loss": 0.819, + "step": 872 + }, + { + "epoch": 1.6166666666666667, + "grad_norm": 1.317558498348599, + "learning_rate": 1.9994309602271299e-07, + "loss": 0.7676, + "step": 873 + }, + { + "epoch": 1.6185185185185185, + "grad_norm": 2.2429093846450203, + "learning_rate": 1.999425891340641e-07, + "loss": 0.78, + "step": 874 + }, + { + "epoch": 1.6203703703703702, + "grad_norm": 1.3729813366205148, + "learning_rate": 1.9994207999843434e-07, + "loss": 0.82, + "step": 875 + }, + { + "epoch": 1.6222222222222222, + "grad_norm": 1.340530718512813, + "learning_rate": 1.999415686158352e-07, + "loss": 0.7205, + "step": 876 + }, + { + "epoch": 1.6240740740740742, + "grad_norm": 1.6237245193881744, + "learning_rate": 1.9994105498627816e-07, + "loss": 0.8091, + "step": 877 + }, + { + "epoch": 1.6259259259259258, + "grad_norm": 1.4879353413850342, + "learning_rate": 1.9994053910977475e-07, + "loss": 0.8508, + "step": 878 + }, + { + "epoch": 1.6277777777777778, + "grad_norm": 1.8625800229898672, + "learning_rate": 1.999400209863366e-07, + "loss": 0.7712, + "step": 879 + }, + { + "epoch": 1.6296296296296298, + "grad_norm": 1.3024264983148957, + "learning_rate": 1.9993950061597534e-07, + "loss": 0.8029, + "step": 880 + }, + { + "epoch": 1.6314814814814815, + "grad_norm": 1.3147706316257266, + "learning_rate": 1.9993897799870266e-07, + "loss": 0.7875, + "step": 881 + }, + { + "epoch": 1.6333333333333333, + "grad_norm": 1.3194809596557366, + "learning_rate": 1.9993845313453038e-07, + "loss": 0.7883, + "step": 882 + }, + { + "epoch": 1.6351851851851853, + "grad_norm": 1.4183417869141677, + "learning_rate": 1.999379260234702e-07, + "loss": 0.7552, + "step": 883 + }, + { + "epoch": 1.637037037037037, + "grad_norm": 1.3947153543970918, + "learning_rate": 1.9993739666553402e-07, + "loss": 0.7563, + "step": 884 + }, + { + "epoch": 1.6388888888888888, + "grad_norm": 1.3268726506713047, + "learning_rate": 1.9993686506073373e-07, + "loss": 0.7818, + "step": 885 + }, + { + "epoch": 1.6407407407407408, + "grad_norm": 1.4612349845390744, + "learning_rate": 1.9993633120908128e-07, + "loss": 0.7743, + "step": 886 + }, + { + "epoch": 1.6425925925925926, + "grad_norm": 1.4598335751397802, + "learning_rate": 1.9993579511058872e-07, + "loss": 0.7799, + "step": 887 + }, + { + "epoch": 1.6444444444444444, + "grad_norm": 1.5354530703186837, + "learning_rate": 1.9993525676526804e-07, + "loss": 0.8388, + "step": 888 + }, + { + "epoch": 1.6462962962962964, + "grad_norm": 1.4793914337304956, + "learning_rate": 1.999347161731314e-07, + "loss": 0.776, + "step": 889 + }, + { + "epoch": 1.6481481481481481, + "grad_norm": 1.6079250075291351, + "learning_rate": 1.9993417333419088e-07, + "loss": 0.7595, + "step": 890 + }, + { + "epoch": 1.65, + "grad_norm": 1.3685160632787292, + "learning_rate": 1.9993362824845874e-07, + "loss": 0.8028, + "step": 891 + }, + { + "epoch": 1.651851851851852, + "grad_norm": 1.4776133538145957, + "learning_rate": 1.9993308091594722e-07, + "loss": 0.7373, + "step": 892 + }, + { + "epoch": 1.6537037037037037, + "grad_norm": 1.3876764163078121, + "learning_rate": 1.9993253133666864e-07, + "loss": 0.7876, + "step": 893 + }, + { + "epoch": 1.6555555555555554, + "grad_norm": 1.5186057063342262, + "learning_rate": 1.9993197951063534e-07, + "loss": 0.7707, + "step": 894 + }, + { + "epoch": 1.6574074074074074, + "grad_norm": 1.6561840526852425, + "learning_rate": 1.999314254378597e-07, + "loss": 0.7678, + "step": 895 + }, + { + "epoch": 1.6592592592592592, + "grad_norm": 3.340669311151622, + "learning_rate": 1.9993086911835424e-07, + "loss": 0.7829, + "step": 896 + }, + { + "epoch": 1.661111111111111, + "grad_norm": 1.5364577004626374, + "learning_rate": 1.9993031055213142e-07, + "loss": 0.767, + "step": 897 + }, + { + "epoch": 1.662962962962963, + "grad_norm": 1.494286308386131, + "learning_rate": 1.999297497392038e-07, + "loss": 0.7709, + "step": 898 + }, + { + "epoch": 1.664814814814815, + "grad_norm": 1.7165139560096818, + "learning_rate": 1.9992918667958404e-07, + "loss": 0.75, + "step": 899 + }, + { + "epoch": 1.6666666666666665, + "grad_norm": 1.5151312622380881, + "learning_rate": 1.999286213732847e-07, + "loss": 0.7655, + "step": 900 + }, + { + "epoch": 1.6685185185185185, + "grad_norm": 1.4393055887815804, + "learning_rate": 1.999280538203186e-07, + "loss": 0.7877, + "step": 901 + }, + { + "epoch": 1.6703703703703705, + "grad_norm": 1.3909681715905493, + "learning_rate": 1.9992748402069846e-07, + "loss": 0.7931, + "step": 902 + }, + { + "epoch": 1.6722222222222223, + "grad_norm": 1.3367514800859155, + "learning_rate": 1.9992691197443703e-07, + "loss": 0.7675, + "step": 903 + }, + { + "epoch": 1.674074074074074, + "grad_norm": 1.5660279955796936, + "learning_rate": 1.9992633768154726e-07, + "loss": 0.7281, + "step": 904 + }, + { + "epoch": 1.675925925925926, + "grad_norm": 1.6593436189179602, + "learning_rate": 1.99925761142042e-07, + "loss": 0.803, + "step": 905 + }, + { + "epoch": 1.6777777777777778, + "grad_norm": 1.4859403459535097, + "learning_rate": 1.9992518235593424e-07, + "loss": 0.7751, + "step": 906 + }, + { + "epoch": 1.6796296296296296, + "grad_norm": 1.4123037224546764, + "learning_rate": 1.99924601323237e-07, + "loss": 0.7416, + "step": 907 + }, + { + "epoch": 1.6814814814814816, + "grad_norm": 1.2990577938387011, + "learning_rate": 1.9992401804396334e-07, + "loss": 0.7485, + "step": 908 + }, + { + "epoch": 1.6833333333333333, + "grad_norm": 1.3759901083208825, + "learning_rate": 1.9992343251812634e-07, + "loss": 0.7779, + "step": 909 + }, + { + "epoch": 1.6851851851851851, + "grad_norm": 1.6708555846734454, + "learning_rate": 1.999228447457392e-07, + "loss": 0.6831, + "step": 910 + }, + { + "epoch": 1.6870370370370371, + "grad_norm": 1.3920984218182832, + "learning_rate": 1.9992225472681513e-07, + "loss": 0.788, + "step": 911 + }, + { + "epoch": 1.6888888888888889, + "grad_norm": 1.3539725437171173, + "learning_rate": 1.9992166246136736e-07, + "loss": 0.745, + "step": 912 + }, + { + "epoch": 1.6907407407407407, + "grad_norm": 1.3179970990736476, + "learning_rate": 1.9992106794940925e-07, + "loss": 0.7424, + "step": 913 + }, + { + "epoch": 1.6925925925925926, + "grad_norm": 1.6647135105039546, + "learning_rate": 1.9992047119095418e-07, + "loss": 0.7891, + "step": 914 + }, + { + "epoch": 1.6944444444444444, + "grad_norm": 1.390235185114431, + "learning_rate": 1.999198721860155e-07, + "loss": 0.7372, + "step": 915 + }, + { + "epoch": 1.6962962962962962, + "grad_norm": 1.250125589913379, + "learning_rate": 1.9991927093460674e-07, + "loss": 0.7532, + "step": 916 + }, + { + "epoch": 1.6981481481481482, + "grad_norm": 1.432752558537211, + "learning_rate": 1.9991866743674137e-07, + "loss": 0.7855, + "step": 917 + }, + { + "epoch": 1.7, + "grad_norm": 1.414140842812916, + "learning_rate": 1.9991806169243298e-07, + "loss": 0.7831, + "step": 918 + }, + { + "epoch": 1.7018518518518517, + "grad_norm": 2.3584733494694605, + "learning_rate": 1.9991745370169523e-07, + "loss": 0.7428, + "step": 919 + }, + { + "epoch": 1.7037037037037037, + "grad_norm": 1.5400767488738312, + "learning_rate": 1.9991684346454168e-07, + "loss": 0.8048, + "step": 920 + }, + { + "epoch": 1.7055555555555557, + "grad_norm": 1.5216429489119565, + "learning_rate": 1.9991623098098617e-07, + "loss": 0.7573, + "step": 921 + }, + { + "epoch": 1.7074074074074073, + "grad_norm": 1.3402999091218104, + "learning_rate": 1.9991561625104242e-07, + "loss": 0.8006, + "step": 922 + }, + { + "epoch": 1.7092592592592593, + "grad_norm": 1.6367634390472985, + "learning_rate": 1.9991499927472425e-07, + "loss": 0.7638, + "step": 923 + }, + { + "epoch": 1.7111111111111112, + "grad_norm": 1.3246648877724783, + "learning_rate": 1.9991438005204552e-07, + "loss": 0.749, + "step": 924 + }, + { + "epoch": 1.7129629629629628, + "grad_norm": 1.637132614117255, + "learning_rate": 1.9991375858302018e-07, + "loss": 0.7832, + "step": 925 + }, + { + "epoch": 1.7148148148148148, + "grad_norm": 1.329745756943769, + "learning_rate": 1.9991313486766215e-07, + "loss": 0.7881, + "step": 926 + }, + { + "epoch": 1.7166666666666668, + "grad_norm": 1.39331932578327, + "learning_rate": 1.9991250890598553e-07, + "loss": 0.7573, + "step": 927 + }, + { + "epoch": 1.7185185185185186, + "grad_norm": 1.3887427303224111, + "learning_rate": 1.9991188069800431e-07, + "loss": 0.8576, + "step": 928 + }, + { + "epoch": 1.7203703703703703, + "grad_norm": 1.30847382261147, + "learning_rate": 1.9991125024373267e-07, + "loss": 0.7765, + "step": 929 + }, + { + "epoch": 1.7222222222222223, + "grad_norm": 1.459885257803674, + "learning_rate": 1.999106175431848e-07, + "loss": 0.7201, + "step": 930 + }, + { + "epoch": 1.724074074074074, + "grad_norm": 1.2439278216824947, + "learning_rate": 1.9990998259637486e-07, + "loss": 0.7867, + "step": 931 + }, + { + "epoch": 1.7259259259259259, + "grad_norm": 1.3240204787554097, + "learning_rate": 1.9990934540331717e-07, + "loss": 0.7553, + "step": 932 + }, + { + "epoch": 1.7277777777777779, + "grad_norm": 1.6480773658958212, + "learning_rate": 1.9990870596402602e-07, + "loss": 0.7713, + "step": 933 + }, + { + "epoch": 1.7296296296296296, + "grad_norm": 1.4536599548496305, + "learning_rate": 1.9990806427851585e-07, + "loss": 0.7525, + "step": 934 + }, + { + "epoch": 1.7314814814814814, + "grad_norm": 1.2636993216040457, + "learning_rate": 1.9990742034680105e-07, + "loss": 0.7284, + "step": 935 + }, + { + "epoch": 1.7333333333333334, + "grad_norm": 2.5749205613917225, + "learning_rate": 1.9990677416889606e-07, + "loss": 0.8042, + "step": 936 + }, + { + "epoch": 1.7351851851851852, + "grad_norm": 1.49376272212408, + "learning_rate": 1.9990612574481547e-07, + "loss": 0.7817, + "step": 937 + }, + { + "epoch": 1.737037037037037, + "grad_norm": 1.7261974542179666, + "learning_rate": 1.9990547507457383e-07, + "loss": 0.7779, + "step": 938 + }, + { + "epoch": 1.738888888888889, + "grad_norm": 1.4497238464789381, + "learning_rate": 1.999048221581858e-07, + "loss": 0.7577, + "step": 939 + }, + { + "epoch": 1.7407407407407407, + "grad_norm": 1.509865668190673, + "learning_rate": 1.9990416699566596e-07, + "loss": 0.7903, + "step": 940 + }, + { + "epoch": 1.7425925925925925, + "grad_norm": 1.3596773090201275, + "learning_rate": 1.9990350958702918e-07, + "loss": 0.7765, + "step": 941 + }, + { + "epoch": 1.7444444444444445, + "grad_norm": 1.423302817350013, + "learning_rate": 1.9990284993229012e-07, + "loss": 0.8126, + "step": 942 + }, + { + "epoch": 1.7462962962962965, + "grad_norm": 1.239754724121764, + "learning_rate": 1.9990218803146368e-07, + "loss": 0.7453, + "step": 943 + }, + { + "epoch": 1.748148148148148, + "grad_norm": 1.3817845223584746, + "learning_rate": 1.9990152388456471e-07, + "loss": 0.7379, + "step": 944 + }, + { + "epoch": 1.75, + "grad_norm": 1.2875603320167348, + "learning_rate": 1.9990085749160818e-07, + "loss": 0.8182, + "step": 945 + }, + { + "epoch": 1.751851851851852, + "grad_norm": 2.0181658079683737, + "learning_rate": 1.9990018885260905e-07, + "loss": 0.8163, + "step": 946 + }, + { + "epoch": 1.7537037037037035, + "grad_norm": 1.4952035292359285, + "learning_rate": 1.9989951796758234e-07, + "loss": 0.7634, + "step": 947 + }, + { + "epoch": 1.7555555555555555, + "grad_norm": 1.340738450776537, + "learning_rate": 1.998988448365431e-07, + "loss": 0.7872, + "step": 948 + }, + { + "epoch": 1.7574074074074075, + "grad_norm": 1.399537935293725, + "learning_rate": 1.9989816945950657e-07, + "loss": 0.7716, + "step": 949 + }, + { + "epoch": 1.7592592592592593, + "grad_norm": 1.535868346196315, + "learning_rate": 1.9989749183648785e-07, + "loss": 0.8115, + "step": 950 + }, + { + "epoch": 1.761111111111111, + "grad_norm": 1.433963445421848, + "learning_rate": 1.998968119675022e-07, + "loss": 0.8099, + "step": 951 + }, + { + "epoch": 1.762962962962963, + "grad_norm": 1.3873999924942957, + "learning_rate": 1.998961298525649e-07, + "loss": 0.7789, + "step": 952 + }, + { + "epoch": 1.7648148148148148, + "grad_norm": 1.3212520883435643, + "learning_rate": 1.9989544549169127e-07, + "loss": 0.7405, + "step": 953 + }, + { + "epoch": 1.7666666666666666, + "grad_norm": 1.3038838284160543, + "learning_rate": 1.9989475888489672e-07, + "loss": 0.7388, + "step": 954 + }, + { + "epoch": 1.7685185185185186, + "grad_norm": 1.4028278928487332, + "learning_rate": 1.998940700321967e-07, + "loss": 0.7484, + "step": 955 + }, + { + "epoch": 1.7703703703703704, + "grad_norm": 1.437825392929185, + "learning_rate": 1.9989337893360667e-07, + "loss": 0.7538, + "step": 956 + }, + { + "epoch": 1.7722222222222221, + "grad_norm": 1.3295417487043837, + "learning_rate": 1.9989268558914217e-07, + "loss": 0.7452, + "step": 957 + }, + { + "epoch": 1.7740740740740741, + "grad_norm": 1.6626271303687088, + "learning_rate": 1.9989198999881878e-07, + "loss": 0.8013, + "step": 958 + }, + { + "epoch": 1.775925925925926, + "grad_norm": 3.698332225453685, + "learning_rate": 1.9989129216265214e-07, + "loss": 0.7616, + "step": 959 + }, + { + "epoch": 1.7777777777777777, + "grad_norm": 1.3816989748673836, + "learning_rate": 1.99890592080658e-07, + "loss": 0.722, + "step": 960 + }, + { + "epoch": 1.7796296296296297, + "grad_norm": 1.3046144809430562, + "learning_rate": 1.9988988975285204e-07, + "loss": 0.7361, + "step": 961 + }, + { + "epoch": 1.7814814814814814, + "grad_norm": 1.4518523974344757, + "learning_rate": 1.9988918517925005e-07, + "loss": 0.8193, + "step": 962 + }, + { + "epoch": 1.7833333333333332, + "grad_norm": 2.3750512495840965, + "learning_rate": 1.9988847835986786e-07, + "loss": 0.7587, + "step": 963 + }, + { + "epoch": 1.7851851851851852, + "grad_norm": 2.192672507249222, + "learning_rate": 1.998877692947214e-07, + "loss": 0.7587, + "step": 964 + }, + { + "epoch": 1.7870370370370372, + "grad_norm": 1.5302685333436545, + "learning_rate": 1.9988705798382661e-07, + "loss": 0.7876, + "step": 965 + }, + { + "epoch": 1.7888888888888888, + "grad_norm": 1.418036599899706, + "learning_rate": 1.9988634442719948e-07, + "loss": 0.7684, + "step": 966 + }, + { + "epoch": 1.7907407407407407, + "grad_norm": 2.419996407593016, + "learning_rate": 1.99885628624856e-07, + "loss": 0.6958, + "step": 967 + }, + { + "epoch": 1.7925925925925927, + "grad_norm": 1.4944578188848778, + "learning_rate": 1.9988491057681232e-07, + "loss": 0.7705, + "step": 968 + }, + { + "epoch": 1.7944444444444443, + "grad_norm": 3.0693779890514, + "learning_rate": 1.9988419028308457e-07, + "loss": 0.739, + "step": 969 + }, + { + "epoch": 1.7962962962962963, + "grad_norm": 1.4341933458154759, + "learning_rate": 1.9988346774368894e-07, + "loss": 0.7418, + "step": 970 + }, + { + "epoch": 1.7981481481481483, + "grad_norm": 1.5736459192659946, + "learning_rate": 1.9988274295864167e-07, + "loss": 0.7601, + "step": 971 + }, + { + "epoch": 1.8, + "grad_norm": 1.4929965237519107, + "learning_rate": 1.9988201592795905e-07, + "loss": 0.803, + "step": 972 + }, + { + "epoch": 1.8018518518518518, + "grad_norm": 2.133063903152309, + "learning_rate": 1.9988128665165746e-07, + "loss": 0.7051, + "step": 973 + }, + { + "epoch": 1.8037037037037038, + "grad_norm": 1.4119194601919294, + "learning_rate": 1.9988055512975329e-07, + "loss": 0.7737, + "step": 974 + }, + { + "epoch": 1.8055555555555556, + "grad_norm": 1.3911171054460025, + "learning_rate": 1.9987982136226292e-07, + "loss": 0.8068, + "step": 975 + }, + { + "epoch": 1.8074074074074074, + "grad_norm": 1.9381069385259089, + "learning_rate": 1.9987908534920291e-07, + "loss": 0.7539, + "step": 976 + }, + { + "epoch": 1.8092592592592593, + "grad_norm": 1.6411214166656005, + "learning_rate": 1.9987834709058983e-07, + "loss": 0.7811, + "step": 977 + }, + { + "epoch": 1.8111111111111111, + "grad_norm": 1.507315883841706, + "learning_rate": 1.9987760658644023e-07, + "loss": 0.7836, + "step": 978 + }, + { + "epoch": 1.8129629629629629, + "grad_norm": 1.869801800461327, + "learning_rate": 1.998768638367707e-07, + "loss": 0.7749, + "step": 979 + }, + { + "epoch": 1.8148148148148149, + "grad_norm": 1.7565505719473904, + "learning_rate": 1.998761188415981e-07, + "loss": 0.7967, + "step": 980 + }, + { + "epoch": 1.8166666666666667, + "grad_norm": 1.6345579550760483, + "learning_rate": 1.9987537160093907e-07, + "loss": 0.792, + "step": 981 + }, + { + "epoch": 1.8185185185185184, + "grad_norm": 1.9607225127855206, + "learning_rate": 1.9987462211481038e-07, + "loss": 0.7928, + "step": 982 + }, + { + "epoch": 1.8203703703703704, + "grad_norm": 1.5188361447171843, + "learning_rate": 1.9987387038322898e-07, + "loss": 0.8108, + "step": 983 + }, + { + "epoch": 1.8222222222222222, + "grad_norm": 1.5413453361100633, + "learning_rate": 1.998731164062117e-07, + "loss": 0.7977, + "step": 984 + }, + { + "epoch": 1.824074074074074, + "grad_norm": 1.3757713481914136, + "learning_rate": 1.998723601837755e-07, + "loss": 0.749, + "step": 985 + }, + { + "epoch": 1.825925925925926, + "grad_norm": 1.4354818948523813, + "learning_rate": 1.9987160171593743e-07, + "loss": 0.8036, + "step": 986 + }, + { + "epoch": 1.8277777777777777, + "grad_norm": 1.597649565887714, + "learning_rate": 1.9987084100271448e-07, + "loss": 0.7323, + "step": 987 + }, + { + "epoch": 1.8296296296296295, + "grad_norm": 1.1936510229152042, + "learning_rate": 1.998700780441238e-07, + "loss": 0.8061, + "step": 988 + }, + { + "epoch": 1.8314814814814815, + "grad_norm": 1.7299553561365208, + "learning_rate": 1.9986931284018248e-07, + "loss": 0.7887, + "step": 989 + }, + { + "epoch": 1.8333333333333335, + "grad_norm": 1.64664118210252, + "learning_rate": 1.9986854539090781e-07, + "loss": 0.7784, + "step": 990 + }, + { + "epoch": 1.835185185185185, + "grad_norm": 1.9177963561964584, + "learning_rate": 1.99867775696317e-07, + "loss": 0.774, + "step": 991 + }, + { + "epoch": 1.837037037037037, + "grad_norm": 1.3093476778746875, + "learning_rate": 1.998670037564273e-07, + "loss": 0.7363, + "step": 992 + }, + { + "epoch": 1.838888888888889, + "grad_norm": 1.319367963893485, + "learning_rate": 1.9986622957125618e-07, + "loss": 0.7447, + "step": 993 + }, + { + "epoch": 1.8407407407407408, + "grad_norm": 2.676685381876739, + "learning_rate": 1.9986545314082097e-07, + "loss": 0.7519, + "step": 994 + }, + { + "epoch": 1.8425925925925926, + "grad_norm": 1.5326917801294218, + "learning_rate": 1.9986467446513915e-07, + "loss": 0.7681, + "step": 995 + }, + { + "epoch": 1.8444444444444446, + "grad_norm": 1.8419702936033213, + "learning_rate": 1.998638935442282e-07, + "loss": 0.7142, + "step": 996 + }, + { + "epoch": 1.8462962962962963, + "grad_norm": 1.3323039771053324, + "learning_rate": 1.998631103781057e-07, + "loss": 0.7286, + "step": 997 + }, + { + "epoch": 1.848148148148148, + "grad_norm": 1.3394903259952737, + "learning_rate": 1.9986232496678928e-07, + "loss": 0.753, + "step": 998 + }, + { + "epoch": 1.85, + "grad_norm": 1.7055420272114223, + "learning_rate": 1.9986153731029653e-07, + "loss": 0.7844, + "step": 999 + }, + { + "epoch": 1.8518518518518519, + "grad_norm": 2.34274316574407, + "learning_rate": 1.9986074740864525e-07, + "loss": 0.7619, + "step": 1000 + }, + { + "epoch": 1.8537037037037036, + "grad_norm": 1.2647293607810843, + "learning_rate": 1.9985995526185313e-07, + "loss": 0.7645, + "step": 1001 + }, + { + "epoch": 1.8555555555555556, + "grad_norm": 1.325895710474424, + "learning_rate": 1.99859160869938e-07, + "loss": 0.728, + "step": 1002 + }, + { + "epoch": 1.8574074074074074, + "grad_norm": 1.6743053959927081, + "learning_rate": 1.9985836423291771e-07, + "loss": 0.8377, + "step": 1003 + }, + { + "epoch": 1.8592592592592592, + "grad_norm": 1.914293654241324, + "learning_rate": 1.998575653508102e-07, + "loss": 0.7333, + "step": 1004 + }, + { + "epoch": 1.8611111111111112, + "grad_norm": 4.547147360648242, + "learning_rate": 1.9985676422363338e-07, + "loss": 0.7765, + "step": 1005 + }, + { + "epoch": 1.862962962962963, + "grad_norm": 1.3432555614109056, + "learning_rate": 1.9985596085140533e-07, + "loss": 0.7946, + "step": 1006 + }, + { + "epoch": 1.8648148148148147, + "grad_norm": 3.539767209438379, + "learning_rate": 1.9985515523414404e-07, + "loss": 0.7472, + "step": 1007 + }, + { + "epoch": 1.8666666666666667, + "grad_norm": 1.4946296123352325, + "learning_rate": 1.9985434737186769e-07, + "loss": 0.7934, + "step": 1008 + }, + { + "epoch": 1.8685185185185185, + "grad_norm": 1.4855968713815713, + "learning_rate": 1.998535372645944e-07, + "loss": 0.7589, + "step": 1009 + }, + { + "epoch": 1.8703703703703702, + "grad_norm": 1.2654104905041015, + "learning_rate": 1.9985272491234238e-07, + "loss": 0.7541, + "step": 1010 + }, + { + "epoch": 1.8722222222222222, + "grad_norm": 1.3082185718578754, + "learning_rate": 1.9985191031512993e-07, + "loss": 0.8079, + "step": 1011 + }, + { + "epoch": 1.8740740740740742, + "grad_norm": 1.6232591014864821, + "learning_rate": 1.9985109347297534e-07, + "loss": 0.8203, + "step": 1012 + }, + { + "epoch": 1.8759259259259258, + "grad_norm": 1.5545942041103336, + "learning_rate": 1.9985027438589696e-07, + "loss": 0.768, + "step": 1013 + }, + { + "epoch": 1.8777777777777778, + "grad_norm": 1.6783166712367008, + "learning_rate": 1.9984945305391323e-07, + "loss": 0.7649, + "step": 1014 + }, + { + "epoch": 1.8796296296296298, + "grad_norm": 1.4934396970379067, + "learning_rate": 1.998486294770426e-07, + "loss": 0.7254, + "step": 1015 + }, + { + "epoch": 1.8814814814814815, + "grad_norm": 1.8125299677321407, + "learning_rate": 1.998478036553036e-07, + "loss": 0.8056, + "step": 1016 + }, + { + "epoch": 1.8833333333333333, + "grad_norm": 1.785947678012694, + "learning_rate": 1.998469755887148e-07, + "loss": 0.756, + "step": 1017 + }, + { + "epoch": 1.8851851851851853, + "grad_norm": 1.5501905077907572, + "learning_rate": 1.9984614527729482e-07, + "loss": 0.7913, + "step": 1018 + }, + { + "epoch": 1.887037037037037, + "grad_norm": 1.4673739574958995, + "learning_rate": 1.9984531272106228e-07, + "loss": 0.786, + "step": 1019 + }, + { + "epoch": 1.8888888888888888, + "grad_norm": 1.5149206975977105, + "learning_rate": 1.9984447792003597e-07, + "loss": 0.7519, + "step": 1020 + }, + { + "epoch": 1.8907407407407408, + "grad_norm": 1.5332146548019172, + "learning_rate": 1.9984364087423463e-07, + "loss": 0.8084, + "step": 1021 + }, + { + "epoch": 1.8925925925925926, + "grad_norm": 1.6798081001422076, + "learning_rate": 1.9984280158367703e-07, + "loss": 0.7245, + "step": 1022 + }, + { + "epoch": 1.8944444444444444, + "grad_norm": 1.3687354583030944, + "learning_rate": 1.998419600483821e-07, + "loss": 0.8057, + "step": 1023 + }, + { + "epoch": 1.8962962962962964, + "grad_norm": 1.4728179311724792, + "learning_rate": 1.9984111626836875e-07, + "loss": 0.7358, + "step": 1024 + }, + { + "epoch": 1.8981481481481481, + "grad_norm": 1.4682529755028493, + "learning_rate": 1.9984027024365596e-07, + "loss": 0.7193, + "step": 1025 + }, + { + "epoch": 1.9, + "grad_norm": 2.353931026811784, + "learning_rate": 1.9983942197426268e-07, + "loss": 0.782, + "step": 1026 + }, + { + "epoch": 1.901851851851852, + "grad_norm": 1.323413112380447, + "learning_rate": 1.998385714602081e-07, + "loss": 0.7608, + "step": 1027 + }, + { + "epoch": 1.9037037037037037, + "grad_norm": 1.349649237885736, + "learning_rate": 1.998377187015112e-07, + "loss": 0.7525, + "step": 1028 + }, + { + "epoch": 1.9055555555555554, + "grad_norm": 1.5133026470740185, + "learning_rate": 1.9983686369819126e-07, + "loss": 0.7252, + "step": 1029 + }, + { + "epoch": 1.9074074074074074, + "grad_norm": 1.529895027913531, + "learning_rate": 1.998360064502675e-07, + "loss": 0.7322, + "step": 1030 + }, + { + "epoch": 1.9092592592592592, + "grad_norm": 1.3683455309371708, + "learning_rate": 1.9983514695775915e-07, + "loss": 0.7654, + "step": 1031 + }, + { + "epoch": 1.911111111111111, + "grad_norm": 1.386781752640885, + "learning_rate": 1.9983428522068553e-07, + "loss": 0.7802, + "step": 1032 + }, + { + "epoch": 1.912962962962963, + "grad_norm": 7.267501283575071, + "learning_rate": 1.9983342123906603e-07, + "loss": 0.7591, + "step": 1033 + }, + { + "epoch": 1.914814814814815, + "grad_norm": 1.5374512127495408, + "learning_rate": 1.9983255501292012e-07, + "loss": 0.7546, + "step": 1034 + }, + { + "epoch": 1.9166666666666665, + "grad_norm": 1.3894365469249104, + "learning_rate": 1.998316865422672e-07, + "loss": 0.7425, + "step": 1035 + }, + { + "epoch": 1.9185185185185185, + "grad_norm": 1.3874291485993693, + "learning_rate": 1.9983081582712683e-07, + "loss": 0.7832, + "step": 1036 + }, + { + "epoch": 1.9203703703703705, + "grad_norm": 1.3064464904357003, + "learning_rate": 1.9982994286751858e-07, + "loss": 0.7978, + "step": 1037 + }, + { + "epoch": 1.9222222222222223, + "grad_norm": 1.582241919580081, + "learning_rate": 1.9982906766346205e-07, + "loss": 0.7479, + "step": 1038 + }, + { + "epoch": 1.924074074074074, + "grad_norm": 1.3279186913188272, + "learning_rate": 1.99828190214977e-07, + "loss": 0.7035, + "step": 1039 + }, + { + "epoch": 1.925925925925926, + "grad_norm": 1.3699480195385578, + "learning_rate": 1.9982731052208307e-07, + "loss": 0.7945, + "step": 1040 + }, + { + "epoch": 1.9277777777777778, + "grad_norm": 1.7259479474815946, + "learning_rate": 1.998264285848001e-07, + "loss": 0.7482, + "step": 1041 + }, + { + "epoch": 1.9296296296296296, + "grad_norm": 1.4844346643748427, + "learning_rate": 1.998255444031479e-07, + "loss": 0.7675, + "step": 1042 + }, + { + "epoch": 1.9314814814814816, + "grad_norm": 1.3975117459361475, + "learning_rate": 1.9982465797714632e-07, + "loss": 0.7678, + "step": 1043 + }, + { + "epoch": 1.9333333333333333, + "grad_norm": 1.3202089517147364, + "learning_rate": 1.998237693068153e-07, + "loss": 0.819, + "step": 1044 + }, + { + "epoch": 1.9351851851851851, + "grad_norm": 1.299947303921116, + "learning_rate": 1.9982287839217483e-07, + "loss": 0.7774, + "step": 1045 + }, + { + "epoch": 1.9370370370370371, + "grad_norm": 1.4987729976260935, + "learning_rate": 1.9982198523324497e-07, + "loss": 0.7579, + "step": 1046 + }, + { + "epoch": 1.9388888888888889, + "grad_norm": 1.5560900777954654, + "learning_rate": 1.9982108983004573e-07, + "loss": 0.7562, + "step": 1047 + }, + { + "epoch": 1.9407407407407407, + "grad_norm": 1.3422556176886826, + "learning_rate": 1.9982019218259728e-07, + "loss": 0.7272, + "step": 1048 + }, + { + "epoch": 1.9425925925925926, + "grad_norm": 1.478719884379062, + "learning_rate": 1.9981929229091982e-07, + "loss": 0.7812, + "step": 1049 + }, + { + "epoch": 1.9444444444444444, + "grad_norm": 1.2455833949829016, + "learning_rate": 1.9981839015503356e-07, + "loss": 0.73, + "step": 1050 + }, + { + "epoch": 1.9462962962962962, + "grad_norm": 1.6427296811374335, + "learning_rate": 1.9981748577495882e-07, + "loss": 0.7531, + "step": 1051 + }, + { + "epoch": 1.9481481481481482, + "grad_norm": 1.3744901950496324, + "learning_rate": 1.9981657915071586e-07, + "loss": 0.753, + "step": 1052 + }, + { + "epoch": 1.95, + "grad_norm": 1.5270359146528198, + "learning_rate": 1.9981567028232512e-07, + "loss": 0.7574, + "step": 1053 + }, + { + "epoch": 1.9518518518518517, + "grad_norm": 1.4389730046470506, + "learning_rate": 1.99814759169807e-07, + "loss": 0.7533, + "step": 1054 + }, + { + "epoch": 1.9537037037037037, + "grad_norm": 1.3096995681968853, + "learning_rate": 1.99813845813182e-07, + "loss": 0.7648, + "step": 1055 + }, + { + "epoch": 1.9555555555555557, + "grad_norm": 1.4176783980098466, + "learning_rate": 1.9981293021247067e-07, + "loss": 0.7046, + "step": 1056 + }, + { + "epoch": 1.9574074074074073, + "grad_norm": 1.2281733478512156, + "learning_rate": 1.998120123676936e-07, + "loss": 0.7684, + "step": 1057 + }, + { + "epoch": 1.9592592592592593, + "grad_norm": 1.2189469708483014, + "learning_rate": 1.998110922788714e-07, + "loss": 0.7526, + "step": 1058 + }, + { + "epoch": 1.9611111111111112, + "grad_norm": 1.4521186933405126, + "learning_rate": 1.998101699460247e-07, + "loss": 0.7919, + "step": 1059 + }, + { + "epoch": 1.9629629629629628, + "grad_norm": 1.4628029193341177, + "learning_rate": 1.9980924536917436e-07, + "loss": 0.7534, + "step": 1060 + }, + { + "epoch": 1.9648148148148148, + "grad_norm": 1.2616788752733763, + "learning_rate": 1.998083185483411e-07, + "loss": 0.7254, + "step": 1061 + }, + { + "epoch": 1.9666666666666668, + "grad_norm": 1.3730086237523185, + "learning_rate": 1.9980738948354574e-07, + "loss": 0.7416, + "step": 1062 + }, + { + "epoch": 1.9685185185185186, + "grad_norm": 1.310619553346444, + "learning_rate": 1.998064581748092e-07, + "loss": 0.7976, + "step": 1063 + }, + { + "epoch": 1.9703703703703703, + "grad_norm": 1.4891024950328993, + "learning_rate": 1.9980552462215243e-07, + "loss": 0.8111, + "step": 1064 + }, + { + "epoch": 1.9722222222222223, + "grad_norm": 2.7226386028805214, + "learning_rate": 1.9980458882559642e-07, + "loss": 0.7614, + "step": 1065 + }, + { + "epoch": 1.974074074074074, + "grad_norm": 1.791562733109985, + "learning_rate": 1.9980365078516212e-07, + "loss": 0.7618, + "step": 1066 + }, + { + "epoch": 1.9759259259259259, + "grad_norm": 1.4722941810114145, + "learning_rate": 1.9980271050087074e-07, + "loss": 0.7998, + "step": 1067 + }, + { + "epoch": 1.9777777777777779, + "grad_norm": 1.421312923217099, + "learning_rate": 1.9980176797274334e-07, + "loss": 0.7356, + "step": 1068 + }, + { + "epoch": 1.9796296296296296, + "grad_norm": 4.404404168548993, + "learning_rate": 1.9980082320080115e-07, + "loss": 0.7729, + "step": 1069 + }, + { + "epoch": 1.9814814814814814, + "grad_norm": 2.807841897993627, + "learning_rate": 1.997998761850654e-07, + "loss": 0.7031, + "step": 1070 + }, + { + "epoch": 1.9833333333333334, + "grad_norm": 1.2146651225346319, + "learning_rate": 1.9979892692555737e-07, + "loss": 0.7458, + "step": 1071 + }, + { + "epoch": 1.9851851851851852, + "grad_norm": 1.6151498113086764, + "learning_rate": 1.9979797542229843e-07, + "loss": 0.7849, + "step": 1072 + }, + { + "epoch": 1.987037037037037, + "grad_norm": 1.9101424059321606, + "learning_rate": 1.9979702167530993e-07, + "loss": 0.7276, + "step": 1073 + }, + { + "epoch": 1.988888888888889, + "grad_norm": 1.8527770325311863, + "learning_rate": 1.9979606568461337e-07, + "loss": 0.7872, + "step": 1074 + }, + { + "epoch": 1.9907407407407407, + "grad_norm": 1.6534336218889853, + "learning_rate": 1.997951074502302e-07, + "loss": 0.715, + "step": 1075 + }, + { + "epoch": 1.9925925925925925, + "grad_norm": 1.5402678844152873, + "learning_rate": 1.9979414697218198e-07, + "loss": 0.7416, + "step": 1076 + }, + { + "epoch": 1.9944444444444445, + "grad_norm": 1.180553748820107, + "learning_rate": 1.997931842504903e-07, + "loss": 0.7176, + "step": 1077 + }, + { + "epoch": 1.9962962962962965, + "grad_norm": 1.308310925597982, + "learning_rate": 1.9979221928517677e-07, + "loss": 0.8047, + "step": 1078 + }, + { + "epoch": 1.998148148148148, + "grad_norm": 1.4086706085230458, + "learning_rate": 1.9979125207626315e-07, + "loss": 0.7973, + "step": 1079 + }, + { + "epoch": 2.0, + "grad_norm": 1.439532492057802, + "learning_rate": 1.9979028262377116e-07, + "loss": 0.7057, + "step": 1080 + }, + { + "epoch": 2.001851851851852, + "grad_norm": 3.742312300553071, + "learning_rate": 1.9978931092772258e-07, + "loss": 0.769, + "step": 1081 + }, + { + "epoch": 2.0037037037037035, + "grad_norm": 1.8585457575425635, + "learning_rate": 1.9978833698813926e-07, + "loss": 0.7508, + "step": 1082 + }, + { + "epoch": 2.0055555555555555, + "grad_norm": 1.5030387887218244, + "learning_rate": 1.997873608050431e-07, + "loss": 0.7611, + "step": 1083 + }, + { + "epoch": 2.0074074074074075, + "grad_norm": 1.300856313413514, + "learning_rate": 1.9978638237845606e-07, + "loss": 0.7669, + "step": 1084 + }, + { + "epoch": 2.009259259259259, + "grad_norm": 1.8991658561495355, + "learning_rate": 1.9978540170840016e-07, + "loss": 0.7982, + "step": 1085 + }, + { + "epoch": 2.011111111111111, + "grad_norm": 1.3891993955211341, + "learning_rate": 1.9978441879489737e-07, + "loss": 0.789, + "step": 1086 + }, + { + "epoch": 2.012962962962963, + "grad_norm": 1.2713680021678209, + "learning_rate": 1.9978343363796983e-07, + "loss": 0.7645, + "step": 1087 + }, + { + "epoch": 2.0148148148148146, + "grad_norm": 1.6233373509493494, + "learning_rate": 1.9978244623763974e-07, + "loss": 0.7466, + "step": 1088 + }, + { + "epoch": 2.0166666666666666, + "grad_norm": 1.2854174566083847, + "learning_rate": 1.9978145659392923e-07, + "loss": 0.7605, + "step": 1089 + }, + { + "epoch": 2.0185185185185186, + "grad_norm": 1.5034516341365665, + "learning_rate": 1.9978046470686056e-07, + "loss": 0.7603, + "step": 1090 + }, + { + "epoch": 2.0203703703703706, + "grad_norm": 1.3607129822819393, + "learning_rate": 1.9977947057645604e-07, + "loss": 0.7587, + "step": 1091 + }, + { + "epoch": 2.022222222222222, + "grad_norm": 1.4252312779020693, + "learning_rate": 1.9977847420273805e-07, + "loss": 0.7339, + "step": 1092 + }, + { + "epoch": 2.024074074074074, + "grad_norm": 2.2850255266168475, + "learning_rate": 1.9977747558572898e-07, + "loss": 0.7495, + "step": 1093 + }, + { + "epoch": 2.025925925925926, + "grad_norm": 1.280053501160805, + "learning_rate": 1.9977647472545121e-07, + "loss": 0.7192, + "step": 1094 + }, + { + "epoch": 2.0277777777777777, + "grad_norm": 1.357419608222813, + "learning_rate": 1.9977547162192737e-07, + "loss": 0.7597, + "step": 1095 + }, + { + "epoch": 2.0296296296296297, + "grad_norm": 1.34589559526732, + "learning_rate": 1.9977446627517988e-07, + "loss": 0.729, + "step": 1096 + }, + { + "epoch": 2.0314814814814817, + "grad_norm": 4.632316368137749, + "learning_rate": 1.9977345868523144e-07, + "loss": 0.7543, + "step": 1097 + }, + { + "epoch": 2.033333333333333, + "grad_norm": 1.3253998226400878, + "learning_rate": 1.9977244885210464e-07, + "loss": 0.7171, + "step": 1098 + }, + { + "epoch": 2.035185185185185, + "grad_norm": 1.3671055208615106, + "learning_rate": 1.9977143677582222e-07, + "loss": 0.7301, + "step": 1099 + }, + { + "epoch": 2.037037037037037, + "grad_norm": 1.7310213780935957, + "learning_rate": 1.9977042245640697e-07, + "loss": 0.7895, + "step": 1100 + }, + { + "epoch": 2.0388888888888888, + "grad_norm": 2.7648618399315814, + "learning_rate": 1.997694058938816e-07, + "loss": 0.7887, + "step": 1101 + }, + { + "epoch": 2.0407407407407407, + "grad_norm": 1.4513820653490532, + "learning_rate": 1.9976838708826903e-07, + "loss": 0.7843, + "step": 1102 + }, + { + "epoch": 2.0425925925925927, + "grad_norm": 1.462192271315431, + "learning_rate": 1.9976736603959217e-07, + "loss": 0.6909, + "step": 1103 + }, + { + "epoch": 2.0444444444444443, + "grad_norm": 1.5338282090107127, + "learning_rate": 1.997663427478739e-07, + "loss": 0.6944, + "step": 1104 + }, + { + "epoch": 2.0462962962962963, + "grad_norm": 1.4166834039190126, + "learning_rate": 1.9976531721313737e-07, + "loss": 0.7619, + "step": 1105 + }, + { + "epoch": 2.0481481481481483, + "grad_norm": 1.3709847244996538, + "learning_rate": 1.9976428943540549e-07, + "loss": 0.7621, + "step": 1106 + }, + { + "epoch": 2.05, + "grad_norm": 3.404650291829635, + "learning_rate": 1.9976325941470145e-07, + "loss": 0.7368, + "step": 1107 + }, + { + "epoch": 2.051851851851852, + "grad_norm": 2.3702115184846546, + "learning_rate": 1.9976222715104836e-07, + "loss": 0.7805, + "step": 1108 + }, + { + "epoch": 2.053703703703704, + "grad_norm": 2.6013403883405126, + "learning_rate": 1.9976119264446946e-07, + "loss": 0.7705, + "step": 1109 + }, + { + "epoch": 2.0555555555555554, + "grad_norm": 1.4160560317452116, + "learning_rate": 1.9976015589498805e-07, + "loss": 0.7694, + "step": 1110 + }, + { + "epoch": 2.0574074074074074, + "grad_norm": 1.5769278435884768, + "learning_rate": 1.9975911690262733e-07, + "loss": 0.778, + "step": 1111 + }, + { + "epoch": 2.0592592592592593, + "grad_norm": 1.500735852563375, + "learning_rate": 1.9975807566741073e-07, + "loss": 0.7356, + "step": 1112 + }, + { + "epoch": 2.061111111111111, + "grad_norm": 1.5896967539997306, + "learning_rate": 1.997570321893617e-07, + "loss": 0.765, + "step": 1113 + }, + { + "epoch": 2.062962962962963, + "grad_norm": 1.3020079809979694, + "learning_rate": 1.997559864685036e-07, + "loss": 0.7594, + "step": 1114 + }, + { + "epoch": 2.064814814814815, + "grad_norm": 1.4196410980954213, + "learning_rate": 1.9975493850486003e-07, + "loss": 0.7762, + "step": 1115 + }, + { + "epoch": 2.066666666666667, + "grad_norm": 1.5604632360783146, + "learning_rate": 1.9975388829845446e-07, + "loss": 0.7745, + "step": 1116 + }, + { + "epoch": 2.0685185185185184, + "grad_norm": 1.4043201527727218, + "learning_rate": 1.997528358493106e-07, + "loss": 0.742, + "step": 1117 + }, + { + "epoch": 2.0703703703703704, + "grad_norm": 1.3973675755732033, + "learning_rate": 1.9975178115745204e-07, + "loss": 0.714, + "step": 1118 + }, + { + "epoch": 2.0722222222222224, + "grad_norm": 1.2952777302609368, + "learning_rate": 1.9975072422290252e-07, + "loss": 0.7583, + "step": 1119 + }, + { + "epoch": 2.074074074074074, + "grad_norm": 1.556133467251769, + "learning_rate": 1.997496650456858e-07, + "loss": 0.7359, + "step": 1120 + }, + { + "epoch": 2.075925925925926, + "grad_norm": 3.4386429960327054, + "learning_rate": 1.997486036258257e-07, + "loss": 0.7409, + "step": 1121 + }, + { + "epoch": 2.077777777777778, + "grad_norm": 1.4861788414353911, + "learning_rate": 1.997475399633461e-07, + "loss": 0.7835, + "step": 1122 + }, + { + "epoch": 2.0796296296296295, + "grad_norm": 1.388262414056101, + "learning_rate": 1.9974647405827088e-07, + "loss": 0.7216, + "step": 1123 + }, + { + "epoch": 2.0814814814814815, + "grad_norm": 1.2440384428919482, + "learning_rate": 1.9974540591062398e-07, + "loss": 0.6982, + "step": 1124 + }, + { + "epoch": 2.0833333333333335, + "grad_norm": 1.631293916240114, + "learning_rate": 1.9974433552042947e-07, + "loss": 0.7499, + "step": 1125 + }, + { + "epoch": 2.085185185185185, + "grad_norm": 1.298597813965211, + "learning_rate": 1.9974326288771143e-07, + "loss": 0.7482, + "step": 1126 + }, + { + "epoch": 2.087037037037037, + "grad_norm": 2.0400880225703513, + "learning_rate": 1.997421880124939e-07, + "loss": 0.7982, + "step": 1127 + }, + { + "epoch": 2.088888888888889, + "grad_norm": 1.7138634485578874, + "learning_rate": 1.997411108948011e-07, + "loss": 0.7892, + "step": 1128 + }, + { + "epoch": 2.0907407407407406, + "grad_norm": 1.435262908773315, + "learning_rate": 1.9974003153465725e-07, + "loss": 0.7813, + "step": 1129 + }, + { + "epoch": 2.0925925925925926, + "grad_norm": 1.3592285851029247, + "learning_rate": 1.9973894993208658e-07, + "loss": 0.7445, + "step": 1130 + }, + { + "epoch": 2.0944444444444446, + "grad_norm": 2.179842417002354, + "learning_rate": 1.9973786608711342e-07, + "loss": 0.7464, + "step": 1131 + }, + { + "epoch": 2.096296296296296, + "grad_norm": 1.2926719020528166, + "learning_rate": 1.9973677999976218e-07, + "loss": 0.8023, + "step": 1132 + }, + { + "epoch": 2.098148148148148, + "grad_norm": 1.6511375747136297, + "learning_rate": 1.9973569167005722e-07, + "loss": 0.7284, + "step": 1133 + }, + { + "epoch": 2.1, + "grad_norm": 1.6715280736077083, + "learning_rate": 1.9973460109802304e-07, + "loss": 0.6956, + "step": 1134 + }, + { + "epoch": 2.1018518518518516, + "grad_norm": 1.6447842784651854, + "learning_rate": 1.9973350828368414e-07, + "loss": 0.754, + "step": 1135 + }, + { + "epoch": 2.1037037037037036, + "grad_norm": 1.3897705242109888, + "learning_rate": 1.9973241322706512e-07, + "loss": 0.7037, + "step": 1136 + }, + { + "epoch": 2.1055555555555556, + "grad_norm": 2.161984534013614, + "learning_rate": 1.997313159281906e-07, + "loss": 0.7844, + "step": 1137 + }, + { + "epoch": 2.1074074074074076, + "grad_norm": 1.4229790022471303, + "learning_rate": 1.9973021638708518e-07, + "loss": 0.7621, + "step": 1138 + }, + { + "epoch": 2.109259259259259, + "grad_norm": 1.3590593506100692, + "learning_rate": 1.9972911460377365e-07, + "loss": 0.7357, + "step": 1139 + }, + { + "epoch": 2.111111111111111, + "grad_norm": 1.408787627489393, + "learning_rate": 1.9972801057828079e-07, + "loss": 0.7683, + "step": 1140 + }, + { + "epoch": 2.112962962962963, + "grad_norm": 1.2304459761898376, + "learning_rate": 1.9972690431063138e-07, + "loss": 0.738, + "step": 1141 + }, + { + "epoch": 2.1148148148148147, + "grad_norm": 3.5184333962917296, + "learning_rate": 1.997257958008503e-07, + "loss": 0.7544, + "step": 1142 + }, + { + "epoch": 2.1166666666666667, + "grad_norm": 1.2813847531083853, + "learning_rate": 1.997246850489625e-07, + "loss": 0.734, + "step": 1143 + }, + { + "epoch": 2.1185185185185187, + "grad_norm": 1.763551807969625, + "learning_rate": 1.997235720549929e-07, + "loss": 0.7159, + "step": 1144 + }, + { + "epoch": 2.1203703703703702, + "grad_norm": 1.330678602342504, + "learning_rate": 1.9972245681896657e-07, + "loss": 0.7259, + "step": 1145 + }, + { + "epoch": 2.1222222222222222, + "grad_norm": 1.506622127641233, + "learning_rate": 1.9972133934090858e-07, + "loss": 0.7881, + "step": 1146 + }, + { + "epoch": 2.1240740740740742, + "grad_norm": 2.4467463972573182, + "learning_rate": 1.9972021962084401e-07, + "loss": 0.7291, + "step": 1147 + }, + { + "epoch": 2.1259259259259258, + "grad_norm": 3.60433915632854, + "learning_rate": 1.997190976587981e-07, + "loss": 0.7796, + "step": 1148 + }, + { + "epoch": 2.1277777777777778, + "grad_norm": 1.4115383709763014, + "learning_rate": 1.9971797345479605e-07, + "loss": 0.7447, + "step": 1149 + }, + { + "epoch": 2.1296296296296298, + "grad_norm": 1.681619890488606, + "learning_rate": 1.9971684700886308e-07, + "loss": 0.7545, + "step": 1150 + }, + { + "epoch": 2.1314814814814813, + "grad_norm": 2.0417354930949325, + "learning_rate": 1.997157183210246e-07, + "loss": 0.6984, + "step": 1151 + }, + { + "epoch": 2.1333333333333333, + "grad_norm": 1.9867810299370199, + "learning_rate": 1.9971458739130595e-07, + "loss": 0.7509, + "step": 1152 + }, + { + "epoch": 2.1351851851851853, + "grad_norm": 1.295106427854352, + "learning_rate": 1.9971345421973256e-07, + "loss": 0.7768, + "step": 1153 + }, + { + "epoch": 2.137037037037037, + "grad_norm": 1.5979162158023175, + "learning_rate": 1.997123188063299e-07, + "loss": 0.7344, + "step": 1154 + }, + { + "epoch": 2.138888888888889, + "grad_norm": 2.377197014235787, + "learning_rate": 1.9971118115112344e-07, + "loss": 0.7425, + "step": 1155 + }, + { + "epoch": 2.140740740740741, + "grad_norm": 1.2497616842818906, + "learning_rate": 1.9971004125413886e-07, + "loss": 0.7177, + "step": 1156 + }, + { + "epoch": 2.1425925925925924, + "grad_norm": 1.4003956100181234, + "learning_rate": 1.9970889911540176e-07, + "loss": 0.7516, + "step": 1157 + }, + { + "epoch": 2.1444444444444444, + "grad_norm": 1.6348103457723056, + "learning_rate": 1.9970775473493778e-07, + "loss": 0.7801, + "step": 1158 + }, + { + "epoch": 2.1462962962962964, + "grad_norm": 1.4136450512382126, + "learning_rate": 1.9970660811277268e-07, + "loss": 0.6966, + "step": 1159 + }, + { + "epoch": 2.148148148148148, + "grad_norm": 1.6142455764087282, + "learning_rate": 1.9970545924893223e-07, + "loss": 0.7495, + "step": 1160 + }, + { + "epoch": 2.15, + "grad_norm": 1.6178997612498114, + "learning_rate": 1.9970430814344228e-07, + "loss": 0.7966, + "step": 1161 + }, + { + "epoch": 2.151851851851852, + "grad_norm": 1.4232253204221887, + "learning_rate": 1.9970315479632866e-07, + "loss": 0.7422, + "step": 1162 + }, + { + "epoch": 2.153703703703704, + "grad_norm": 1.5044370898039416, + "learning_rate": 1.9970199920761735e-07, + "loss": 0.7694, + "step": 1163 + }, + { + "epoch": 2.1555555555555554, + "grad_norm": 1.3437644544948428, + "learning_rate": 1.9970084137733428e-07, + "loss": 0.7736, + "step": 1164 + }, + { + "epoch": 2.1574074074074074, + "grad_norm": 1.6890754307428033, + "learning_rate": 1.9969968130550554e-07, + "loss": 0.7808, + "step": 1165 + }, + { + "epoch": 2.1592592592592594, + "grad_norm": 1.361561129812041, + "learning_rate": 1.9969851899215716e-07, + "loss": 0.8053, + "step": 1166 + }, + { + "epoch": 2.161111111111111, + "grad_norm": 4.838929318922867, + "learning_rate": 1.9969735443731533e-07, + "loss": 0.7754, + "step": 1167 + }, + { + "epoch": 2.162962962962963, + "grad_norm": 1.5476326656490054, + "learning_rate": 1.996961876410062e-07, + "loss": 0.7765, + "step": 1168 + }, + { + "epoch": 2.164814814814815, + "grad_norm": 1.4084510628858158, + "learning_rate": 1.9969501860325597e-07, + "loss": 0.7676, + "step": 1169 + }, + { + "epoch": 2.1666666666666665, + "grad_norm": 1.4163701593789408, + "learning_rate": 1.9969384732409097e-07, + "loss": 0.7505, + "step": 1170 + }, + { + "epoch": 2.1685185185185185, + "grad_norm": 1.771755355140723, + "learning_rate": 1.996926738035375e-07, + "loss": 0.695, + "step": 1171 + }, + { + "epoch": 2.1703703703703705, + "grad_norm": 1.4669304605030435, + "learning_rate": 1.99691498041622e-07, + "loss": 0.7229, + "step": 1172 + }, + { + "epoch": 2.172222222222222, + "grad_norm": 1.3267890524408694, + "learning_rate": 1.9969032003837084e-07, + "loss": 0.7234, + "step": 1173 + }, + { + "epoch": 2.174074074074074, + "grad_norm": 1.4194547818969117, + "learning_rate": 1.9968913979381052e-07, + "loss": 0.7217, + "step": 1174 + }, + { + "epoch": 2.175925925925926, + "grad_norm": 1.5153391098990912, + "learning_rate": 1.9968795730796758e-07, + "loss": 0.7904, + "step": 1175 + }, + { + "epoch": 2.1777777777777776, + "grad_norm": 1.2940548838760708, + "learning_rate": 1.9968677258086865e-07, + "loss": 0.7094, + "step": 1176 + }, + { + "epoch": 2.1796296296296296, + "grad_norm": 1.490808723435155, + "learning_rate": 1.996855856125403e-07, + "loss": 0.7833, + "step": 1177 + }, + { + "epoch": 2.1814814814814816, + "grad_norm": 2.396912831421858, + "learning_rate": 1.9968439640300926e-07, + "loss": 0.7285, + "step": 1178 + }, + { + "epoch": 2.183333333333333, + "grad_norm": 2.6549758199448843, + "learning_rate": 1.9968320495230223e-07, + "loss": 0.6962, + "step": 1179 + }, + { + "epoch": 2.185185185185185, + "grad_norm": 1.5147539697111214, + "learning_rate": 1.9968201126044605e-07, + "loss": 0.7599, + "step": 1180 + }, + { + "epoch": 2.187037037037037, + "grad_norm": 1.4378491090761631, + "learning_rate": 1.996808153274675e-07, + "loss": 0.7822, + "step": 1181 + }, + { + "epoch": 2.188888888888889, + "grad_norm": 1.4882275792897923, + "learning_rate": 1.996796171533935e-07, + "loss": 0.789, + "step": 1182 + }, + { + "epoch": 2.1907407407407407, + "grad_norm": 1.4527471550256748, + "learning_rate": 1.99678416738251e-07, + "loss": 0.7387, + "step": 1183 + }, + { + "epoch": 2.1925925925925926, + "grad_norm": 1.4190799171358848, + "learning_rate": 1.9967721408206694e-07, + "loss": 0.6648, + "step": 1184 + }, + { + "epoch": 2.1944444444444446, + "grad_norm": 1.583024267678944, + "learning_rate": 1.9967600918486838e-07, + "loss": 0.7173, + "step": 1185 + }, + { + "epoch": 2.196296296296296, + "grad_norm": 1.512672592951793, + "learning_rate": 1.9967480204668244e-07, + "loss": 0.7384, + "step": 1186 + }, + { + "epoch": 2.198148148148148, + "grad_norm": 1.3230071571320559, + "learning_rate": 1.9967359266753623e-07, + "loss": 0.7779, + "step": 1187 + }, + { + "epoch": 2.2, + "grad_norm": 3.527666430187374, + "learning_rate": 1.9967238104745695e-07, + "loss": 0.7709, + "step": 1188 + }, + { + "epoch": 2.2018518518518517, + "grad_norm": 1.5321498753215999, + "learning_rate": 1.996711671864718e-07, + "loss": 0.7489, + "step": 1189 + }, + { + "epoch": 2.2037037037037037, + "grad_norm": 1.62265937753039, + "learning_rate": 1.9966995108460816e-07, + "loss": 0.689, + "step": 1190 + }, + { + "epoch": 2.2055555555555557, + "grad_norm": 1.4880134468881105, + "learning_rate": 1.996687327418933e-07, + "loss": 0.7238, + "step": 1191 + }, + { + "epoch": 2.2074074074074073, + "grad_norm": 1.6630532477636921, + "learning_rate": 1.9966751215835465e-07, + "loss": 0.75, + "step": 1192 + }, + { + "epoch": 2.2092592592592593, + "grad_norm": 1.9684648966364346, + "learning_rate": 1.996662893340196e-07, + "loss": 0.7486, + "step": 1193 + }, + { + "epoch": 2.2111111111111112, + "grad_norm": 1.4030132391334154, + "learning_rate": 1.996650642689157e-07, + "loss": 0.7315, + "step": 1194 + }, + { + "epoch": 2.212962962962963, + "grad_norm": 1.911094077117738, + "learning_rate": 1.9966383696307045e-07, + "loss": 0.7229, + "step": 1195 + }, + { + "epoch": 2.214814814814815, + "grad_norm": 1.5840556545521352, + "learning_rate": 1.996626074165115e-07, + "loss": 0.7349, + "step": 1196 + }, + { + "epoch": 2.216666666666667, + "grad_norm": 1.756288854094917, + "learning_rate": 1.9966137562926643e-07, + "loss": 0.7007, + "step": 1197 + }, + { + "epoch": 2.2185185185185183, + "grad_norm": 1.3073936842333558, + "learning_rate": 1.9966014160136296e-07, + "loss": 0.7532, + "step": 1198 + }, + { + "epoch": 2.2203703703703703, + "grad_norm": 4.035589368652393, + "learning_rate": 1.9965890533282887e-07, + "loss": 0.7112, + "step": 1199 + }, + { + "epoch": 2.2222222222222223, + "grad_norm": 1.4732753028362748, + "learning_rate": 1.9965766682369185e-07, + "loss": 0.7302, + "step": 1200 + }, + { + "epoch": 2.224074074074074, + "grad_norm": 2.763755114368663, + "learning_rate": 1.9965642607397987e-07, + "loss": 0.7505, + "step": 1201 + }, + { + "epoch": 2.225925925925926, + "grad_norm": 1.339919554346128, + "learning_rate": 1.996551830837207e-07, + "loss": 0.6981, + "step": 1202 + }, + { + "epoch": 2.227777777777778, + "grad_norm": 1.4443921967016284, + "learning_rate": 1.9965393785294244e-07, + "loss": 0.7439, + "step": 1203 + }, + { + "epoch": 2.2296296296296294, + "grad_norm": 1.4073117309434156, + "learning_rate": 1.9965269038167295e-07, + "loss": 0.7162, + "step": 1204 + }, + { + "epoch": 2.2314814814814814, + "grad_norm": 1.5692579691249378, + "learning_rate": 1.9965144066994033e-07, + "loss": 0.7359, + "step": 1205 + }, + { + "epoch": 2.2333333333333334, + "grad_norm": 1.5463640968234122, + "learning_rate": 1.9965018871777271e-07, + "loss": 0.6792, + "step": 1206 + }, + { + "epoch": 2.2351851851851854, + "grad_norm": 1.4976390579987924, + "learning_rate": 1.9964893452519816e-07, + "loss": 0.7765, + "step": 1207 + }, + { + "epoch": 2.237037037037037, + "grad_norm": 1.43084691544816, + "learning_rate": 1.9964767809224494e-07, + "loss": 0.7612, + "step": 1208 + }, + { + "epoch": 2.238888888888889, + "grad_norm": 1.6311366341923859, + "learning_rate": 1.996464194189413e-07, + "loss": 0.7094, + "step": 1209 + }, + { + "epoch": 2.240740740740741, + "grad_norm": 2.576922359812976, + "learning_rate": 1.9964515850531548e-07, + "loss": 0.7372, + "step": 1210 + }, + { + "epoch": 2.2425925925925925, + "grad_norm": 2.2331007759152834, + "learning_rate": 1.9964389535139588e-07, + "loss": 0.768, + "step": 1211 + }, + { + "epoch": 2.2444444444444445, + "grad_norm": 1.2417765903652613, + "learning_rate": 1.9964262995721085e-07, + "loss": 0.7452, + "step": 1212 + }, + { + "epoch": 2.2462962962962965, + "grad_norm": 1.3297827731211054, + "learning_rate": 1.9964136232278893e-07, + "loss": 0.7143, + "step": 1213 + }, + { + "epoch": 2.248148148148148, + "grad_norm": 1.4082222752683888, + "learning_rate": 1.9964009244815854e-07, + "loss": 0.734, + "step": 1214 + }, + { + "epoch": 2.25, + "grad_norm": 1.3311351221397543, + "learning_rate": 1.9963882033334824e-07, + "loss": 0.7364, + "step": 1215 + }, + { + "epoch": 2.251851851851852, + "grad_norm": 1.5331442442302905, + "learning_rate": 1.9963754597838666e-07, + "loss": 0.7347, + "step": 1216 + }, + { + "epoch": 2.2537037037037035, + "grad_norm": 1.5942800975078237, + "learning_rate": 1.9963626938330242e-07, + "loss": 0.7264, + "step": 1217 + }, + { + "epoch": 2.2555555555555555, + "grad_norm": 1.4532867896154087, + "learning_rate": 1.9963499054812422e-07, + "loss": 0.8134, + "step": 1218 + }, + { + "epoch": 2.2574074074074075, + "grad_norm": 1.4106077017962686, + "learning_rate": 1.9963370947288087e-07, + "loss": 0.7602, + "step": 1219 + }, + { + "epoch": 2.259259259259259, + "grad_norm": 1.4617468746958917, + "learning_rate": 1.996324261576011e-07, + "loss": 0.7383, + "step": 1220 + }, + { + "epoch": 2.261111111111111, + "grad_norm": 1.3703956515571203, + "learning_rate": 1.996311406023138e-07, + "loss": 0.7622, + "step": 1221 + }, + { + "epoch": 2.262962962962963, + "grad_norm": 1.4280626801500316, + "learning_rate": 1.9962985280704787e-07, + "loss": 0.7503, + "step": 1222 + }, + { + "epoch": 2.264814814814815, + "grad_norm": 1.4779680443888599, + "learning_rate": 1.996285627718323e-07, + "loss": 0.7718, + "step": 1223 + }, + { + "epoch": 2.2666666666666666, + "grad_norm": 1.4319549088731647, + "learning_rate": 1.9962727049669598e-07, + "loss": 0.7636, + "step": 1224 + }, + { + "epoch": 2.2685185185185186, + "grad_norm": 1.5659256944422904, + "learning_rate": 1.9962597598166807e-07, + "loss": 0.7127, + "step": 1225 + }, + { + "epoch": 2.2703703703703706, + "grad_norm": 1.382465822642072, + "learning_rate": 1.9962467922677763e-07, + "loss": 0.6671, + "step": 1226 + }, + { + "epoch": 2.272222222222222, + "grad_norm": 1.9137542357073167, + "learning_rate": 1.9962338023205385e-07, + "loss": 0.7209, + "step": 1227 + }, + { + "epoch": 2.274074074074074, + "grad_norm": 1.3210331280527465, + "learning_rate": 1.9962207899752588e-07, + "loss": 0.741, + "step": 1228 + }, + { + "epoch": 2.275925925925926, + "grad_norm": 1.2074600758268053, + "learning_rate": 1.9962077552322301e-07, + "loss": 0.7324, + "step": 1229 + }, + { + "epoch": 2.2777777777777777, + "grad_norm": 1.400795484828022, + "learning_rate": 1.9961946980917453e-07, + "loss": 0.7263, + "step": 1230 + }, + { + "epoch": 2.2796296296296297, + "grad_norm": 2.166761449238475, + "learning_rate": 1.9961816185540985e-07, + "loss": 0.7601, + "step": 1231 + }, + { + "epoch": 2.2814814814814817, + "grad_norm": 1.3370869307741051, + "learning_rate": 1.996168516619583e-07, + "loss": 0.7751, + "step": 1232 + }, + { + "epoch": 2.283333333333333, + "grad_norm": 1.4225726500809228, + "learning_rate": 1.9961553922884935e-07, + "loss": 0.7314, + "step": 1233 + }, + { + "epoch": 2.285185185185185, + "grad_norm": 1.5971809673079405, + "learning_rate": 1.9961422455611257e-07, + "loss": 0.691, + "step": 1234 + }, + { + "epoch": 2.287037037037037, + "grad_norm": 1.8156131681572563, + "learning_rate": 1.9961290764377745e-07, + "loss": 0.7631, + "step": 1235 + }, + { + "epoch": 2.2888888888888888, + "grad_norm": 1.8245386969298583, + "learning_rate": 1.996115884918736e-07, + "loss": 0.7428, + "step": 1236 + }, + { + "epoch": 2.2907407407407407, + "grad_norm": 1.6666003394322357, + "learning_rate": 1.9961026710043075e-07, + "loss": 0.7396, + "step": 1237 + }, + { + "epoch": 2.2925925925925927, + "grad_norm": 3.2775651045587724, + "learning_rate": 1.9960894346947855e-07, + "loss": 0.7456, + "step": 1238 + }, + { + "epoch": 2.2944444444444443, + "grad_norm": 1.313254787643138, + "learning_rate": 1.9960761759904675e-07, + "loss": 0.7371, + "step": 1239 + }, + { + "epoch": 2.2962962962962963, + "grad_norm": 1.4514041104248467, + "learning_rate": 1.9960628948916516e-07, + "loss": 0.7223, + "step": 1240 + }, + { + "epoch": 2.2981481481481483, + "grad_norm": 1.6307566339317496, + "learning_rate": 1.9960495913986367e-07, + "loss": 0.7649, + "step": 1241 + }, + { + "epoch": 2.3, + "grad_norm": 1.3757045434901785, + "learning_rate": 1.9960362655117216e-07, + "loss": 0.7395, + "step": 1242 + }, + { + "epoch": 2.301851851851852, + "grad_norm": 1.3181530481192756, + "learning_rate": 1.9960229172312064e-07, + "loss": 0.7341, + "step": 1243 + }, + { + "epoch": 2.303703703703704, + "grad_norm": 1.3159019446998637, + "learning_rate": 1.9960095465573909e-07, + "loss": 0.7108, + "step": 1244 + }, + { + "epoch": 2.3055555555555554, + "grad_norm": 1.439739292413786, + "learning_rate": 1.995996153490575e-07, + "loss": 0.7208, + "step": 1245 + }, + { + "epoch": 2.3074074074074074, + "grad_norm": 1.5639247460941688, + "learning_rate": 1.9959827380310613e-07, + "loss": 0.7648, + "step": 1246 + }, + { + "epoch": 2.3092592592592593, + "grad_norm": 1.3664602578227352, + "learning_rate": 1.99596930017915e-07, + "loss": 0.7544, + "step": 1247 + }, + { + "epoch": 2.311111111111111, + "grad_norm": 2.074099644035599, + "learning_rate": 1.995955839935144e-07, + "loss": 0.7118, + "step": 1248 + }, + { + "epoch": 2.312962962962963, + "grad_norm": 1.465785641487809, + "learning_rate": 1.995942357299346e-07, + "loss": 0.7399, + "step": 1249 + }, + { + "epoch": 2.314814814814815, + "grad_norm": 2.4772583402368995, + "learning_rate": 1.9959288522720588e-07, + "loss": 0.7644, + "step": 1250 + }, + { + "epoch": 2.3166666666666664, + "grad_norm": 1.4168197817198627, + "learning_rate": 1.995915324853586e-07, + "loss": 0.7667, + "step": 1251 + }, + { + "epoch": 2.3185185185185184, + "grad_norm": 1.315761213911486, + "learning_rate": 1.9959017750442316e-07, + "loss": 0.7354, + "step": 1252 + }, + { + "epoch": 2.3203703703703704, + "grad_norm": 1.6012055837160273, + "learning_rate": 1.9958882028443007e-07, + "loss": 0.7702, + "step": 1253 + }, + { + "epoch": 2.3222222222222224, + "grad_norm": 3.869447565213466, + "learning_rate": 1.9958746082540984e-07, + "loss": 0.7378, + "step": 1254 + }, + { + "epoch": 2.324074074074074, + "grad_norm": 1.3896457987577644, + "learning_rate": 1.99586099127393e-07, + "loss": 0.7938, + "step": 1255 + }, + { + "epoch": 2.325925925925926, + "grad_norm": 1.3285227542943618, + "learning_rate": 1.9958473519041014e-07, + "loss": 0.7564, + "step": 1256 + }, + { + "epoch": 2.327777777777778, + "grad_norm": 1.5622805090178404, + "learning_rate": 1.99583369014492e-07, + "loss": 0.7465, + "step": 1257 + }, + { + "epoch": 2.3296296296296295, + "grad_norm": 1.7133870464560872, + "learning_rate": 1.9958200059966928e-07, + "loss": 0.7404, + "step": 1258 + }, + { + "epoch": 2.3314814814814815, + "grad_norm": 1.226662033422179, + "learning_rate": 1.9958062994597272e-07, + "loss": 0.7507, + "step": 1259 + }, + { + "epoch": 2.3333333333333335, + "grad_norm": 1.3357050930059793, + "learning_rate": 1.995792570534331e-07, + "loss": 0.7348, + "step": 1260 + }, + { + "epoch": 2.335185185185185, + "grad_norm": 1.4116054751822662, + "learning_rate": 1.9957788192208137e-07, + "loss": 0.7208, + "step": 1261 + }, + { + "epoch": 2.337037037037037, + "grad_norm": 1.4944051441425161, + "learning_rate": 1.9957650455194836e-07, + "loss": 0.7459, + "step": 1262 + }, + { + "epoch": 2.338888888888889, + "grad_norm": 1.3274322361818032, + "learning_rate": 1.9957512494306513e-07, + "loss": 0.7394, + "step": 1263 + }, + { + "epoch": 2.3407407407407406, + "grad_norm": 1.2784429905748773, + "learning_rate": 1.9957374309546263e-07, + "loss": 0.7193, + "step": 1264 + }, + { + "epoch": 2.3425925925925926, + "grad_norm": 1.347064083367779, + "learning_rate": 1.9957235900917195e-07, + "loss": 0.7014, + "step": 1265 + }, + { + "epoch": 2.3444444444444446, + "grad_norm": 1.3100819833547577, + "learning_rate": 1.995709726842242e-07, + "loss": 0.7503, + "step": 1266 + }, + { + "epoch": 2.346296296296296, + "grad_norm": 2.040831965584761, + "learning_rate": 1.9956958412065056e-07, + "loss": 0.7427, + "step": 1267 + }, + { + "epoch": 2.348148148148148, + "grad_norm": 5.09687394405009, + "learning_rate": 1.9956819331848226e-07, + "loss": 0.7324, + "step": 1268 + }, + { + "epoch": 2.35, + "grad_norm": 1.436584781908511, + "learning_rate": 1.995668002777505e-07, + "loss": 0.7291, + "step": 1269 + }, + { + "epoch": 2.351851851851852, + "grad_norm": 1.390596417737228, + "learning_rate": 1.9956540499848668e-07, + "loss": 0.7284, + "step": 1270 + }, + { + "epoch": 2.3537037037037036, + "grad_norm": 2.217339272217483, + "learning_rate": 1.9956400748072214e-07, + "loss": 0.726, + "step": 1271 + }, + { + "epoch": 2.3555555555555556, + "grad_norm": 1.3019079412972838, + "learning_rate": 1.9956260772448833e-07, + "loss": 0.771, + "step": 1272 + }, + { + "epoch": 2.3574074074074076, + "grad_norm": 1.4091549947469788, + "learning_rate": 1.9956120572981666e-07, + "loss": 0.6765, + "step": 1273 + }, + { + "epoch": 2.359259259259259, + "grad_norm": 1.7272101322961766, + "learning_rate": 1.9955980149673867e-07, + "loss": 0.7287, + "step": 1274 + }, + { + "epoch": 2.361111111111111, + "grad_norm": 1.7592990089891896, + "learning_rate": 1.9955839502528597e-07, + "loss": 0.7055, + "step": 1275 + }, + { + "epoch": 2.362962962962963, + "grad_norm": 1.2522616599956227, + "learning_rate": 1.9955698631549015e-07, + "loss": 0.71, + "step": 1276 + }, + { + "epoch": 2.3648148148148147, + "grad_norm": 1.4979857231808944, + "learning_rate": 1.9955557536738284e-07, + "loss": 0.775, + "step": 1277 + }, + { + "epoch": 2.3666666666666667, + "grad_norm": 1.9229564453154884, + "learning_rate": 1.9955416218099585e-07, + "loss": 0.7172, + "step": 1278 + }, + { + "epoch": 2.3685185185185187, + "grad_norm": 1.7020988661355843, + "learning_rate": 1.9955274675636092e-07, + "loss": 0.7076, + "step": 1279 + }, + { + "epoch": 2.3703703703703702, + "grad_norm": 1.3211692785207134, + "learning_rate": 1.9955132909350983e-07, + "loss": 0.7127, + "step": 1280 + }, + { + "epoch": 2.3722222222222222, + "grad_norm": 1.3728623978053636, + "learning_rate": 1.9954990919247447e-07, + "loss": 0.7479, + "step": 1281 + }, + { + "epoch": 2.3740740740740742, + "grad_norm": 1.4129051162117727, + "learning_rate": 1.9954848705328682e-07, + "loss": 0.706, + "step": 1282 + }, + { + "epoch": 2.3759259259259258, + "grad_norm": 1.4864202567030413, + "learning_rate": 1.995470626759788e-07, + "loss": 0.7829, + "step": 1283 + }, + { + "epoch": 2.3777777777777778, + "grad_norm": 1.4322383635756901, + "learning_rate": 1.9954563606058243e-07, + "loss": 0.7504, + "step": 1284 + }, + { + "epoch": 2.3796296296296298, + "grad_norm": 1.2738809704083554, + "learning_rate": 1.9954420720712982e-07, + "loss": 0.7072, + "step": 1285 + }, + { + "epoch": 2.3814814814814813, + "grad_norm": 1.504796334831779, + "learning_rate": 1.9954277611565307e-07, + "loss": 0.7309, + "step": 1286 + }, + { + "epoch": 2.3833333333333333, + "grad_norm": 1.3528167079894706, + "learning_rate": 1.9954134278618432e-07, + "loss": 0.7315, + "step": 1287 + }, + { + "epoch": 2.3851851851851853, + "grad_norm": 1.544388573843145, + "learning_rate": 1.995399072187559e-07, + "loss": 0.7539, + "step": 1288 + }, + { + "epoch": 2.387037037037037, + "grad_norm": 1.3101024883482584, + "learning_rate": 1.995384694134e-07, + "loss": 0.7848, + "step": 1289 + }, + { + "epoch": 2.388888888888889, + "grad_norm": 2.739151201465498, + "learning_rate": 1.9953702937014892e-07, + "loss": 0.7352, + "step": 1290 + }, + { + "epoch": 2.390740740740741, + "grad_norm": 1.594452795133058, + "learning_rate": 1.9953558708903513e-07, + "loss": 0.768, + "step": 1291 + }, + { + "epoch": 2.3925925925925924, + "grad_norm": 1.4577210153692408, + "learning_rate": 1.9953414257009098e-07, + "loss": 0.7099, + "step": 1292 + }, + { + "epoch": 2.3944444444444444, + "grad_norm": 1.4523890645160162, + "learning_rate": 1.9953269581334898e-07, + "loss": 0.7724, + "step": 1293 + }, + { + "epoch": 2.3962962962962964, + "grad_norm": 2.38572112543369, + "learning_rate": 1.9953124681884166e-07, + "loss": 0.8075, + "step": 1294 + }, + { + "epoch": 2.398148148148148, + "grad_norm": 1.5545941224078266, + "learning_rate": 1.995297955866016e-07, + "loss": 0.7447, + "step": 1295 + }, + { + "epoch": 2.4, + "grad_norm": 1.3918212878154332, + "learning_rate": 1.9952834211666138e-07, + "loss": 0.7315, + "step": 1296 + }, + { + "epoch": 2.401851851851852, + "grad_norm": 1.5217648568059665, + "learning_rate": 1.9952688640905375e-07, + "loss": 0.718, + "step": 1297 + }, + { + "epoch": 2.403703703703704, + "grad_norm": 1.3679870358225215, + "learning_rate": 1.995254284638114e-07, + "loss": 0.6739, + "step": 1298 + }, + { + "epoch": 2.4055555555555554, + "grad_norm": 1.4514544719088338, + "learning_rate": 1.9952396828096707e-07, + "loss": 0.7539, + "step": 1299 + }, + { + "epoch": 2.4074074074074074, + "grad_norm": 1.5798593765026705, + "learning_rate": 1.9952250586055367e-07, + "loss": 0.6912, + "step": 1300 + }, + { + "epoch": 2.4092592592592594, + "grad_norm": 1.323997059452625, + "learning_rate": 1.9952104120260405e-07, + "loss": 0.7257, + "step": 1301 + }, + { + "epoch": 2.411111111111111, + "grad_norm": 2.0748887970357517, + "learning_rate": 1.995195743071511e-07, + "loss": 0.7233, + "step": 1302 + }, + { + "epoch": 2.412962962962963, + "grad_norm": 1.2538093316094987, + "learning_rate": 1.9951810517422783e-07, + "loss": 0.7338, + "step": 1303 + }, + { + "epoch": 2.414814814814815, + "grad_norm": 1.3034081544174396, + "learning_rate": 1.9951663380386727e-07, + "loss": 0.7689, + "step": 1304 + }, + { + "epoch": 2.4166666666666665, + "grad_norm": 1.4305553163289428, + "learning_rate": 1.995151601961025e-07, + "loss": 0.7043, + "step": 1305 + }, + { + "epoch": 2.4185185185185185, + "grad_norm": 1.3918206507554012, + "learning_rate": 1.9951368435096666e-07, + "loss": 0.6981, + "step": 1306 + }, + { + "epoch": 2.4203703703703705, + "grad_norm": 2.115482621301663, + "learning_rate": 1.995122062684929e-07, + "loss": 0.7181, + "step": 1307 + }, + { + "epoch": 2.422222222222222, + "grad_norm": 1.5353132536292329, + "learning_rate": 1.995107259487145e-07, + "loss": 0.7267, + "step": 1308 + }, + { + "epoch": 2.424074074074074, + "grad_norm": 1.3220417503313013, + "learning_rate": 1.995092433916647e-07, + "loss": 0.742, + "step": 1309 + }, + { + "epoch": 2.425925925925926, + "grad_norm": 1.735163383524067, + "learning_rate": 1.9950775859737683e-07, + "loss": 0.7582, + "step": 1310 + }, + { + "epoch": 2.4277777777777776, + "grad_norm": 1.4669934710497194, + "learning_rate": 1.995062715658843e-07, + "loss": 0.7234, + "step": 1311 + }, + { + "epoch": 2.4296296296296296, + "grad_norm": 1.3589612687031636, + "learning_rate": 1.9950478229722054e-07, + "loss": 0.7331, + "step": 1312 + }, + { + "epoch": 2.4314814814814816, + "grad_norm": 1.4510512231537043, + "learning_rate": 1.9950329079141904e-07, + "loss": 0.7162, + "step": 1313 + }, + { + "epoch": 2.4333333333333336, + "grad_norm": 6.43609263841738, + "learning_rate": 1.995017970485133e-07, + "loss": 0.7439, + "step": 1314 + }, + { + "epoch": 2.435185185185185, + "grad_norm": 1.4550032112648212, + "learning_rate": 1.995003010685369e-07, + "loss": 0.7496, + "step": 1315 + }, + { + "epoch": 2.437037037037037, + "grad_norm": 1.28730307738335, + "learning_rate": 1.994988028515235e-07, + "loss": 0.7367, + "step": 1316 + }, + { + "epoch": 2.438888888888889, + "grad_norm": 1.493038642709901, + "learning_rate": 1.994973023975068e-07, + "loss": 0.7887, + "step": 1317 + }, + { + "epoch": 2.4407407407407407, + "grad_norm": 1.6992223520238683, + "learning_rate": 1.994957997065205e-07, + "loss": 0.6892, + "step": 1318 + }, + { + "epoch": 2.4425925925925926, + "grad_norm": 1.3734996204251164, + "learning_rate": 1.9949429477859838e-07, + "loss": 0.771, + "step": 1319 + }, + { + "epoch": 2.4444444444444446, + "grad_norm": 1.3051818716362529, + "learning_rate": 1.9949278761377428e-07, + "loss": 0.689, + "step": 1320 + }, + { + "epoch": 2.446296296296296, + "grad_norm": 1.5177936272081083, + "learning_rate": 1.9949127821208212e-07, + "loss": 0.7089, + "step": 1321 + }, + { + "epoch": 2.448148148148148, + "grad_norm": 1.403616389139682, + "learning_rate": 1.994897665735558e-07, + "loss": 0.7246, + "step": 1322 + }, + { + "epoch": 2.45, + "grad_norm": 1.7081335790237482, + "learning_rate": 1.9948825269822933e-07, + "loss": 0.7574, + "step": 1323 + }, + { + "epoch": 2.4518518518518517, + "grad_norm": 2.498650256362956, + "learning_rate": 1.9948673658613675e-07, + "loss": 0.7201, + "step": 1324 + }, + { + "epoch": 2.4537037037037037, + "grad_norm": 1.3487085473802438, + "learning_rate": 1.994852182373121e-07, + "loss": 0.736, + "step": 1325 + }, + { + "epoch": 2.4555555555555557, + "grad_norm": 1.3782873827988964, + "learning_rate": 1.9948369765178955e-07, + "loss": 0.735, + "step": 1326 + }, + { + "epoch": 2.4574074074074073, + "grad_norm": 1.621600873009482, + "learning_rate": 1.994821748296033e-07, + "loss": 0.7499, + "step": 1327 + }, + { + "epoch": 2.4592592592592593, + "grad_norm": 1.3240194327308723, + "learning_rate": 1.9948064977078755e-07, + "loss": 0.7066, + "step": 1328 + }, + { + "epoch": 2.4611111111111112, + "grad_norm": 1.4321939897417713, + "learning_rate": 1.9947912247537665e-07, + "loss": 0.7502, + "step": 1329 + }, + { + "epoch": 2.462962962962963, + "grad_norm": 1.476595812814863, + "learning_rate": 1.9947759294340485e-07, + "loss": 0.6964, + "step": 1330 + }, + { + "epoch": 2.464814814814815, + "grad_norm": 1.6128769573405382, + "learning_rate": 1.9947606117490657e-07, + "loss": 0.7027, + "step": 1331 + }, + { + "epoch": 2.466666666666667, + "grad_norm": 2.436776070200861, + "learning_rate": 1.994745271699163e-07, + "loss": 0.7566, + "step": 1332 + }, + { + "epoch": 2.4685185185185183, + "grad_norm": 2.201402659636539, + "learning_rate": 1.994729909284685e-07, + "loss": 0.7294, + "step": 1333 + }, + { + "epoch": 2.4703703703703703, + "grad_norm": 1.4525911019143531, + "learning_rate": 1.9947145245059765e-07, + "loss": 0.7577, + "step": 1334 + }, + { + "epoch": 2.4722222222222223, + "grad_norm": 1.47649127251014, + "learning_rate": 1.9946991173633841e-07, + "loss": 0.7068, + "step": 1335 + }, + { + "epoch": 2.474074074074074, + "grad_norm": 1.3796124237134613, + "learning_rate": 1.9946836878572543e-07, + "loss": 0.7753, + "step": 1336 + }, + { + "epoch": 2.475925925925926, + "grad_norm": 1.3455624677273677, + "learning_rate": 1.9946682359879333e-07, + "loss": 0.7383, + "step": 1337 + }, + { + "epoch": 2.477777777777778, + "grad_norm": 1.9675019541056298, + "learning_rate": 1.994652761755769e-07, + "loss": 0.732, + "step": 1338 + }, + { + "epoch": 2.4796296296296294, + "grad_norm": 2.2608283646835274, + "learning_rate": 1.9946372651611092e-07, + "loss": 0.7165, + "step": 1339 + }, + { + "epoch": 2.4814814814814814, + "grad_norm": 1.4798478832314137, + "learning_rate": 1.9946217462043023e-07, + "loss": 0.7298, + "step": 1340 + }, + { + "epoch": 2.4833333333333334, + "grad_norm": 1.5757807776503756, + "learning_rate": 1.994606204885697e-07, + "loss": 0.7077, + "step": 1341 + }, + { + "epoch": 2.485185185185185, + "grad_norm": 1.7699428796798822, + "learning_rate": 1.9945906412056432e-07, + "loss": 0.7411, + "step": 1342 + }, + { + "epoch": 2.487037037037037, + "grad_norm": 1.5205513973035123, + "learning_rate": 1.9945750551644905e-07, + "loss": 0.7566, + "step": 1343 + }, + { + "epoch": 2.488888888888889, + "grad_norm": 1.4683735155241477, + "learning_rate": 1.9945594467625892e-07, + "loss": 0.7031, + "step": 1344 + }, + { + "epoch": 2.490740740740741, + "grad_norm": 1.665967937097463, + "learning_rate": 1.9945438160002902e-07, + "loss": 0.7377, + "step": 1345 + }, + { + "epoch": 2.4925925925925925, + "grad_norm": 1.4088919029718936, + "learning_rate": 1.9945281628779453e-07, + "loss": 0.782, + "step": 1346 + }, + { + "epoch": 2.4944444444444445, + "grad_norm": 2.5187312198219125, + "learning_rate": 1.994512487395906e-07, + "loss": 0.7331, + "step": 1347 + }, + { + "epoch": 2.4962962962962965, + "grad_norm": 1.9306301542382656, + "learning_rate": 1.9944967895545253e-07, + "loss": 0.7386, + "step": 1348 + }, + { + "epoch": 2.498148148148148, + "grad_norm": 2.3790078314720136, + "learning_rate": 1.9944810693541556e-07, + "loss": 0.7261, + "step": 1349 + }, + { + "epoch": 2.5, + "grad_norm": 1.487959192167854, + "learning_rate": 1.9944653267951503e-07, + "loss": 0.722, + "step": 1350 + }, + { + "epoch": 2.501851851851852, + "grad_norm": 1.3848128439458334, + "learning_rate": 1.9944495618778635e-07, + "loss": 0.7249, + "step": 1351 + }, + { + "epoch": 2.5037037037037035, + "grad_norm": 1.3473791688552903, + "learning_rate": 1.99443377460265e-07, + "loss": 0.737, + "step": 1352 + }, + { + "epoch": 2.5055555555555555, + "grad_norm": 1.7181077632032955, + "learning_rate": 1.994417964969864e-07, + "loss": 0.6886, + "step": 1353 + }, + { + "epoch": 2.5074074074074075, + "grad_norm": 2.25765993896238, + "learning_rate": 1.9944021329798616e-07, + "loss": 0.7461, + "step": 1354 + }, + { + "epoch": 2.5092592592592595, + "grad_norm": 3.2947546451288794, + "learning_rate": 1.9943862786329988e-07, + "loss": 0.7655, + "step": 1355 + }, + { + "epoch": 2.511111111111111, + "grad_norm": 1.66835857253872, + "learning_rate": 1.9943704019296313e-07, + "loss": 0.7143, + "step": 1356 + }, + { + "epoch": 2.512962962962963, + "grad_norm": 1.3785416513400173, + "learning_rate": 1.9943545028701166e-07, + "loss": 0.7233, + "step": 1357 + }, + { + "epoch": 2.514814814814815, + "grad_norm": 1.2805502946902554, + "learning_rate": 1.9943385814548122e-07, + "loss": 0.755, + "step": 1358 + }, + { + "epoch": 2.5166666666666666, + "grad_norm": 1.419040120188716, + "learning_rate": 1.9943226376840755e-07, + "loss": 0.7479, + "step": 1359 + }, + { + "epoch": 2.5185185185185186, + "grad_norm": 1.4365460621979755, + "learning_rate": 1.9943066715582658e-07, + "loss": 0.6918, + "step": 1360 + }, + { + "epoch": 2.5203703703703706, + "grad_norm": 1.6644583860050746, + "learning_rate": 1.9942906830777412e-07, + "loss": 0.733, + "step": 1361 + }, + { + "epoch": 2.522222222222222, + "grad_norm": 2.8935566897237788, + "learning_rate": 1.994274672242862e-07, + "loss": 0.7312, + "step": 1362 + }, + { + "epoch": 2.524074074074074, + "grad_norm": 1.4212290242919747, + "learning_rate": 1.9942586390539877e-07, + "loss": 0.7392, + "step": 1363 + }, + { + "epoch": 2.525925925925926, + "grad_norm": 1.377153985466174, + "learning_rate": 1.994242583511479e-07, + "loss": 0.7452, + "step": 1364 + }, + { + "epoch": 2.5277777777777777, + "grad_norm": 2.3282075988119817, + "learning_rate": 1.994226505615696e-07, + "loss": 0.7243, + "step": 1365 + }, + { + "epoch": 2.5296296296296297, + "grad_norm": 1.6438608755254531, + "learning_rate": 1.9942104053670015e-07, + "loss": 0.7521, + "step": 1366 + }, + { + "epoch": 2.5314814814814817, + "grad_norm": 1.5157406787614836, + "learning_rate": 1.9941942827657567e-07, + "loss": 0.6806, + "step": 1367 + }, + { + "epoch": 2.533333333333333, + "grad_norm": 1.836797133163215, + "learning_rate": 1.994178137812324e-07, + "loss": 0.757, + "step": 1368 + }, + { + "epoch": 2.535185185185185, + "grad_norm": 1.4177530089585573, + "learning_rate": 1.9941619705070672e-07, + "loss": 0.728, + "step": 1369 + }, + { + "epoch": 2.537037037037037, + "grad_norm": 1.4183179301649709, + "learning_rate": 1.9941457808503485e-07, + "loss": 0.7261, + "step": 1370 + }, + { + "epoch": 2.5388888888888888, + "grad_norm": 1.2933395766770701, + "learning_rate": 1.994129568842533e-07, + "loss": 0.7443, + "step": 1371 + }, + { + "epoch": 2.5407407407407407, + "grad_norm": 1.377139936520184, + "learning_rate": 1.9941133344839847e-07, + "loss": 0.7305, + "step": 1372 + }, + { + "epoch": 2.5425925925925927, + "grad_norm": 1.447004496252634, + "learning_rate": 1.9940970777750685e-07, + "loss": 0.7111, + "step": 1373 + }, + { + "epoch": 2.5444444444444443, + "grad_norm": 1.4073204308846854, + "learning_rate": 1.9940807987161504e-07, + "loss": 0.7492, + "step": 1374 + }, + { + "epoch": 2.5462962962962963, + "grad_norm": 1.3087732721771723, + "learning_rate": 1.9940644973075956e-07, + "loss": 0.711, + "step": 1375 + }, + { + "epoch": 2.5481481481481483, + "grad_norm": 1.489240328186331, + "learning_rate": 1.9940481735497716e-07, + "loss": 0.7501, + "step": 1376 + }, + { + "epoch": 2.55, + "grad_norm": 1.7552615420474227, + "learning_rate": 1.9940318274430447e-07, + "loss": 0.7166, + "step": 1377 + }, + { + "epoch": 2.551851851851852, + "grad_norm": 1.4688327907631578, + "learning_rate": 1.9940154589877823e-07, + "loss": 0.6983, + "step": 1378 + }, + { + "epoch": 2.553703703703704, + "grad_norm": 1.4345992939228318, + "learning_rate": 1.993999068184353e-07, + "loss": 0.7107, + "step": 1379 + }, + { + "epoch": 2.5555555555555554, + "grad_norm": 1.962583253413522, + "learning_rate": 1.993982655033125e-07, + "loss": 0.6895, + "step": 1380 + }, + { + "epoch": 2.5574074074074074, + "grad_norm": 1.4584206018477688, + "learning_rate": 1.993966219534467e-07, + "loss": 0.7353, + "step": 1381 + }, + { + "epoch": 2.5592592592592593, + "grad_norm": 7.344289544183646, + "learning_rate": 1.9939497616887488e-07, + "loss": 0.7287, + "step": 1382 + }, + { + "epoch": 2.561111111111111, + "grad_norm": 1.606620569367819, + "learning_rate": 1.9939332814963405e-07, + "loss": 0.7304, + "step": 1383 + }, + { + "epoch": 2.562962962962963, + "grad_norm": 1.601226709074676, + "learning_rate": 1.9939167789576128e-07, + "loss": 0.748, + "step": 1384 + }, + { + "epoch": 2.564814814814815, + "grad_norm": 1.2828542326409174, + "learning_rate": 1.9939002540729365e-07, + "loss": 0.7691, + "step": 1385 + }, + { + "epoch": 2.5666666666666664, + "grad_norm": 2.049625059439289, + "learning_rate": 1.993883706842683e-07, + "loss": 0.7137, + "step": 1386 + }, + { + "epoch": 2.5685185185185184, + "grad_norm": 1.8496576138698904, + "learning_rate": 1.993867137267224e-07, + "loss": 0.6919, + "step": 1387 + }, + { + "epoch": 2.5703703703703704, + "grad_norm": 1.311962953399448, + "learning_rate": 1.993850545346933e-07, + "loss": 0.757, + "step": 1388 + }, + { + "epoch": 2.572222222222222, + "grad_norm": 1.2462191087023502, + "learning_rate": 1.993833931082182e-07, + "loss": 0.7222, + "step": 1389 + }, + { + "epoch": 2.574074074074074, + "grad_norm": 2.1964495667189095, + "learning_rate": 1.9938172944733454e-07, + "loss": 0.7353, + "step": 1390 + }, + { + "epoch": 2.575925925925926, + "grad_norm": 1.4278099404405515, + "learning_rate": 1.9938006355207966e-07, + "loss": 0.7526, + "step": 1391 + }, + { + "epoch": 2.5777777777777775, + "grad_norm": 1.4407481870984324, + "learning_rate": 1.9937839542249106e-07, + "loss": 0.7132, + "step": 1392 + }, + { + "epoch": 2.5796296296296295, + "grad_norm": 1.6406838100182555, + "learning_rate": 1.9937672505860622e-07, + "loss": 0.7211, + "step": 1393 + }, + { + "epoch": 2.5814814814814815, + "grad_norm": 1.5264800298000034, + "learning_rate": 1.993750524604627e-07, + "loss": 0.7311, + "step": 1394 + }, + { + "epoch": 2.5833333333333335, + "grad_norm": 2.2747301835962737, + "learning_rate": 1.993733776280981e-07, + "loss": 0.7087, + "step": 1395 + }, + { + "epoch": 2.585185185185185, + "grad_norm": 1.237741273066771, + "learning_rate": 1.9937170056155006e-07, + "loss": 0.7222, + "step": 1396 + }, + { + "epoch": 2.587037037037037, + "grad_norm": 1.3213607402135585, + "learning_rate": 1.993700212608563e-07, + "loss": 0.755, + "step": 1397 + }, + { + "epoch": 2.588888888888889, + "grad_norm": 1.4684444180742577, + "learning_rate": 1.993683397260546e-07, + "loss": 0.704, + "step": 1398 + }, + { + "epoch": 2.590740740740741, + "grad_norm": 1.2577211850671282, + "learning_rate": 1.9936665595718273e-07, + "loss": 0.7304, + "step": 1399 + }, + { + "epoch": 2.5925925925925926, + "grad_norm": 1.4732745439814985, + "learning_rate": 1.9936496995427857e-07, + "loss": 0.7101, + "step": 1400 + }, + { + "epoch": 2.5944444444444446, + "grad_norm": 1.4041608303881208, + "learning_rate": 1.9936328171738e-07, + "loss": 0.7039, + "step": 1401 + }, + { + "epoch": 2.5962962962962965, + "grad_norm": 1.4545056768760085, + "learning_rate": 1.99361591246525e-07, + "loss": 0.7094, + "step": 1402 + }, + { + "epoch": 2.598148148148148, + "grad_norm": 1.4346120185530165, + "learning_rate": 1.9935989854175156e-07, + "loss": 0.7105, + "step": 1403 + }, + { + "epoch": 2.6, + "grad_norm": 1.5121752691030814, + "learning_rate": 1.9935820360309775e-07, + "loss": 0.7099, + "step": 1404 + }, + { + "epoch": 2.601851851851852, + "grad_norm": 2.8549940448132385, + "learning_rate": 1.9935650643060168e-07, + "loss": 0.7345, + "step": 1405 + }, + { + "epoch": 2.6037037037037036, + "grad_norm": 1.3666763697107338, + "learning_rate": 1.993548070243015e-07, + "loss": 0.7059, + "step": 1406 + }, + { + "epoch": 2.6055555555555556, + "grad_norm": 1.4739599529807506, + "learning_rate": 1.9935310538423538e-07, + "loss": 0.739, + "step": 1407 + }, + { + "epoch": 2.6074074074074076, + "grad_norm": 1.4019653185119212, + "learning_rate": 1.9935140151044164e-07, + "loss": 0.76, + "step": 1408 + }, + { + "epoch": 2.609259259259259, + "grad_norm": 1.9540045908761232, + "learning_rate": 1.9934969540295853e-07, + "loss": 0.7597, + "step": 1409 + }, + { + "epoch": 2.611111111111111, + "grad_norm": 1.3182232284969628, + "learning_rate": 1.9934798706182448e-07, + "loss": 0.6808, + "step": 1410 + }, + { + "epoch": 2.612962962962963, + "grad_norm": 1.6506572134468076, + "learning_rate": 1.993462764870778e-07, + "loss": 0.7151, + "step": 1411 + }, + { + "epoch": 2.6148148148148147, + "grad_norm": 1.4244205286532778, + "learning_rate": 1.9934456367875706e-07, + "loss": 0.7234, + "step": 1412 + }, + { + "epoch": 2.6166666666666667, + "grad_norm": 1.5765445142615166, + "learning_rate": 1.993428486369007e-07, + "loss": 0.72, + "step": 1413 + }, + { + "epoch": 2.6185185185185187, + "grad_norm": 1.3517511430814697, + "learning_rate": 1.9934113136154726e-07, + "loss": 0.7293, + "step": 1414 + }, + { + "epoch": 2.6203703703703702, + "grad_norm": 1.4218381796670414, + "learning_rate": 1.9933941185273543e-07, + "loss": 0.714, + "step": 1415 + }, + { + "epoch": 2.6222222222222222, + "grad_norm": 1.6556829414793377, + "learning_rate": 1.9933769011050378e-07, + "loss": 0.73, + "step": 1416 + }, + { + "epoch": 2.6240740740740742, + "grad_norm": 1.6957755935795544, + "learning_rate": 1.993359661348911e-07, + "loss": 0.7166, + "step": 1417 + }, + { + "epoch": 2.6259259259259258, + "grad_norm": 1.441535191694854, + "learning_rate": 1.9933423992593607e-07, + "loss": 0.7141, + "step": 1418 + }, + { + "epoch": 2.6277777777777778, + "grad_norm": 1.647606304559682, + "learning_rate": 1.9933251148367755e-07, + "loss": 0.7395, + "step": 1419 + }, + { + "epoch": 2.6296296296296298, + "grad_norm": 1.7169659387407783, + "learning_rate": 1.993307808081544e-07, + "loss": 0.7043, + "step": 1420 + }, + { + "epoch": 2.6314814814814813, + "grad_norm": 1.648043118822608, + "learning_rate": 1.993290478994055e-07, + "loss": 0.6766, + "step": 1421 + }, + { + "epoch": 2.6333333333333333, + "grad_norm": 1.8411238108635852, + "learning_rate": 1.9932731275746984e-07, + "loss": 0.7535, + "step": 1422 + }, + { + "epoch": 2.6351851851851853, + "grad_norm": 1.3398947894551478, + "learning_rate": 1.9932557538238642e-07, + "loss": 0.7046, + "step": 1423 + }, + { + "epoch": 2.637037037037037, + "grad_norm": 1.5219901478819284, + "learning_rate": 1.9932383577419428e-07, + "loss": 0.7018, + "step": 1424 + }, + { + "epoch": 2.638888888888889, + "grad_norm": 1.8535331870555205, + "learning_rate": 1.9932209393293258e-07, + "loss": 0.6998, + "step": 1425 + }, + { + "epoch": 2.640740740740741, + "grad_norm": 1.3948215769488879, + "learning_rate": 1.9932034985864043e-07, + "loss": 0.7281, + "step": 1426 + }, + { + "epoch": 2.6425925925925924, + "grad_norm": 1.4687380050627368, + "learning_rate": 1.993186035513571e-07, + "loss": 0.6579, + "step": 1427 + }, + { + "epoch": 2.6444444444444444, + "grad_norm": 1.5444439131304921, + "learning_rate": 1.9931685501112182e-07, + "loss": 0.7495, + "step": 1428 + }, + { + "epoch": 2.6462962962962964, + "grad_norm": 1.4573829658949196, + "learning_rate": 1.9931510423797384e-07, + "loss": 0.7507, + "step": 1429 + }, + { + "epoch": 2.648148148148148, + "grad_norm": 1.5164027382102139, + "learning_rate": 1.9931335123195262e-07, + "loss": 0.729, + "step": 1430 + }, + { + "epoch": 2.65, + "grad_norm": 1.7807591373454794, + "learning_rate": 1.9931159599309755e-07, + "loss": 0.6606, + "step": 1431 + }, + { + "epoch": 2.651851851851852, + "grad_norm": 1.9537125393507286, + "learning_rate": 1.9930983852144804e-07, + "loss": 0.7524, + "step": 1432 + }, + { + "epoch": 2.6537037037037035, + "grad_norm": 1.5076329470963121, + "learning_rate": 1.9930807881704368e-07, + "loss": 0.7292, + "step": 1433 + }, + { + "epoch": 2.6555555555555554, + "grad_norm": 1.7149101274722025, + "learning_rate": 1.9930631687992396e-07, + "loss": 0.7527, + "step": 1434 + }, + { + "epoch": 2.6574074074074074, + "grad_norm": 1.9147017997919344, + "learning_rate": 1.993045527101285e-07, + "loss": 0.7356, + "step": 1435 + }, + { + "epoch": 2.659259259259259, + "grad_norm": 1.4184128796472453, + "learning_rate": 1.9930278630769704e-07, + "loss": 0.7204, + "step": 1436 + }, + { + "epoch": 2.661111111111111, + "grad_norm": 1.3583863839665171, + "learning_rate": 1.9930101767266923e-07, + "loss": 0.7351, + "step": 1437 + }, + { + "epoch": 2.662962962962963, + "grad_norm": 1.604272902725679, + "learning_rate": 1.9929924680508484e-07, + "loss": 0.7012, + "step": 1438 + }, + { + "epoch": 2.664814814814815, + "grad_norm": 1.6204049473591053, + "learning_rate": 1.9929747370498368e-07, + "loss": 0.7633, + "step": 1439 + }, + { + "epoch": 2.6666666666666665, + "grad_norm": 1.4753625913045245, + "learning_rate": 1.9929569837240564e-07, + "loss": 0.7193, + "step": 1440 + }, + { + "epoch": 2.6685185185185185, + "grad_norm": 1.3481003940160432, + "learning_rate": 1.9929392080739057e-07, + "loss": 0.7271, + "step": 1441 + }, + { + "epoch": 2.6703703703703705, + "grad_norm": 1.5428727208360309, + "learning_rate": 1.9929214100997854e-07, + "loss": 0.721, + "step": 1442 + }, + { + "epoch": 2.6722222222222225, + "grad_norm": 1.3178025410945733, + "learning_rate": 1.9929035898020948e-07, + "loss": 0.7191, + "step": 1443 + }, + { + "epoch": 2.674074074074074, + "grad_norm": 1.376741882777622, + "learning_rate": 1.9928857471812347e-07, + "loss": 0.7082, + "step": 1444 + }, + { + "epoch": 2.675925925925926, + "grad_norm": 1.3999587100507738, + "learning_rate": 1.992867882237607e-07, + "loss": 0.736, + "step": 1445 + }, + { + "epoch": 2.677777777777778, + "grad_norm": 1.4187994381577271, + "learning_rate": 1.9928499949716122e-07, + "loss": 0.688, + "step": 1446 + }, + { + "epoch": 2.6796296296296296, + "grad_norm": 1.3688063742951624, + "learning_rate": 1.9928320853836529e-07, + "loss": 0.7288, + "step": 1447 + }, + { + "epoch": 2.6814814814814816, + "grad_norm": 1.5174559167512862, + "learning_rate": 1.992814153474132e-07, + "loss": 0.7429, + "step": 1448 + }, + { + "epoch": 2.6833333333333336, + "grad_norm": 1.345651670511597, + "learning_rate": 1.9927961992434526e-07, + "loss": 0.6998, + "step": 1449 + }, + { + "epoch": 2.685185185185185, + "grad_norm": 1.2843568364875257, + "learning_rate": 1.992778222692018e-07, + "loss": 0.7112, + "step": 1450 + }, + { + "epoch": 2.687037037037037, + "grad_norm": 1.3332390141321087, + "learning_rate": 1.9927602238202326e-07, + "loss": 0.7223, + "step": 1451 + }, + { + "epoch": 2.688888888888889, + "grad_norm": 1.22297008094667, + "learning_rate": 1.9927422026285012e-07, + "loss": 0.7217, + "step": 1452 + }, + { + "epoch": 2.6907407407407407, + "grad_norm": 2.744603741519903, + "learning_rate": 1.9927241591172287e-07, + "loss": 0.6965, + "step": 1453 + }, + { + "epoch": 2.6925925925925926, + "grad_norm": 1.5452297436865128, + "learning_rate": 1.992706093286821e-07, + "loss": 0.7409, + "step": 1454 + }, + { + "epoch": 2.6944444444444446, + "grad_norm": 1.3773033529222274, + "learning_rate": 1.9926880051376841e-07, + "loss": 0.7949, + "step": 1455 + }, + { + "epoch": 2.696296296296296, + "grad_norm": 1.4404363061410415, + "learning_rate": 1.9926698946702251e-07, + "loss": 0.7191, + "step": 1456 + }, + { + "epoch": 2.698148148148148, + "grad_norm": 1.5785980737770442, + "learning_rate": 1.9926517618848506e-07, + "loss": 0.6962, + "step": 1457 + }, + { + "epoch": 2.7, + "grad_norm": 1.9138041993716404, + "learning_rate": 1.9926336067819682e-07, + "loss": 0.7411, + "step": 1458 + }, + { + "epoch": 2.7018518518518517, + "grad_norm": 1.2699927395404789, + "learning_rate": 1.9926154293619868e-07, + "loss": 0.6818, + "step": 1459 + }, + { + "epoch": 2.7037037037037037, + "grad_norm": 2.652407051472237, + "learning_rate": 1.9925972296253144e-07, + "loss": 0.7473, + "step": 1460 + }, + { + "epoch": 2.7055555555555557, + "grad_norm": 1.753958906352326, + "learning_rate": 1.9925790075723605e-07, + "loss": 0.7087, + "step": 1461 + }, + { + "epoch": 2.7074074074074073, + "grad_norm": 1.3529865864936963, + "learning_rate": 1.9925607632035345e-07, + "loss": 0.6758, + "step": 1462 + }, + { + "epoch": 2.7092592592592593, + "grad_norm": 1.37313979976493, + "learning_rate": 1.9925424965192473e-07, + "loss": 0.7214, + "step": 1463 + }, + { + "epoch": 2.7111111111111112, + "grad_norm": 1.407332563273367, + "learning_rate": 1.9925242075199088e-07, + "loss": 0.7252, + "step": 1464 + }, + { + "epoch": 2.712962962962963, + "grad_norm": 1.4053114667167617, + "learning_rate": 1.9925058962059302e-07, + "loss": 0.6868, + "step": 1465 + }, + { + "epoch": 2.714814814814815, + "grad_norm": 1.3720977567276125, + "learning_rate": 1.9924875625777238e-07, + "loss": 0.7308, + "step": 1466 + }, + { + "epoch": 2.716666666666667, + "grad_norm": 2.031860874764236, + "learning_rate": 1.9924692066357012e-07, + "loss": 0.724, + "step": 1467 + }, + { + "epoch": 2.7185185185185183, + "grad_norm": 1.6766940600295293, + "learning_rate": 1.9924508283802755e-07, + "loss": 0.7078, + "step": 1468 + }, + { + "epoch": 2.7203703703703703, + "grad_norm": 1.702889744510947, + "learning_rate": 1.9924324278118597e-07, + "loss": 0.7516, + "step": 1469 + }, + { + "epoch": 2.7222222222222223, + "grad_norm": 1.4106313576934042, + "learning_rate": 1.9924140049308673e-07, + "loss": 0.7098, + "step": 1470 + }, + { + "epoch": 2.724074074074074, + "grad_norm": 1.5248315352087267, + "learning_rate": 1.9923955597377131e-07, + "loss": 0.7323, + "step": 1471 + }, + { + "epoch": 2.725925925925926, + "grad_norm": 2.0992228365384964, + "learning_rate": 1.9923770922328112e-07, + "loss": 0.7515, + "step": 1472 + }, + { + "epoch": 2.727777777777778, + "grad_norm": 1.8377903683184047, + "learning_rate": 1.9923586024165767e-07, + "loss": 0.7468, + "step": 1473 + }, + { + "epoch": 2.7296296296296294, + "grad_norm": 2.1622808185079907, + "learning_rate": 1.9923400902894261e-07, + "loss": 0.7584, + "step": 1474 + }, + { + "epoch": 2.7314814814814814, + "grad_norm": 1.380384955132983, + "learning_rate": 1.992321555851775e-07, + "loss": 0.702, + "step": 1475 + }, + { + "epoch": 2.7333333333333334, + "grad_norm": 1.3547755070212955, + "learning_rate": 1.99230299910404e-07, + "loss": 0.6587, + "step": 1476 + }, + { + "epoch": 2.735185185185185, + "grad_norm": 1.6097112785705026, + "learning_rate": 1.992284420046639e-07, + "loss": 0.6953, + "step": 1477 + }, + { + "epoch": 2.737037037037037, + "grad_norm": 1.2901271040411313, + "learning_rate": 1.9922658186799887e-07, + "loss": 0.7578, + "step": 1478 + }, + { + "epoch": 2.738888888888889, + "grad_norm": 1.4345757814223434, + "learning_rate": 1.992247195004508e-07, + "loss": 0.6525, + "step": 1479 + }, + { + "epoch": 2.7407407407407405, + "grad_norm": 1.3629627367709394, + "learning_rate": 1.9922285490206154e-07, + "loss": 0.7624, + "step": 1480 + }, + { + "epoch": 2.7425925925925925, + "grad_norm": 1.3013276237592246, + "learning_rate": 1.9922098807287304e-07, + "loss": 0.6946, + "step": 1481 + }, + { + "epoch": 2.7444444444444445, + "grad_norm": 1.5417261700302196, + "learning_rate": 1.9921911901292722e-07, + "loss": 0.7015, + "step": 1482 + }, + { + "epoch": 2.7462962962962965, + "grad_norm": 1.429741192286043, + "learning_rate": 1.9921724772226618e-07, + "loss": 0.7385, + "step": 1483 + }, + { + "epoch": 2.748148148148148, + "grad_norm": 1.3542948761957248, + "learning_rate": 1.992153742009319e-07, + "loss": 0.6911, + "step": 1484 + }, + { + "epoch": 2.75, + "grad_norm": 1.367272080925192, + "learning_rate": 1.9921349844896654e-07, + "loss": 0.7095, + "step": 1485 + }, + { + "epoch": 2.751851851851852, + "grad_norm": 1.4296083497101486, + "learning_rate": 1.9921162046641228e-07, + "loss": 0.738, + "step": 1486 + }, + { + "epoch": 2.7537037037037035, + "grad_norm": 1.3315970825011056, + "learning_rate": 1.9920974025331138e-07, + "loss": 0.7091, + "step": 1487 + }, + { + "epoch": 2.7555555555555555, + "grad_norm": 1.4888919175325739, + "learning_rate": 1.99207857809706e-07, + "loss": 0.6967, + "step": 1488 + }, + { + "epoch": 2.7574074074074075, + "grad_norm": 1.611702710474475, + "learning_rate": 1.992059731356386e-07, + "loss": 0.6973, + "step": 1489 + }, + { + "epoch": 2.7592592592592595, + "grad_norm": 1.5556533935399386, + "learning_rate": 1.992040862311514e-07, + "loss": 0.7161, + "step": 1490 + }, + { + "epoch": 2.761111111111111, + "grad_norm": 1.888419657741896, + "learning_rate": 1.9920219709628699e-07, + "loss": 0.7144, + "step": 1491 + }, + { + "epoch": 2.762962962962963, + "grad_norm": 2.711646101270849, + "learning_rate": 1.9920030573108772e-07, + "loss": 0.7775, + "step": 1492 + }, + { + "epoch": 2.764814814814815, + "grad_norm": 1.7408632639415766, + "learning_rate": 1.9919841213559615e-07, + "loss": 0.6922, + "step": 1493 + }, + { + "epoch": 2.7666666666666666, + "grad_norm": 1.6941950561119365, + "learning_rate": 1.9919651630985487e-07, + "loss": 0.6921, + "step": 1494 + }, + { + "epoch": 2.7685185185185186, + "grad_norm": 1.3737057114402744, + "learning_rate": 1.9919461825390646e-07, + "loss": 0.6744, + "step": 1495 + }, + { + "epoch": 2.7703703703703706, + "grad_norm": 1.9139374928838895, + "learning_rate": 1.9919271796779367e-07, + "loss": 0.7123, + "step": 1496 + }, + { + "epoch": 2.772222222222222, + "grad_norm": 1.6811225389365376, + "learning_rate": 1.9919081545155916e-07, + "loss": 0.7564, + "step": 1497 + }, + { + "epoch": 2.774074074074074, + "grad_norm": 1.4332878176164237, + "learning_rate": 1.991889107052457e-07, + "loss": 0.7419, + "step": 1498 + }, + { + "epoch": 2.775925925925926, + "grad_norm": 1.4935347248276818, + "learning_rate": 1.9918700372889616e-07, + "loss": 0.7517, + "step": 1499 + }, + { + "epoch": 2.7777777777777777, + "grad_norm": 1.384083210231503, + "learning_rate": 1.9918509452255336e-07, + "loss": 0.7019, + "step": 1500 + }, + { + "epoch": 2.7796296296296297, + "grad_norm": 1.413106769413716, + "learning_rate": 1.991831830862603e-07, + "loss": 0.7631, + "step": 1501 + }, + { + "epoch": 2.7814814814814817, + "grad_norm": 1.4970556202287384, + "learning_rate": 1.9918126942005985e-07, + "loss": 0.7165, + "step": 1502 + }, + { + "epoch": 2.783333333333333, + "grad_norm": 1.7728976124533433, + "learning_rate": 1.9917935352399512e-07, + "loss": 0.7331, + "step": 1503 + }, + { + "epoch": 2.785185185185185, + "grad_norm": 1.4171460514978556, + "learning_rate": 1.991774353981092e-07, + "loss": 0.7258, + "step": 1504 + }, + { + "epoch": 2.787037037037037, + "grad_norm": 1.8826535957849944, + "learning_rate": 1.991755150424451e-07, + "loss": 0.7141, + "step": 1505 + }, + { + "epoch": 2.7888888888888888, + "grad_norm": 1.6579873958720395, + "learning_rate": 1.9917359245704607e-07, + "loss": 0.7554, + "step": 1506 + }, + { + "epoch": 2.7907407407407407, + "grad_norm": 1.489400972138541, + "learning_rate": 1.9917166764195535e-07, + "loss": 0.7218, + "step": 1507 + }, + { + "epoch": 2.7925925925925927, + "grad_norm": 2.313111244149141, + "learning_rate": 1.991697405972162e-07, + "loss": 0.7267, + "step": 1508 + }, + { + "epoch": 2.7944444444444443, + "grad_norm": 1.3952291567025221, + "learning_rate": 1.9916781132287194e-07, + "loss": 0.7303, + "step": 1509 + }, + { + "epoch": 2.7962962962962963, + "grad_norm": 1.5890406608455308, + "learning_rate": 1.9916587981896596e-07, + "loss": 0.7241, + "step": 1510 + }, + { + "epoch": 2.7981481481481483, + "grad_norm": 1.3718581660900853, + "learning_rate": 1.9916394608554166e-07, + "loss": 0.7481, + "step": 1511 + }, + { + "epoch": 2.8, + "grad_norm": 1.5160270562013836, + "learning_rate": 1.991620101226425e-07, + "loss": 0.6682, + "step": 1512 + }, + { + "epoch": 2.801851851851852, + "grad_norm": 1.3118551173038357, + "learning_rate": 1.9916007193031205e-07, + "loss": 0.7204, + "step": 1513 + }, + { + "epoch": 2.803703703703704, + "grad_norm": 1.2949838309786599, + "learning_rate": 1.9915813150859385e-07, + "loss": 0.7178, + "step": 1514 + }, + { + "epoch": 2.8055555555555554, + "grad_norm": 1.7004135791952835, + "learning_rate": 1.9915618885753158e-07, + "loss": 0.7734, + "step": 1515 + }, + { + "epoch": 2.8074074074074074, + "grad_norm": 1.352637727783801, + "learning_rate": 1.9915424397716886e-07, + "loss": 0.6849, + "step": 1516 + }, + { + "epoch": 2.8092592592592593, + "grad_norm": 1.5473451725920722, + "learning_rate": 1.9915229686754946e-07, + "loss": 0.6832, + "step": 1517 + }, + { + "epoch": 2.811111111111111, + "grad_norm": 1.3287342631789687, + "learning_rate": 1.9915034752871711e-07, + "loss": 0.69, + "step": 1518 + }, + { + "epoch": 2.812962962962963, + "grad_norm": 1.3720750580165226, + "learning_rate": 1.9914839596071566e-07, + "loss": 0.7501, + "step": 1519 + }, + { + "epoch": 2.814814814814815, + "grad_norm": 1.4232971879299827, + "learning_rate": 1.99146442163589e-07, + "loss": 0.7237, + "step": 1520 + }, + { + "epoch": 2.8166666666666664, + "grad_norm": 1.5821146734751461, + "learning_rate": 1.9914448613738103e-07, + "loss": 0.7185, + "step": 1521 + }, + { + "epoch": 2.8185185185185184, + "grad_norm": 1.4414412852138414, + "learning_rate": 1.9914252788213574e-07, + "loss": 0.6824, + "step": 1522 + }, + { + "epoch": 2.8203703703703704, + "grad_norm": 1.7144281826093233, + "learning_rate": 1.9914056739789716e-07, + "loss": 0.7105, + "step": 1523 + }, + { + "epoch": 2.822222222222222, + "grad_norm": 1.69747855090801, + "learning_rate": 1.9913860468470937e-07, + "loss": 0.7183, + "step": 1524 + }, + { + "epoch": 2.824074074074074, + "grad_norm": 1.9398439185782745, + "learning_rate": 1.9913663974261648e-07, + "loss": 0.6894, + "step": 1525 + }, + { + "epoch": 2.825925925925926, + "grad_norm": 3.6921179461241813, + "learning_rate": 1.9913467257166269e-07, + "loss": 0.6879, + "step": 1526 + }, + { + "epoch": 2.8277777777777775, + "grad_norm": 2.158776601881717, + "learning_rate": 1.991327031718922e-07, + "loss": 0.7145, + "step": 1527 + }, + { + "epoch": 2.8296296296296295, + "grad_norm": 1.3493932056090379, + "learning_rate": 1.9913073154334928e-07, + "loss": 0.7385, + "step": 1528 + }, + { + "epoch": 2.8314814814814815, + "grad_norm": 1.2785099004157228, + "learning_rate": 1.9912875768607832e-07, + "loss": 0.7401, + "step": 1529 + }, + { + "epoch": 2.8333333333333335, + "grad_norm": 1.8081616343782652, + "learning_rate": 1.9912678160012364e-07, + "loss": 0.7125, + "step": 1530 + }, + { + "epoch": 2.835185185185185, + "grad_norm": 1.5246545046562165, + "learning_rate": 1.991248032855297e-07, + "loss": 0.6957, + "step": 1531 + }, + { + "epoch": 2.837037037037037, + "grad_norm": 1.4880472415222572, + "learning_rate": 1.9912282274234094e-07, + "loss": 0.707, + "step": 1532 + }, + { + "epoch": 2.838888888888889, + "grad_norm": 1.383381944786622, + "learning_rate": 1.9912083997060192e-07, + "loss": 0.7503, + "step": 1533 + }, + { + "epoch": 2.840740740740741, + "grad_norm": 1.3895322485746548, + "learning_rate": 1.9911885497035724e-07, + "loss": 0.6786, + "step": 1534 + }, + { + "epoch": 2.8425925925925926, + "grad_norm": 1.393414344160297, + "learning_rate": 1.9911686774165145e-07, + "loss": 0.7138, + "step": 1535 + }, + { + "epoch": 2.8444444444444446, + "grad_norm": 1.4157120923346107, + "learning_rate": 1.991148782845293e-07, + "loss": 0.7342, + "step": 1536 + }, + { + "epoch": 2.8462962962962965, + "grad_norm": 1.4113867767478832, + "learning_rate": 1.991128865990355e-07, + "loss": 0.7532, + "step": 1537 + }, + { + "epoch": 2.848148148148148, + "grad_norm": 1.2641943736247099, + "learning_rate": 1.991108926852148e-07, + "loss": 0.7414, + "step": 1538 + }, + { + "epoch": 2.85, + "grad_norm": 1.3477940433471907, + "learning_rate": 1.9910889654311208e-07, + "loss": 0.7254, + "step": 1539 + }, + { + "epoch": 2.851851851851852, + "grad_norm": 1.4723486302190898, + "learning_rate": 1.9910689817277215e-07, + "loss": 0.7422, + "step": 1540 + }, + { + "epoch": 2.8537037037037036, + "grad_norm": 1.4767158522435488, + "learning_rate": 1.9910489757424004e-07, + "loss": 0.744, + "step": 1541 + }, + { + "epoch": 2.8555555555555556, + "grad_norm": 1.4095226595398418, + "learning_rate": 1.9910289474756062e-07, + "loss": 0.7678, + "step": 1542 + }, + { + "epoch": 2.8574074074074076, + "grad_norm": 1.5217756251669254, + "learning_rate": 1.9910088969277895e-07, + "loss": 0.7445, + "step": 1543 + }, + { + "epoch": 2.859259259259259, + "grad_norm": 1.3469934114055042, + "learning_rate": 1.9909888240994017e-07, + "loss": 0.7406, + "step": 1544 + }, + { + "epoch": 2.861111111111111, + "grad_norm": 1.631959340029328, + "learning_rate": 1.9909687289908935e-07, + "loss": 0.7059, + "step": 1545 + }, + { + "epoch": 2.862962962962963, + "grad_norm": 1.4958800264921208, + "learning_rate": 1.9909486116027168e-07, + "loss": 0.7058, + "step": 1546 + }, + { + "epoch": 2.8648148148148147, + "grad_norm": 1.4236333734618511, + "learning_rate": 1.990928471935324e-07, + "loss": 0.7094, + "step": 1547 + }, + { + "epoch": 2.8666666666666667, + "grad_norm": 1.5178503788104258, + "learning_rate": 1.990908309989168e-07, + "loss": 0.7443, + "step": 1548 + }, + { + "epoch": 2.8685185185185187, + "grad_norm": 1.5105425013016076, + "learning_rate": 1.9908881257647016e-07, + "loss": 0.7396, + "step": 1549 + }, + { + "epoch": 2.8703703703703702, + "grad_norm": 2.355577851446489, + "learning_rate": 1.990867919262379e-07, + "loss": 0.7447, + "step": 1550 + }, + { + "epoch": 2.8722222222222222, + "grad_norm": 1.4677595499689686, + "learning_rate": 1.9908476904826547e-07, + "loss": 0.759, + "step": 1551 + }, + { + "epoch": 2.8740740740740742, + "grad_norm": 1.552627302641364, + "learning_rate": 1.9908274394259828e-07, + "loss": 0.751, + "step": 1552 + }, + { + "epoch": 2.8759259259259258, + "grad_norm": 1.6555852580999029, + "learning_rate": 1.9908071660928193e-07, + "loss": 0.7355, + "step": 1553 + }, + { + "epoch": 2.8777777777777778, + "grad_norm": 1.3723374885598025, + "learning_rate": 1.9907868704836194e-07, + "loss": 0.7222, + "step": 1554 + }, + { + "epoch": 2.8796296296296298, + "grad_norm": 1.5900472687022837, + "learning_rate": 1.9907665525988402e-07, + "loss": 0.7413, + "step": 1555 + }, + { + "epoch": 2.8814814814814813, + "grad_norm": 3.5880722849062674, + "learning_rate": 1.9907462124389377e-07, + "loss": 0.734, + "step": 1556 + }, + { + "epoch": 2.8833333333333333, + "grad_norm": 3.1509341016854746, + "learning_rate": 1.9907258500043694e-07, + "loss": 0.7355, + "step": 1557 + }, + { + "epoch": 2.8851851851851853, + "grad_norm": 1.5360821041028956, + "learning_rate": 1.9907054652955932e-07, + "loss": 0.6681, + "step": 1558 + }, + { + "epoch": 2.887037037037037, + "grad_norm": 1.4547394731485015, + "learning_rate": 1.9906850583130677e-07, + "loss": 0.7625, + "step": 1559 + }, + { + "epoch": 2.888888888888889, + "grad_norm": 1.709178223262143, + "learning_rate": 1.9906646290572513e-07, + "loss": 0.7002, + "step": 1560 + }, + { + "epoch": 2.890740740740741, + "grad_norm": 1.2054478082821016, + "learning_rate": 1.9906441775286035e-07, + "loss": 0.6804, + "step": 1561 + }, + { + "epoch": 2.8925925925925924, + "grad_norm": 1.666213109998135, + "learning_rate": 1.9906237037275843e-07, + "loss": 0.7851, + "step": 1562 + }, + { + "epoch": 2.8944444444444444, + "grad_norm": 1.6798417204777845, + "learning_rate": 1.9906032076546532e-07, + "loss": 0.6579, + "step": 1563 + }, + { + "epoch": 2.8962962962962964, + "grad_norm": 1.3826982573769029, + "learning_rate": 1.990582689310272e-07, + "loss": 0.7176, + "step": 1564 + }, + { + "epoch": 2.898148148148148, + "grad_norm": 1.4657547018707673, + "learning_rate": 1.9905621486949013e-07, + "loss": 0.7429, + "step": 1565 + }, + { + "epoch": 2.9, + "grad_norm": 2.126039634312253, + "learning_rate": 1.9905415858090032e-07, + "loss": 0.7256, + "step": 1566 + }, + { + "epoch": 2.901851851851852, + "grad_norm": 1.3237019339778255, + "learning_rate": 1.9905210006530398e-07, + "loss": 0.7053, + "step": 1567 + }, + { + "epoch": 2.9037037037037035, + "grad_norm": 1.3647018670925397, + "learning_rate": 1.9905003932274746e-07, + "loss": 0.7146, + "step": 1568 + }, + { + "epoch": 2.9055555555555554, + "grad_norm": 1.458602935413366, + "learning_rate": 1.9904797635327702e-07, + "loss": 0.7084, + "step": 1569 + }, + { + "epoch": 2.9074074074074074, + "grad_norm": 1.3781877310261341, + "learning_rate": 1.9904591115693905e-07, + "loss": 0.7504, + "step": 1570 + }, + { + "epoch": 2.909259259259259, + "grad_norm": 2.221669754642774, + "learning_rate": 1.9904384373378e-07, + "loss": 0.7876, + "step": 1571 + }, + { + "epoch": 2.911111111111111, + "grad_norm": 1.3812586932654014, + "learning_rate": 1.9904177408384633e-07, + "loss": 0.7444, + "step": 1572 + }, + { + "epoch": 2.912962962962963, + "grad_norm": 1.6747140864461514, + "learning_rate": 1.990397022071846e-07, + "loss": 0.6335, + "step": 1573 + }, + { + "epoch": 2.914814814814815, + "grad_norm": 1.4317123943799357, + "learning_rate": 1.9903762810384138e-07, + "loss": 0.7324, + "step": 1574 + }, + { + "epoch": 2.9166666666666665, + "grad_norm": 2.0302283230660456, + "learning_rate": 1.990355517738633e-07, + "loss": 0.6766, + "step": 1575 + }, + { + "epoch": 2.9185185185185185, + "grad_norm": 1.5482722587521691, + "learning_rate": 1.9903347321729699e-07, + "loss": 0.7131, + "step": 1576 + }, + { + "epoch": 2.9203703703703705, + "grad_norm": 1.6357874902641074, + "learning_rate": 1.990313924341893e-07, + "loss": 0.6799, + "step": 1577 + }, + { + "epoch": 2.9222222222222225, + "grad_norm": 1.4018801800672824, + "learning_rate": 1.9902930942458692e-07, + "loss": 0.7061, + "step": 1578 + }, + { + "epoch": 2.924074074074074, + "grad_norm": 1.565864855508223, + "learning_rate": 1.990272241885367e-07, + "loss": 0.7363, + "step": 1579 + }, + { + "epoch": 2.925925925925926, + "grad_norm": 1.4209708460174861, + "learning_rate": 1.9902513672608552e-07, + "loss": 0.7049, + "step": 1580 + }, + { + "epoch": 2.927777777777778, + "grad_norm": 1.601211762325991, + "learning_rate": 1.9902304703728034e-07, + "loss": 0.6928, + "step": 1581 + }, + { + "epoch": 2.9296296296296296, + "grad_norm": 1.941811892938044, + "learning_rate": 1.9902095512216808e-07, + "loss": 0.7122, + "step": 1582 + }, + { + "epoch": 2.9314814814814816, + "grad_norm": 1.563309407291449, + "learning_rate": 1.9901886098079582e-07, + "loss": 0.7014, + "step": 1583 + }, + { + "epoch": 2.9333333333333336, + "grad_norm": 2.25580973584275, + "learning_rate": 1.9901676461321067e-07, + "loss": 0.7151, + "step": 1584 + }, + { + "epoch": 2.935185185185185, + "grad_norm": 1.3662637398853443, + "learning_rate": 1.990146660194597e-07, + "loss": 0.7043, + "step": 1585 + }, + { + "epoch": 2.937037037037037, + "grad_norm": 2.8500493066284487, + "learning_rate": 1.990125651995901e-07, + "loss": 0.7456, + "step": 1586 + }, + { + "epoch": 2.938888888888889, + "grad_norm": 1.4101988677031434, + "learning_rate": 1.9901046215364917e-07, + "loss": 0.74, + "step": 1587 + }, + { + "epoch": 2.9407407407407407, + "grad_norm": 1.544313036377568, + "learning_rate": 1.9900835688168408e-07, + "loss": 0.6936, + "step": 1588 + }, + { + "epoch": 2.9425925925925926, + "grad_norm": 1.4074380588460953, + "learning_rate": 1.9900624938374227e-07, + "loss": 0.658, + "step": 1589 + }, + { + "epoch": 2.9444444444444446, + "grad_norm": 1.3162581344440665, + "learning_rate": 1.990041396598711e-07, + "loss": 0.6656, + "step": 1590 + }, + { + "epoch": 2.946296296296296, + "grad_norm": 1.520048816301999, + "learning_rate": 1.9900202771011795e-07, + "loss": 0.7155, + "step": 1591 + }, + { + "epoch": 2.948148148148148, + "grad_norm": 1.4340900834721495, + "learning_rate": 1.989999135345303e-07, + "loss": 0.7333, + "step": 1592 + }, + { + "epoch": 2.95, + "grad_norm": 11.286555927192733, + "learning_rate": 1.9899779713315575e-07, + "loss": 0.7405, + "step": 1593 + }, + { + "epoch": 2.9518518518518517, + "grad_norm": 1.5293603892505057, + "learning_rate": 1.9899567850604185e-07, + "loss": 0.7062, + "step": 1594 + }, + { + "epoch": 2.9537037037037037, + "grad_norm": 1.3140294351980084, + "learning_rate": 1.9899355765323623e-07, + "loss": 0.7083, + "step": 1595 + }, + { + "epoch": 2.9555555555555557, + "grad_norm": 1.3405231484968647, + "learning_rate": 1.9899143457478655e-07, + "loss": 0.6909, + "step": 1596 + }, + { + "epoch": 2.9574074074074073, + "grad_norm": 1.3672900881444934, + "learning_rate": 1.9898930927074055e-07, + "loss": 0.712, + "step": 1597 + }, + { + "epoch": 2.9592592592592593, + "grad_norm": 1.703842596725543, + "learning_rate": 1.9898718174114605e-07, + "loss": 0.7021, + "step": 1598 + }, + { + "epoch": 2.9611111111111112, + "grad_norm": 1.7375989794196072, + "learning_rate": 1.9898505198605088e-07, + "loss": 0.758, + "step": 1599 + }, + { + "epoch": 2.962962962962963, + "grad_norm": 1.6224928533143612, + "learning_rate": 1.9898292000550289e-07, + "loss": 0.7396, + "step": 1600 + }, + { + "epoch": 2.964814814814815, + "grad_norm": 1.6303459865409737, + "learning_rate": 1.9898078579955002e-07, + "loss": 0.6431, + "step": 1601 + }, + { + "epoch": 2.966666666666667, + "grad_norm": 1.654190860525658, + "learning_rate": 1.9897864936824024e-07, + "loss": 0.6561, + "step": 1602 + }, + { + "epoch": 2.9685185185185183, + "grad_norm": 1.788151948461757, + "learning_rate": 1.989765107116216e-07, + "loss": 0.7187, + "step": 1603 + }, + { + "epoch": 2.9703703703703703, + "grad_norm": 1.4738733364351084, + "learning_rate": 1.9897436982974223e-07, + "loss": 0.7513, + "step": 1604 + }, + { + "epoch": 2.9722222222222223, + "grad_norm": 1.4176979977823627, + "learning_rate": 1.9897222672265017e-07, + "loss": 0.7311, + "step": 1605 + }, + { + "epoch": 2.974074074074074, + "grad_norm": 1.2131457107880457, + "learning_rate": 1.9897008139039366e-07, + "loss": 0.6931, + "step": 1606 + }, + { + "epoch": 2.975925925925926, + "grad_norm": 1.4790144518991266, + "learning_rate": 1.9896793383302094e-07, + "loss": 0.6802, + "step": 1607 + }, + { + "epoch": 2.977777777777778, + "grad_norm": 1.7385149317862878, + "learning_rate": 1.9896578405058025e-07, + "loss": 0.7609, + "step": 1608 + }, + { + "epoch": 2.9796296296296294, + "grad_norm": 1.3892458405453747, + "learning_rate": 1.9896363204311996e-07, + "loss": 0.7373, + "step": 1609 + }, + { + "epoch": 2.9814814814814814, + "grad_norm": 1.3627744163643796, + "learning_rate": 1.9896147781068842e-07, + "loss": 0.7432, + "step": 1610 + }, + { + "epoch": 2.9833333333333334, + "grad_norm": 1.3074709448165318, + "learning_rate": 1.989593213533341e-07, + "loss": 0.6811, + "step": 1611 + }, + { + "epoch": 2.985185185185185, + "grad_norm": 1.448861395054024, + "learning_rate": 1.9895716267110548e-07, + "loss": 0.652, + "step": 1612 + }, + { + "epoch": 2.987037037037037, + "grad_norm": 1.2823434727948166, + "learning_rate": 1.9895500176405106e-07, + "loss": 0.6876, + "step": 1613 + }, + { + "epoch": 2.988888888888889, + "grad_norm": 1.2754660886216431, + "learning_rate": 1.9895283863221945e-07, + "loss": 0.697, + "step": 1614 + }, + { + "epoch": 2.9907407407407405, + "grad_norm": 1.9914142825761545, + "learning_rate": 1.9895067327565927e-07, + "loss": 0.74, + "step": 1615 + }, + { + "epoch": 2.9925925925925925, + "grad_norm": 1.505506782590237, + "learning_rate": 1.989485056944192e-07, + "loss": 0.6696, + "step": 1616 + }, + { + "epoch": 2.9944444444444445, + "grad_norm": 3.1005063199012457, + "learning_rate": 1.98946335888548e-07, + "loss": 0.7325, + "step": 1617 + }, + { + "epoch": 2.9962962962962965, + "grad_norm": 1.348441663216238, + "learning_rate": 1.9894416385809442e-07, + "loss": 0.7448, + "step": 1618 + }, + { + "epoch": 2.998148148148148, + "grad_norm": 2.1157934108696135, + "learning_rate": 1.9894198960310736e-07, + "loss": 0.7123, + "step": 1619 + }, + { + "epoch": 3.0, + "grad_norm": 1.5400644539695503, + "learning_rate": 1.989398131236356e-07, + "loss": 0.7479, + "step": 1620 + }, + { + "epoch": 3.001851851851852, + "grad_norm": 1.3465290403646841, + "learning_rate": 1.9893763441972812e-07, + "loss": 0.7176, + "step": 1621 + }, + { + "epoch": 3.0037037037037035, + "grad_norm": 2.5116095724844363, + "learning_rate": 1.9893545349143393e-07, + "loss": 0.7012, + "step": 1622 + }, + { + "epoch": 3.0055555555555555, + "grad_norm": 1.3581760933892026, + "learning_rate": 1.9893327033880203e-07, + "loss": 0.6863, + "step": 1623 + }, + { + "epoch": 3.0074074074074075, + "grad_norm": 1.2994520920497343, + "learning_rate": 1.9893108496188155e-07, + "loss": 0.716, + "step": 1624 + }, + { + "epoch": 3.009259259259259, + "grad_norm": 1.3055433011569353, + "learning_rate": 1.9892889736072154e-07, + "loss": 0.6888, + "step": 1625 + }, + { + "epoch": 3.011111111111111, + "grad_norm": 1.2499442093256048, + "learning_rate": 1.9892670753537124e-07, + "loss": 0.6973, + "step": 1626 + }, + { + "epoch": 3.012962962962963, + "grad_norm": 1.3692448570675324, + "learning_rate": 1.9892451548587986e-07, + "loss": 0.68, + "step": 1627 + }, + { + "epoch": 3.0148148148148146, + "grad_norm": 5.254663109492828, + "learning_rate": 1.9892232121229672e-07, + "loss": 0.7396, + "step": 1628 + }, + { + "epoch": 3.0166666666666666, + "grad_norm": 1.4458957757007047, + "learning_rate": 1.9892012471467114e-07, + "loss": 0.7068, + "step": 1629 + }, + { + "epoch": 3.0185185185185186, + "grad_norm": 1.4568042426347534, + "learning_rate": 1.9891792599305245e-07, + "loss": 0.708, + "step": 1630 + }, + { + "epoch": 3.0203703703703706, + "grad_norm": 1.5898842089100773, + "learning_rate": 1.9891572504749016e-07, + "loss": 0.6825, + "step": 1631 + }, + { + "epoch": 3.022222222222222, + "grad_norm": 1.4093596408780869, + "learning_rate": 1.9891352187803373e-07, + "loss": 0.741, + "step": 1632 + }, + { + "epoch": 3.024074074074074, + "grad_norm": 1.4536944611496536, + "learning_rate": 1.9891131648473266e-07, + "loss": 0.752, + "step": 1633 + }, + { + "epoch": 3.025925925925926, + "grad_norm": 1.4395753321416864, + "learning_rate": 1.989091088676366e-07, + "loss": 0.6715, + "step": 1634 + }, + { + "epoch": 3.0277777777777777, + "grad_norm": 1.4961611160424322, + "learning_rate": 1.9890689902679507e-07, + "loss": 0.7228, + "step": 1635 + }, + { + "epoch": 3.0296296296296297, + "grad_norm": 1.4922576209506542, + "learning_rate": 1.9890468696225788e-07, + "loss": 0.7549, + "step": 1636 + }, + { + "epoch": 3.0314814814814817, + "grad_norm": 1.6955670865230463, + "learning_rate": 1.9890247267407468e-07, + "loss": 0.7276, + "step": 1637 + }, + { + "epoch": 3.033333333333333, + "grad_norm": 1.4550193857556193, + "learning_rate": 1.9890025616229526e-07, + "loss": 0.7127, + "step": 1638 + }, + { + "epoch": 3.035185185185185, + "grad_norm": 1.249933019312746, + "learning_rate": 1.9889803742696952e-07, + "loss": 0.6699, + "step": 1639 + }, + { + "epoch": 3.037037037037037, + "grad_norm": 1.5506750348497658, + "learning_rate": 1.9889581646814725e-07, + "loss": 0.7337, + "step": 1640 + }, + { + "epoch": 3.0388888888888888, + "grad_norm": 2.755592884413332, + "learning_rate": 1.9889359328587848e-07, + "loss": 0.7405, + "step": 1641 + }, + { + "epoch": 3.0407407407407407, + "grad_norm": 1.6330365000153635, + "learning_rate": 1.988913678802131e-07, + "loss": 0.73, + "step": 1642 + }, + { + "epoch": 3.0425925925925927, + "grad_norm": 1.5619368172557622, + "learning_rate": 1.988891402512012e-07, + "loss": 0.7563, + "step": 1643 + }, + { + "epoch": 3.0444444444444443, + "grad_norm": 1.4887327335824572, + "learning_rate": 1.9888691039889285e-07, + "loss": 0.7684, + "step": 1644 + }, + { + "epoch": 3.0462962962962963, + "grad_norm": 1.7746831756420167, + "learning_rate": 1.9888467832333817e-07, + "loss": 0.7693, + "step": 1645 + }, + { + "epoch": 3.0481481481481483, + "grad_norm": 1.4558601823973059, + "learning_rate": 1.9888244402458737e-07, + "loss": 0.6964, + "step": 1646 + }, + { + "epoch": 3.05, + "grad_norm": 2.730373109051131, + "learning_rate": 1.9888020750269066e-07, + "loss": 0.6868, + "step": 1647 + }, + { + "epoch": 3.051851851851852, + "grad_norm": 1.3171058015474684, + "learning_rate": 1.9887796875769835e-07, + "loss": 0.6741, + "step": 1648 + }, + { + "epoch": 3.053703703703704, + "grad_norm": 1.6828570889766399, + "learning_rate": 1.988757277896607e-07, + "loss": 0.6984, + "step": 1649 + }, + { + "epoch": 3.0555555555555554, + "grad_norm": 1.7102743906255844, + "learning_rate": 1.988734845986282e-07, + "loss": 0.7288, + "step": 1650 + }, + { + "epoch": 3.0574074074074074, + "grad_norm": 1.4648711735671105, + "learning_rate": 1.988712391846512e-07, + "loss": 0.7536, + "step": 1651 + }, + { + "epoch": 3.0592592592592593, + "grad_norm": 2.7389505677907153, + "learning_rate": 1.9886899154778023e-07, + "loss": 0.7341, + "step": 1652 + }, + { + "epoch": 3.061111111111111, + "grad_norm": 1.3866068178133961, + "learning_rate": 1.988667416880658e-07, + "loss": 0.6619, + "step": 1653 + }, + { + "epoch": 3.062962962962963, + "grad_norm": 1.542935055861403, + "learning_rate": 1.9886448960555852e-07, + "loss": 0.7048, + "step": 1654 + }, + { + "epoch": 3.064814814814815, + "grad_norm": 1.5361354972717536, + "learning_rate": 1.9886223530030897e-07, + "loss": 0.7004, + "step": 1655 + }, + { + "epoch": 3.066666666666667, + "grad_norm": 1.6880469698301295, + "learning_rate": 1.9885997877236786e-07, + "loss": 0.7209, + "step": 1656 + }, + { + "epoch": 3.0685185185185184, + "grad_norm": 1.7682470421796426, + "learning_rate": 1.9885772002178594e-07, + "loss": 0.7509, + "step": 1657 + }, + { + "epoch": 3.0703703703703704, + "grad_norm": 2.0197081624474857, + "learning_rate": 1.9885545904861401e-07, + "loss": 0.7501, + "step": 1658 + }, + { + "epoch": 3.0722222222222224, + "grad_norm": 1.3349564461482262, + "learning_rate": 1.988531958529028e-07, + "loss": 0.7179, + "step": 1659 + }, + { + "epoch": 3.074074074074074, + "grad_norm": 1.6653903828205585, + "learning_rate": 1.9885093043470334e-07, + "loss": 0.7032, + "step": 1660 + }, + { + "epoch": 3.075925925925926, + "grad_norm": 1.7863874143951282, + "learning_rate": 1.9884866279406647e-07, + "loss": 0.7324, + "step": 1661 + }, + { + "epoch": 3.077777777777778, + "grad_norm": 1.5414246900775297, + "learning_rate": 1.9884639293104318e-07, + "loss": 0.7215, + "step": 1662 + }, + { + "epoch": 3.0796296296296295, + "grad_norm": 1.3500687796087663, + "learning_rate": 1.988441208456845e-07, + "loss": 0.7436, + "step": 1663 + }, + { + "epoch": 3.0814814814814815, + "grad_norm": 1.5677478826583047, + "learning_rate": 1.9884184653804157e-07, + "loss": 0.7004, + "step": 1664 + }, + { + "epoch": 3.0833333333333335, + "grad_norm": 1.4095622928721916, + "learning_rate": 1.9883957000816545e-07, + "loss": 0.6872, + "step": 1665 + }, + { + "epoch": 3.085185185185185, + "grad_norm": 1.5182424388915532, + "learning_rate": 1.9883729125610734e-07, + "loss": 0.7276, + "step": 1666 + }, + { + "epoch": 3.087037037037037, + "grad_norm": 1.4318940099260271, + "learning_rate": 1.988350102819185e-07, + "loss": 0.6919, + "step": 1667 + }, + { + "epoch": 3.088888888888889, + "grad_norm": 1.8798896662429445, + "learning_rate": 1.988327270856502e-07, + "loss": 0.713, + "step": 1668 + }, + { + "epoch": 3.0907407407407406, + "grad_norm": 1.7396830969265804, + "learning_rate": 1.9883044166735375e-07, + "loss": 0.7096, + "step": 1669 + }, + { + "epoch": 3.0925925925925926, + "grad_norm": 1.4675500179598062, + "learning_rate": 1.9882815402708058e-07, + "loss": 0.6673, + "step": 1670 + }, + { + "epoch": 3.0944444444444446, + "grad_norm": 3.908630997338015, + "learning_rate": 1.9882586416488208e-07, + "loss": 0.6757, + "step": 1671 + }, + { + "epoch": 3.096296296296296, + "grad_norm": 1.5773921026996343, + "learning_rate": 1.9882357208080977e-07, + "loss": 0.7608, + "step": 1672 + }, + { + "epoch": 3.098148148148148, + "grad_norm": 1.5807244741734738, + "learning_rate": 1.9882127777491514e-07, + "loss": 0.7144, + "step": 1673 + }, + { + "epoch": 3.1, + "grad_norm": 1.8922916047260412, + "learning_rate": 1.9881898124724978e-07, + "loss": 0.6948, + "step": 1674 + }, + { + "epoch": 3.1018518518518516, + "grad_norm": 1.3615325912481913, + "learning_rate": 1.9881668249786533e-07, + "loss": 0.751, + "step": 1675 + }, + { + "epoch": 3.1037037037037036, + "grad_norm": 1.4705548481990303, + "learning_rate": 1.988143815268135e-07, + "loss": 0.6496, + "step": 1676 + }, + { + "epoch": 3.1055555555555556, + "grad_norm": 2.33415194489758, + "learning_rate": 1.9881207833414599e-07, + "loss": 0.7061, + "step": 1677 + }, + { + "epoch": 3.1074074074074076, + "grad_norm": 2.5969211621662427, + "learning_rate": 1.9880977291991456e-07, + "loss": 0.688, + "step": 1678 + }, + { + "epoch": 3.109259259259259, + "grad_norm": 1.5417996015192457, + "learning_rate": 1.9880746528417112e-07, + "loss": 0.6758, + "step": 1679 + }, + { + "epoch": 3.111111111111111, + "grad_norm": 3.269981042236833, + "learning_rate": 1.9880515542696746e-07, + "loss": 0.7392, + "step": 1680 + }, + { + "epoch": 3.112962962962963, + "grad_norm": 1.4054886995259201, + "learning_rate": 1.988028433483556e-07, + "loss": 0.6951, + "step": 1681 + }, + { + "epoch": 3.1148148148148147, + "grad_norm": 2.15587450450883, + "learning_rate": 1.9880052904838744e-07, + "loss": 0.672, + "step": 1682 + }, + { + "epoch": 3.1166666666666667, + "grad_norm": 1.5399876800245342, + "learning_rate": 1.9879821252711507e-07, + "loss": 0.6865, + "step": 1683 + }, + { + "epoch": 3.1185185185185187, + "grad_norm": 1.4940979279637496, + "learning_rate": 1.987958937845905e-07, + "loss": 0.7134, + "step": 1684 + }, + { + "epoch": 3.1203703703703702, + "grad_norm": 1.6600346744212056, + "learning_rate": 1.9879357282086596e-07, + "loss": 0.7183, + "step": 1685 + }, + { + "epoch": 3.1222222222222222, + "grad_norm": 1.4661779934176216, + "learning_rate": 1.987912496359936e-07, + "loss": 0.6695, + "step": 1686 + }, + { + "epoch": 3.1240740740740742, + "grad_norm": 1.4292786918613043, + "learning_rate": 1.9878892423002562e-07, + "loss": 0.7261, + "step": 1687 + }, + { + "epoch": 3.1259259259259258, + "grad_norm": 1.5135291921133704, + "learning_rate": 1.987865966030143e-07, + "loss": 0.7349, + "step": 1688 + }, + { + "epoch": 3.1277777777777778, + "grad_norm": 1.465639785609288, + "learning_rate": 1.98784266755012e-07, + "loss": 0.6916, + "step": 1689 + }, + { + "epoch": 3.1296296296296298, + "grad_norm": 1.292903294793804, + "learning_rate": 1.987819346860711e-07, + "loss": 0.6983, + "step": 1690 + }, + { + "epoch": 3.1314814814814813, + "grad_norm": 1.5769108742797056, + "learning_rate": 1.98779600396244e-07, + "loss": 0.7236, + "step": 1691 + }, + { + "epoch": 3.1333333333333333, + "grad_norm": 1.439158811004055, + "learning_rate": 1.9877726388558322e-07, + "loss": 0.708, + "step": 1692 + }, + { + "epoch": 3.1351851851851853, + "grad_norm": 2.7838377971471076, + "learning_rate": 1.9877492515414125e-07, + "loss": 0.6605, + "step": 1693 + }, + { + "epoch": 3.137037037037037, + "grad_norm": 1.770285195097917, + "learning_rate": 1.987725842019707e-07, + "loss": 0.7156, + "step": 1694 + }, + { + "epoch": 3.138888888888889, + "grad_norm": 1.7884590387188937, + "learning_rate": 1.987702410291242e-07, + "loss": 0.7023, + "step": 1695 + }, + { + "epoch": 3.140740740740741, + "grad_norm": 1.3955629912551473, + "learning_rate": 1.9876789563565443e-07, + "loss": 0.7182, + "step": 1696 + }, + { + "epoch": 3.1425925925925924, + "grad_norm": 1.4431664841162344, + "learning_rate": 1.987655480216141e-07, + "loss": 0.6731, + "step": 1697 + }, + { + "epoch": 3.1444444444444444, + "grad_norm": 1.3261493602966676, + "learning_rate": 1.9876319818705602e-07, + "loss": 0.646, + "step": 1698 + }, + { + "epoch": 3.1462962962962964, + "grad_norm": 1.3959952532044104, + "learning_rate": 1.9876084613203302e-07, + "loss": 0.6796, + "step": 1699 + }, + { + "epoch": 3.148148148148148, + "grad_norm": 2.031273973137189, + "learning_rate": 1.9875849185659795e-07, + "loss": 0.6788, + "step": 1700 + }, + { + "epoch": 3.15, + "grad_norm": 1.5551745296788908, + "learning_rate": 1.9875613536080377e-07, + "loss": 0.7309, + "step": 1701 + }, + { + "epoch": 3.151851851851852, + "grad_norm": 1.6289415891619674, + "learning_rate": 1.9875377664470345e-07, + "loss": 0.7677, + "step": 1702 + }, + { + "epoch": 3.153703703703704, + "grad_norm": 1.4390675140572768, + "learning_rate": 1.9875141570835004e-07, + "loss": 0.7382, + "step": 1703 + }, + { + "epoch": 3.1555555555555554, + "grad_norm": 1.372965455625381, + "learning_rate": 1.9874905255179657e-07, + "loss": 0.7105, + "step": 1704 + }, + { + "epoch": 3.1574074074074074, + "grad_norm": 1.6015794601427793, + "learning_rate": 1.987466871750962e-07, + "loss": 0.6916, + "step": 1705 + }, + { + "epoch": 3.1592592592592594, + "grad_norm": 1.5890351606935418, + "learning_rate": 1.987443195783021e-07, + "loss": 0.7614, + "step": 1706 + }, + { + "epoch": 3.161111111111111, + "grad_norm": 1.5303485579252196, + "learning_rate": 1.9874194976146755e-07, + "loss": 0.6869, + "step": 1707 + }, + { + "epoch": 3.162962962962963, + "grad_norm": 1.5551718907204275, + "learning_rate": 1.9873957772464576e-07, + "loss": 0.6605, + "step": 1708 + }, + { + "epoch": 3.164814814814815, + "grad_norm": 1.4360847819466684, + "learning_rate": 1.9873720346789008e-07, + "loss": 0.738, + "step": 1709 + }, + { + "epoch": 3.1666666666666665, + "grad_norm": 1.4184993620680044, + "learning_rate": 1.9873482699125393e-07, + "loss": 0.6959, + "step": 1710 + }, + { + "epoch": 3.1685185185185185, + "grad_norm": 1.9224815358750618, + "learning_rate": 1.987324482947907e-07, + "loss": 0.74, + "step": 1711 + }, + { + "epoch": 3.1703703703703705, + "grad_norm": 1.5359644124916254, + "learning_rate": 1.9873006737855386e-07, + "loss": 0.6937, + "step": 1712 + }, + { + "epoch": 3.172222222222222, + "grad_norm": 1.6687035934098722, + "learning_rate": 1.9872768424259698e-07, + "loss": 0.7322, + "step": 1713 + }, + { + "epoch": 3.174074074074074, + "grad_norm": 1.7354524302271312, + "learning_rate": 1.987252988869736e-07, + "loss": 0.6876, + "step": 1714 + }, + { + "epoch": 3.175925925925926, + "grad_norm": 1.8078074991161097, + "learning_rate": 1.987229113117374e-07, + "loss": 0.7763, + "step": 1715 + }, + { + "epoch": 3.1777777777777776, + "grad_norm": 1.5431668606476954, + "learning_rate": 1.9872052151694202e-07, + "loss": 0.7032, + "step": 1716 + }, + { + "epoch": 3.1796296296296296, + "grad_norm": 2.0257639254774085, + "learning_rate": 1.9871812950264117e-07, + "loss": 0.69, + "step": 1717 + }, + { + "epoch": 3.1814814814814816, + "grad_norm": 1.7816118494622915, + "learning_rate": 1.9871573526888869e-07, + "loss": 0.6912, + "step": 1718 + }, + { + "epoch": 3.183333333333333, + "grad_norm": 1.367941302477703, + "learning_rate": 1.987133388157383e-07, + "loss": 0.6725, + "step": 1719 + }, + { + "epoch": 3.185185185185185, + "grad_norm": 1.9654016251306183, + "learning_rate": 1.9871094014324402e-07, + "loss": 0.6604, + "step": 1720 + }, + { + "epoch": 3.187037037037037, + "grad_norm": 1.548416789303848, + "learning_rate": 1.9870853925145968e-07, + "loss": 0.7079, + "step": 1721 + }, + { + "epoch": 3.188888888888889, + "grad_norm": 1.3323893288004114, + "learning_rate": 1.9870613614043933e-07, + "loss": 0.6793, + "step": 1722 + }, + { + "epoch": 3.1907407407407407, + "grad_norm": 1.3496211227864021, + "learning_rate": 1.9870373081023691e-07, + "loss": 0.7095, + "step": 1723 + }, + { + "epoch": 3.1925925925925926, + "grad_norm": 1.5871603174102669, + "learning_rate": 1.9870132326090656e-07, + "loss": 0.7386, + "step": 1724 + }, + { + "epoch": 3.1944444444444446, + "grad_norm": 1.4320257269335934, + "learning_rate": 1.986989134925024e-07, + "loss": 0.7216, + "step": 1725 + }, + { + "epoch": 3.196296296296296, + "grad_norm": 1.5076149315040464, + "learning_rate": 1.9869650150507859e-07, + "loss": 0.7277, + "step": 1726 + }, + { + "epoch": 3.198148148148148, + "grad_norm": 1.4735141461354706, + "learning_rate": 1.986940872986894e-07, + "loss": 0.7014, + "step": 1727 + }, + { + "epoch": 3.2, + "grad_norm": 1.4614415608173328, + "learning_rate": 1.9869167087338905e-07, + "loss": 0.6927, + "step": 1728 + }, + { + "epoch": 3.2018518518518517, + "grad_norm": 1.3521199024968569, + "learning_rate": 1.9868925222923192e-07, + "loss": 0.6945, + "step": 1729 + }, + { + "epoch": 3.2037037037037037, + "grad_norm": 1.3345681083578147, + "learning_rate": 1.9868683136627235e-07, + "loss": 0.7484, + "step": 1730 + }, + { + "epoch": 3.2055555555555557, + "grad_norm": 1.4216257424639744, + "learning_rate": 1.9868440828456477e-07, + "loss": 0.718, + "step": 1731 + }, + { + "epoch": 3.2074074074074073, + "grad_norm": 1.5742081718620373, + "learning_rate": 1.9868198298416368e-07, + "loss": 0.7128, + "step": 1732 + }, + { + "epoch": 3.2092592592592593, + "grad_norm": 1.660929669372897, + "learning_rate": 1.986795554651236e-07, + "loss": 0.7197, + "step": 1733 + }, + { + "epoch": 3.2111111111111112, + "grad_norm": 1.418589702033337, + "learning_rate": 1.986771257274991e-07, + "loss": 0.6862, + "step": 1734 + }, + { + "epoch": 3.212962962962963, + "grad_norm": 1.4879585482571616, + "learning_rate": 1.9867469377134483e-07, + "loss": 0.7015, + "step": 1735 + }, + { + "epoch": 3.214814814814815, + "grad_norm": 1.4155096043374973, + "learning_rate": 1.986722595967154e-07, + "loss": 0.7119, + "step": 1736 + }, + { + "epoch": 3.216666666666667, + "grad_norm": 1.6719920845608907, + "learning_rate": 1.9866982320366564e-07, + "loss": 0.6801, + "step": 1737 + }, + { + "epoch": 3.2185185185185183, + "grad_norm": 1.917493852131688, + "learning_rate": 1.9866738459225023e-07, + "loss": 0.693, + "step": 1738 + }, + { + "epoch": 3.2203703703703703, + "grad_norm": 1.4048596837328573, + "learning_rate": 1.9866494376252406e-07, + "loss": 0.6552, + "step": 1739 + }, + { + "epoch": 3.2222222222222223, + "grad_norm": 1.3107955464321916, + "learning_rate": 1.9866250071454198e-07, + "loss": 0.7368, + "step": 1740 + }, + { + "epoch": 3.224074074074074, + "grad_norm": 1.6415296097464296, + "learning_rate": 1.9866005544835894e-07, + "loss": 0.6746, + "step": 1741 + }, + { + "epoch": 3.225925925925926, + "grad_norm": 1.3668511561598151, + "learning_rate": 1.9865760796402986e-07, + "loss": 0.7314, + "step": 1742 + }, + { + "epoch": 3.227777777777778, + "grad_norm": 1.278809798682975, + "learning_rate": 1.9865515826160982e-07, + "loss": 0.6867, + "step": 1743 + }, + { + "epoch": 3.2296296296296294, + "grad_norm": 1.3961839263779474, + "learning_rate": 1.9865270634115387e-07, + "loss": 0.7131, + "step": 1744 + }, + { + "epoch": 3.2314814814814814, + "grad_norm": 1.4091498954729598, + "learning_rate": 1.9865025220271715e-07, + "loss": 0.6982, + "step": 1745 + }, + { + "epoch": 3.2333333333333334, + "grad_norm": 1.5630621979796469, + "learning_rate": 1.9864779584635484e-07, + "loss": 0.7226, + "step": 1746 + }, + { + "epoch": 3.2351851851851854, + "grad_norm": 1.7604298806648988, + "learning_rate": 1.9864533727212214e-07, + "loss": 0.7291, + "step": 1747 + }, + { + "epoch": 3.237037037037037, + "grad_norm": 1.3320347511429715, + "learning_rate": 1.9864287648007436e-07, + "loss": 0.714, + "step": 1748 + }, + { + "epoch": 3.238888888888889, + "grad_norm": 1.6794335700204426, + "learning_rate": 1.986404134702668e-07, + "loss": 0.6914, + "step": 1749 + }, + { + "epoch": 3.240740740740741, + "grad_norm": 3.0510458140298247, + "learning_rate": 1.9863794824275483e-07, + "loss": 0.7281, + "step": 1750 + }, + { + "epoch": 3.2425925925925925, + "grad_norm": 1.3230940795455612, + "learning_rate": 1.9863548079759388e-07, + "loss": 0.7322, + "step": 1751 + }, + { + "epoch": 3.2444444444444445, + "grad_norm": 1.4313794759766358, + "learning_rate": 1.9863301113483946e-07, + "loss": 0.7064, + "step": 1752 + }, + { + "epoch": 3.2462962962962965, + "grad_norm": 1.4628379933589932, + "learning_rate": 1.9863053925454705e-07, + "loss": 0.6756, + "step": 1753 + }, + { + "epoch": 3.248148148148148, + "grad_norm": 1.406276310367317, + "learning_rate": 1.9862806515677227e-07, + "loss": 0.7186, + "step": 1754 + }, + { + "epoch": 3.25, + "grad_norm": 1.4343961238620486, + "learning_rate": 1.9862558884157065e-07, + "loss": 0.6865, + "step": 1755 + }, + { + "epoch": 3.251851851851852, + "grad_norm": 2.366793024973306, + "learning_rate": 1.9862311030899798e-07, + "loss": 0.6603, + "step": 1756 + }, + { + "epoch": 3.2537037037037035, + "grad_norm": 1.6388609763325115, + "learning_rate": 1.9862062955910992e-07, + "loss": 0.6923, + "step": 1757 + }, + { + "epoch": 3.2555555555555555, + "grad_norm": 2.0986713998445925, + "learning_rate": 1.9861814659196228e-07, + "loss": 0.7106, + "step": 1758 + }, + { + "epoch": 3.2574074074074075, + "grad_norm": 1.2530713132208617, + "learning_rate": 1.9861566140761083e-07, + "loss": 0.666, + "step": 1759 + }, + { + "epoch": 3.259259259259259, + "grad_norm": 1.7827017227512207, + "learning_rate": 1.9861317400611148e-07, + "loss": 0.7117, + "step": 1760 + }, + { + "epoch": 3.261111111111111, + "grad_norm": 1.6668804613145347, + "learning_rate": 1.986106843875202e-07, + "loss": 0.6844, + "step": 1761 + }, + { + "epoch": 3.262962962962963, + "grad_norm": 2.433122523000563, + "learning_rate": 1.9860819255189285e-07, + "loss": 0.6974, + "step": 1762 + }, + { + "epoch": 3.264814814814815, + "grad_norm": 1.5368541544360192, + "learning_rate": 1.9860569849928552e-07, + "loss": 0.7131, + "step": 1763 + }, + { + "epoch": 3.2666666666666666, + "grad_norm": 1.5907604703887104, + "learning_rate": 1.986032022297543e-07, + "loss": 0.6813, + "step": 1764 + }, + { + "epoch": 3.2685185185185186, + "grad_norm": 2.054360580524189, + "learning_rate": 1.986007037433553e-07, + "loss": 0.718, + "step": 1765 + }, + { + "epoch": 3.2703703703703706, + "grad_norm": 1.9050273075340438, + "learning_rate": 1.9859820304014466e-07, + "loss": 0.7575, + "step": 1766 + }, + { + "epoch": 3.272222222222222, + "grad_norm": 1.6060824521122723, + "learning_rate": 1.9859570012017866e-07, + "loss": 0.6988, + "step": 1767 + }, + { + "epoch": 3.274074074074074, + "grad_norm": 1.4382538556506632, + "learning_rate": 1.985931949835135e-07, + "loss": 0.7233, + "step": 1768 + }, + { + "epoch": 3.275925925925926, + "grad_norm": 1.7742874379466718, + "learning_rate": 1.9859068763020557e-07, + "loss": 0.6661, + "step": 1769 + }, + { + "epoch": 3.2777777777777777, + "grad_norm": 1.6377630137693746, + "learning_rate": 1.985881780603112e-07, + "loss": 0.7553, + "step": 1770 + }, + { + "epoch": 3.2796296296296297, + "grad_norm": 1.5376859283539732, + "learning_rate": 1.9858566627388681e-07, + "loss": 0.6796, + "step": 1771 + }, + { + "epoch": 3.2814814814814817, + "grad_norm": 1.5021536429969475, + "learning_rate": 1.9858315227098892e-07, + "loss": 0.7195, + "step": 1772 + }, + { + "epoch": 3.283333333333333, + "grad_norm": 1.4369446812207705, + "learning_rate": 1.98580636051674e-07, + "loss": 0.7122, + "step": 1773 + }, + { + "epoch": 3.285185185185185, + "grad_norm": 1.6431245894661268, + "learning_rate": 1.9857811761599864e-07, + "loss": 0.7535, + "step": 1774 + }, + { + "epoch": 3.287037037037037, + "grad_norm": 1.4250330847004862, + "learning_rate": 1.9857559696401948e-07, + "loss": 0.6714, + "step": 1775 + }, + { + "epoch": 3.2888888888888888, + "grad_norm": 1.722137182459568, + "learning_rate": 1.9857307409579317e-07, + "loss": 0.7529, + "step": 1776 + }, + { + "epoch": 3.2907407407407407, + "grad_norm": 1.47834174654245, + "learning_rate": 1.985705490113764e-07, + "loss": 0.6833, + "step": 1777 + }, + { + "epoch": 3.2925925925925927, + "grad_norm": 1.6717925145228725, + "learning_rate": 1.98568021710826e-07, + "loss": 0.6778, + "step": 1778 + }, + { + "epoch": 3.2944444444444443, + "grad_norm": 1.3389523840169226, + "learning_rate": 1.9856549219419878e-07, + "loss": 0.7261, + "step": 1779 + }, + { + "epoch": 3.2962962962962963, + "grad_norm": 1.508202190548861, + "learning_rate": 1.9856296046155158e-07, + "loss": 0.6902, + "step": 1780 + }, + { + "epoch": 3.2981481481481483, + "grad_norm": 1.3055339718961279, + "learning_rate": 1.9856042651294132e-07, + "loss": 0.7292, + "step": 1781 + }, + { + "epoch": 3.3, + "grad_norm": 1.3784260823512247, + "learning_rate": 1.9855789034842504e-07, + "loss": 0.69, + "step": 1782 + }, + { + "epoch": 3.301851851851852, + "grad_norm": 1.4540692938786228, + "learning_rate": 1.9855535196805966e-07, + "loss": 0.662, + "step": 1783 + }, + { + "epoch": 3.303703703703704, + "grad_norm": 1.4179602530351605, + "learning_rate": 1.985528113719023e-07, + "loss": 0.7272, + "step": 1784 + }, + { + "epoch": 3.3055555555555554, + "grad_norm": 1.431388432484835, + "learning_rate": 1.9855026856001007e-07, + "loss": 0.6861, + "step": 1785 + }, + { + "epoch": 3.3074074074074074, + "grad_norm": 1.474845184737349, + "learning_rate": 1.9854772353244016e-07, + "loss": 0.7601, + "step": 1786 + }, + { + "epoch": 3.3092592592592593, + "grad_norm": 1.4115749619499567, + "learning_rate": 1.9854517628924977e-07, + "loss": 0.6965, + "step": 1787 + }, + { + "epoch": 3.311111111111111, + "grad_norm": 2.031703428532122, + "learning_rate": 1.9854262683049616e-07, + "loss": 0.7051, + "step": 1788 + }, + { + "epoch": 3.312962962962963, + "grad_norm": 1.4210085566074337, + "learning_rate": 1.9854007515623667e-07, + "loss": 0.6889, + "step": 1789 + }, + { + "epoch": 3.314814814814815, + "grad_norm": 1.4557990808669314, + "learning_rate": 1.9853752126652864e-07, + "loss": 0.7347, + "step": 1790 + }, + { + "epoch": 3.3166666666666664, + "grad_norm": 1.367150528394596, + "learning_rate": 1.9853496516142955e-07, + "loss": 0.7006, + "step": 1791 + }, + { + "epoch": 3.3185185185185184, + "grad_norm": 2.470382011775795, + "learning_rate": 1.9853240684099682e-07, + "loss": 0.7246, + "step": 1792 + }, + { + "epoch": 3.3203703703703704, + "grad_norm": 1.4075732692730825, + "learning_rate": 1.9852984630528792e-07, + "loss": 0.6332, + "step": 1793 + }, + { + "epoch": 3.3222222222222224, + "grad_norm": 1.377107831679937, + "learning_rate": 1.9852728355436052e-07, + "loss": 0.7314, + "step": 1794 + }, + { + "epoch": 3.324074074074074, + "grad_norm": 1.5179482731981948, + "learning_rate": 1.9852471858827219e-07, + "loss": 0.6982, + "step": 1795 + }, + { + "epoch": 3.325925925925926, + "grad_norm": 2.01220396343066, + "learning_rate": 1.985221514070806e-07, + "loss": 0.7429, + "step": 1796 + }, + { + "epoch": 3.327777777777778, + "grad_norm": 1.3123577076162922, + "learning_rate": 1.9851958201084345e-07, + "loss": 0.6921, + "step": 1797 + }, + { + "epoch": 3.3296296296296295, + "grad_norm": 1.4965491778271642, + "learning_rate": 1.985170103996185e-07, + "loss": 0.7382, + "step": 1798 + }, + { + "epoch": 3.3314814814814815, + "grad_norm": 1.4850693258187, + "learning_rate": 1.9851443657346364e-07, + "loss": 0.7577, + "step": 1799 + }, + { + "epoch": 3.3333333333333335, + "grad_norm": 1.7904392962818572, + "learning_rate": 1.9851186053243664e-07, + "loss": 0.6895, + "step": 1800 + }, + { + "epoch": 3.335185185185185, + "grad_norm": 2.8057665234037055, + "learning_rate": 1.9850928227659547e-07, + "loss": 0.6865, + "step": 1801 + }, + { + "epoch": 3.337037037037037, + "grad_norm": 1.371401778740363, + "learning_rate": 1.9850670180599808e-07, + "loss": 0.7548, + "step": 1802 + }, + { + "epoch": 3.338888888888889, + "grad_norm": 2.215297675271137, + "learning_rate": 1.985041191207025e-07, + "loss": 0.6867, + "step": 1803 + }, + { + "epoch": 3.3407407407407406, + "grad_norm": 1.6420758470306913, + "learning_rate": 1.9850153422076682e-07, + "loss": 0.6951, + "step": 1804 + }, + { + "epoch": 3.3425925925925926, + "grad_norm": 1.4527864639750439, + "learning_rate": 1.9849894710624912e-07, + "loss": 0.7336, + "step": 1805 + }, + { + "epoch": 3.3444444444444446, + "grad_norm": 1.3217069089705615, + "learning_rate": 1.9849635777720755e-07, + "loss": 0.6662, + "step": 1806 + }, + { + "epoch": 3.346296296296296, + "grad_norm": 2.067910615618579, + "learning_rate": 1.9849376623370034e-07, + "loss": 0.6142, + "step": 1807 + }, + { + "epoch": 3.348148148148148, + "grad_norm": 1.6148500108855797, + "learning_rate": 1.9849117247578575e-07, + "loss": 0.7426, + "step": 1808 + }, + { + "epoch": 3.35, + "grad_norm": 1.5223554238171622, + "learning_rate": 1.9848857650352212e-07, + "loss": 0.7149, + "step": 1809 + }, + { + "epoch": 3.351851851851852, + "grad_norm": 1.5893635024278807, + "learning_rate": 1.984859783169678e-07, + "loss": 0.6712, + "step": 1810 + }, + { + "epoch": 3.3537037037037036, + "grad_norm": 2.0638628133874355, + "learning_rate": 1.984833779161812e-07, + "loss": 0.713, + "step": 1811 + }, + { + "epoch": 3.3555555555555556, + "grad_norm": 1.3625841468778455, + "learning_rate": 1.984807753012208e-07, + "loss": 0.7031, + "step": 1812 + }, + { + "epoch": 3.3574074074074076, + "grad_norm": 1.4926336279072059, + "learning_rate": 1.9847817047214508e-07, + "loss": 0.7021, + "step": 1813 + }, + { + "epoch": 3.359259259259259, + "grad_norm": 1.4858279133201107, + "learning_rate": 1.9847556342901265e-07, + "loss": 0.6981, + "step": 1814 + }, + { + "epoch": 3.361111111111111, + "grad_norm": 1.3795031843352072, + "learning_rate": 1.9847295417188207e-07, + "loss": 0.741, + "step": 1815 + }, + { + "epoch": 3.362962962962963, + "grad_norm": 2.2109787933302876, + "learning_rate": 1.9847034270081206e-07, + "loss": 0.6689, + "step": 1816 + }, + { + "epoch": 3.3648148148148147, + "grad_norm": 1.8045167707521255, + "learning_rate": 1.984677290158613e-07, + "loss": 0.7523, + "step": 1817 + }, + { + "epoch": 3.3666666666666667, + "grad_norm": 1.4517747645748975, + "learning_rate": 1.9846511311708855e-07, + "loss": 0.7491, + "step": 1818 + }, + { + "epoch": 3.3685185185185187, + "grad_norm": 1.2931889574317914, + "learning_rate": 1.9846249500455263e-07, + "loss": 0.6776, + "step": 1819 + }, + { + "epoch": 3.3703703703703702, + "grad_norm": 1.4044752792037491, + "learning_rate": 1.984598746783124e-07, + "loss": 0.6783, + "step": 1820 + }, + { + "epoch": 3.3722222222222222, + "grad_norm": 1.6478546800324836, + "learning_rate": 1.9845725213842676e-07, + "loss": 0.7151, + "step": 1821 + }, + { + "epoch": 3.3740740740740742, + "grad_norm": 1.8568294350567034, + "learning_rate": 1.9845462738495473e-07, + "loss": 0.7111, + "step": 1822 + }, + { + "epoch": 3.3759259259259258, + "grad_norm": 1.4979035309359159, + "learning_rate": 1.9845200041795523e-07, + "loss": 0.6891, + "step": 1823 + }, + { + "epoch": 3.3777777777777778, + "grad_norm": 1.4950173832590299, + "learning_rate": 1.984493712374874e-07, + "loss": 0.6792, + "step": 1824 + }, + { + "epoch": 3.3796296296296298, + "grad_norm": 1.4108846458026787, + "learning_rate": 1.984467398436103e-07, + "loss": 0.778, + "step": 1825 + }, + { + "epoch": 3.3814814814814813, + "grad_norm": 1.5839148313364935, + "learning_rate": 1.984441062363831e-07, + "loss": 0.703, + "step": 1826 + }, + { + "epoch": 3.3833333333333333, + "grad_norm": 1.4385794392575448, + "learning_rate": 1.9844147041586507e-07, + "loss": 0.7153, + "step": 1827 + }, + { + "epoch": 3.3851851851851853, + "grad_norm": 1.5272095334171423, + "learning_rate": 1.9843883238211538e-07, + "loss": 0.7068, + "step": 1828 + }, + { + "epoch": 3.387037037037037, + "grad_norm": 2.6314228093941163, + "learning_rate": 1.984361921351934e-07, + "loss": 0.7182, + "step": 1829 + }, + { + "epoch": 3.388888888888889, + "grad_norm": 1.4089460445025883, + "learning_rate": 1.984335496751585e-07, + "loss": 0.6674, + "step": 1830 + }, + { + "epoch": 3.390740740740741, + "grad_norm": 3.0085161164177805, + "learning_rate": 1.9843090500207e-07, + "loss": 0.7527, + "step": 1831 + }, + { + "epoch": 3.3925925925925924, + "grad_norm": 1.4130940415035835, + "learning_rate": 1.9842825811598747e-07, + "loss": 0.7204, + "step": 1832 + }, + { + "epoch": 3.3944444444444444, + "grad_norm": 1.5105516145504305, + "learning_rate": 1.9842560901697035e-07, + "loss": 0.6699, + "step": 1833 + }, + { + "epoch": 3.3962962962962964, + "grad_norm": 1.491999933569889, + "learning_rate": 1.9842295770507822e-07, + "loss": 0.6861, + "step": 1834 + }, + { + "epoch": 3.398148148148148, + "grad_norm": 1.253748570846862, + "learning_rate": 1.984203041803707e-07, + "loss": 0.6806, + "step": 1835 + }, + { + "epoch": 3.4, + "grad_norm": 1.5545753659869026, + "learning_rate": 1.9841764844290743e-07, + "loss": 0.6927, + "step": 1836 + }, + { + "epoch": 3.401851851851852, + "grad_norm": 1.501761073419926, + "learning_rate": 1.984149904927481e-07, + "loss": 0.7158, + "step": 1837 + }, + { + "epoch": 3.403703703703704, + "grad_norm": 1.7265965770255634, + "learning_rate": 1.9841233032995252e-07, + "loss": 0.6741, + "step": 1838 + }, + { + "epoch": 3.4055555555555554, + "grad_norm": 1.6305566817488533, + "learning_rate": 1.9840966795458045e-07, + "loss": 0.6849, + "step": 1839 + }, + { + "epoch": 3.4074074074074074, + "grad_norm": 1.4081321041940684, + "learning_rate": 1.984070033666918e-07, + "loss": 0.7227, + "step": 1840 + }, + { + "epoch": 3.4092592592592594, + "grad_norm": 1.5194649070203468, + "learning_rate": 1.984043365663464e-07, + "loss": 0.6884, + "step": 1841 + }, + { + "epoch": 3.411111111111111, + "grad_norm": 1.7708797081830638, + "learning_rate": 1.984016675536043e-07, + "loss": 0.677, + "step": 1842 + }, + { + "epoch": 3.412962962962963, + "grad_norm": 1.387747704612861, + "learning_rate": 1.9839899632852545e-07, + "loss": 0.7512, + "step": 1843 + }, + { + "epoch": 3.414814814814815, + "grad_norm": 2.0993706208511282, + "learning_rate": 1.983963228911699e-07, + "loss": 0.6885, + "step": 1844 + }, + { + "epoch": 3.4166666666666665, + "grad_norm": 1.3781961699995935, + "learning_rate": 1.9839364724159776e-07, + "loss": 0.7207, + "step": 1845 + }, + { + "epoch": 3.4185185185185185, + "grad_norm": 1.5712083425661263, + "learning_rate": 1.9839096937986922e-07, + "loss": 0.6917, + "step": 1846 + }, + { + "epoch": 3.4203703703703705, + "grad_norm": 1.7443941577490623, + "learning_rate": 1.9838828930604444e-07, + "loss": 0.7186, + "step": 1847 + }, + { + "epoch": 3.422222222222222, + "grad_norm": 1.4627056735055046, + "learning_rate": 1.9838560702018373e-07, + "loss": 0.7947, + "step": 1848 + }, + { + "epoch": 3.424074074074074, + "grad_norm": 1.4323519976183854, + "learning_rate": 1.9838292252234734e-07, + "loss": 0.7513, + "step": 1849 + }, + { + "epoch": 3.425925925925926, + "grad_norm": 1.5282606769898337, + "learning_rate": 1.9838023581259564e-07, + "loss": 0.6961, + "step": 1850 + }, + { + "epoch": 3.4277777777777776, + "grad_norm": 1.5200764897120511, + "learning_rate": 1.9837754689098903e-07, + "loss": 0.6774, + "step": 1851 + }, + { + "epoch": 3.4296296296296296, + "grad_norm": 1.5836343448700176, + "learning_rate": 1.9837485575758805e-07, + "loss": 0.7075, + "step": 1852 + }, + { + "epoch": 3.4314814814814816, + "grad_norm": 1.533404345674884, + "learning_rate": 1.9837216241245306e-07, + "loss": 0.6937, + "step": 1853 + }, + { + "epoch": 3.4333333333333336, + "grad_norm": 1.7884097474333618, + "learning_rate": 1.983694668556447e-07, + "loss": 0.7169, + "step": 1854 + }, + { + "epoch": 3.435185185185185, + "grad_norm": 1.689564754648096, + "learning_rate": 1.983667690872236e-07, + "loss": 0.7194, + "step": 1855 + }, + { + "epoch": 3.437037037037037, + "grad_norm": 1.584534178578136, + "learning_rate": 1.983640691072503e-07, + "loss": 0.7312, + "step": 1856 + }, + { + "epoch": 3.438888888888889, + "grad_norm": 2.6423491864346618, + "learning_rate": 1.9836136691578562e-07, + "loss": 0.6967, + "step": 1857 + }, + { + "epoch": 3.4407407407407407, + "grad_norm": 1.5964527218129614, + "learning_rate": 1.9835866251289023e-07, + "loss": 0.6772, + "step": 1858 + }, + { + "epoch": 3.4425925925925926, + "grad_norm": 1.4982229630711585, + "learning_rate": 1.98355955898625e-07, + "loss": 0.6997, + "step": 1859 + }, + { + "epoch": 3.4444444444444446, + "grad_norm": 1.4673660259822097, + "learning_rate": 1.9835324707305074e-07, + "loss": 0.6836, + "step": 1860 + }, + { + "epoch": 3.446296296296296, + "grad_norm": 1.842385103664792, + "learning_rate": 1.9835053603622838e-07, + "loss": 0.7167, + "step": 1861 + }, + { + "epoch": 3.448148148148148, + "grad_norm": 1.5177275684482423, + "learning_rate": 1.9834782278821883e-07, + "loss": 0.6853, + "step": 1862 + }, + { + "epoch": 3.45, + "grad_norm": 2.167833033772285, + "learning_rate": 1.983451073290831e-07, + "loss": 0.7273, + "step": 1863 + }, + { + "epoch": 3.4518518518518517, + "grad_norm": 1.560681261677744, + "learning_rate": 1.983423896588823e-07, + "loss": 0.6453, + "step": 1864 + }, + { + "epoch": 3.4537037037037037, + "grad_norm": 1.5380173541655122, + "learning_rate": 1.983396697776775e-07, + "loss": 0.7512, + "step": 1865 + }, + { + "epoch": 3.4555555555555557, + "grad_norm": 1.4693529725852381, + "learning_rate": 1.983369476855298e-07, + "loss": 0.7107, + "step": 1866 + }, + { + "epoch": 3.4574074074074073, + "grad_norm": 1.4125957262066153, + "learning_rate": 1.983342233825004e-07, + "loss": 0.7222, + "step": 1867 + }, + { + "epoch": 3.4592592592592593, + "grad_norm": 1.5180714260911108, + "learning_rate": 1.9833149686865066e-07, + "loss": 0.7133, + "step": 1868 + }, + { + "epoch": 3.4611111111111112, + "grad_norm": 1.3970107683126725, + "learning_rate": 1.9832876814404175e-07, + "loss": 0.6759, + "step": 1869 + }, + { + "epoch": 3.462962962962963, + "grad_norm": 1.3742674833403816, + "learning_rate": 1.983260372087351e-07, + "loss": 0.689, + "step": 1870 + }, + { + "epoch": 3.464814814814815, + "grad_norm": 1.533264797025038, + "learning_rate": 1.983233040627921e-07, + "loss": 0.6807, + "step": 1871 + }, + { + "epoch": 3.466666666666667, + "grad_norm": 1.5775174685083997, + "learning_rate": 1.9832056870627416e-07, + "loss": 0.6532, + "step": 1872 + }, + { + "epoch": 3.4685185185185183, + "grad_norm": 1.57956971869636, + "learning_rate": 1.983178311392428e-07, + "loss": 0.7, + "step": 1873 + }, + { + "epoch": 3.4703703703703703, + "grad_norm": 1.60845502213651, + "learning_rate": 1.983150913617596e-07, + "loss": 0.711, + "step": 1874 + }, + { + "epoch": 3.4722222222222223, + "grad_norm": 1.5212273014649118, + "learning_rate": 1.983123493738861e-07, + "loss": 0.7242, + "step": 1875 + }, + { + "epoch": 3.474074074074074, + "grad_norm": 1.8248193052095798, + "learning_rate": 1.9830960517568395e-07, + "loss": 0.7044, + "step": 1876 + }, + { + "epoch": 3.475925925925926, + "grad_norm": 1.4933086652691223, + "learning_rate": 1.9830685876721488e-07, + "loss": 0.668, + "step": 1877 + }, + { + "epoch": 3.477777777777778, + "grad_norm": 1.409484200081249, + "learning_rate": 1.9830411014854068e-07, + "loss": 0.7498, + "step": 1878 + }, + { + "epoch": 3.4796296296296294, + "grad_norm": 1.450000499176415, + "learning_rate": 1.9830135931972304e-07, + "loss": 0.7118, + "step": 1879 + }, + { + "epoch": 3.4814814814814814, + "grad_norm": 1.3893754012226047, + "learning_rate": 1.982986062808239e-07, + "loss": 0.7373, + "step": 1880 + }, + { + "epoch": 3.4833333333333334, + "grad_norm": 1.3333801218890933, + "learning_rate": 1.982958510319051e-07, + "loss": 0.6864, + "step": 1881 + }, + { + "epoch": 3.485185185185185, + "grad_norm": 1.6738239699454398, + "learning_rate": 1.9829309357302857e-07, + "loss": 0.681, + "step": 1882 + }, + { + "epoch": 3.487037037037037, + "grad_norm": 1.8996939479255008, + "learning_rate": 1.9829033390425635e-07, + "loss": 0.6832, + "step": 1883 + }, + { + "epoch": 3.488888888888889, + "grad_norm": 1.4297045913556865, + "learning_rate": 1.982875720256505e-07, + "loss": 0.6724, + "step": 1884 + }, + { + "epoch": 3.490740740740741, + "grad_norm": 1.3175346513259658, + "learning_rate": 1.9828480793727308e-07, + "loss": 0.7191, + "step": 1885 + }, + { + "epoch": 3.4925925925925925, + "grad_norm": 1.2809625303647758, + "learning_rate": 1.982820416391862e-07, + "loss": 0.6955, + "step": 1886 + }, + { + "epoch": 3.4944444444444445, + "grad_norm": 1.7030736431840674, + "learning_rate": 1.9827927313145214e-07, + "loss": 0.6785, + "step": 1887 + }, + { + "epoch": 3.4962962962962965, + "grad_norm": 1.4782137077450757, + "learning_rate": 1.9827650241413308e-07, + "loss": 0.6843, + "step": 1888 + }, + { + "epoch": 3.498148148148148, + "grad_norm": 11.343940043858082, + "learning_rate": 1.9827372948729132e-07, + "loss": 0.67, + "step": 1889 + }, + { + "epoch": 3.5, + "grad_norm": 2.4214805025973365, + "learning_rate": 1.9827095435098923e-07, + "loss": 0.6929, + "step": 1890 + }, + { + "epoch": 3.501851851851852, + "grad_norm": 1.446837042767488, + "learning_rate": 1.9826817700528918e-07, + "loss": 0.7051, + "step": 1891 + }, + { + "epoch": 3.5037037037037035, + "grad_norm": 1.525076726251059, + "learning_rate": 1.9826539745025364e-07, + "loss": 0.7174, + "step": 1892 + }, + { + "epoch": 3.5055555555555555, + "grad_norm": 1.8971822298167618, + "learning_rate": 1.9826261568594505e-07, + "loss": 0.7369, + "step": 1893 + }, + { + "epoch": 3.5074074074074075, + "grad_norm": 1.402419883821451, + "learning_rate": 1.98259831712426e-07, + "loss": 0.7071, + "step": 1894 + }, + { + "epoch": 3.5092592592592595, + "grad_norm": 1.4348258278852095, + "learning_rate": 1.9825704552975905e-07, + "loss": 0.7685, + "step": 1895 + }, + { + "epoch": 3.511111111111111, + "grad_norm": 1.3758144198481757, + "learning_rate": 1.9825425713800684e-07, + "loss": 0.7234, + "step": 1896 + }, + { + "epoch": 3.512962962962963, + "grad_norm": 1.432153376179934, + "learning_rate": 1.9825146653723212e-07, + "loss": 0.683, + "step": 1897 + }, + { + "epoch": 3.514814814814815, + "grad_norm": 1.592847660149842, + "learning_rate": 1.9824867372749754e-07, + "loss": 0.7368, + "step": 1898 + }, + { + "epoch": 3.5166666666666666, + "grad_norm": 1.4260888262639537, + "learning_rate": 1.9824587870886599e-07, + "loss": 0.7514, + "step": 1899 + }, + { + "epoch": 3.5185185185185186, + "grad_norm": 1.425142455196794, + "learning_rate": 1.9824308148140018e-07, + "loss": 0.6946, + "step": 1900 + }, + { + "epoch": 3.5203703703703706, + "grad_norm": 1.486037666687816, + "learning_rate": 1.9824028204516313e-07, + "loss": 0.687, + "step": 1901 + }, + { + "epoch": 3.522222222222222, + "grad_norm": 1.5990733904107073, + "learning_rate": 1.9823748040021772e-07, + "loss": 0.7399, + "step": 1902 + }, + { + "epoch": 3.524074074074074, + "grad_norm": 1.9209588481351703, + "learning_rate": 1.982346765466269e-07, + "loss": 0.6501, + "step": 1903 + }, + { + "epoch": 3.525925925925926, + "grad_norm": 1.4512384059639374, + "learning_rate": 1.982318704844538e-07, + "loss": 0.7175, + "step": 1904 + }, + { + "epoch": 3.5277777777777777, + "grad_norm": 1.6544460474443168, + "learning_rate": 1.9822906221376142e-07, + "loss": 0.6819, + "step": 1905 + }, + { + "epoch": 3.5296296296296297, + "grad_norm": 2.256026112534218, + "learning_rate": 1.98226251734613e-07, + "loss": 0.7003, + "step": 1906 + }, + { + "epoch": 3.5314814814814817, + "grad_norm": 1.595422499695949, + "learning_rate": 1.9822343904707157e-07, + "loss": 0.705, + "step": 1907 + }, + { + "epoch": 3.533333333333333, + "grad_norm": 1.7061819333454802, + "learning_rate": 1.9822062415120051e-07, + "loss": 0.7048, + "step": 1908 + }, + { + "epoch": 3.535185185185185, + "grad_norm": 1.4195221640502955, + "learning_rate": 1.9821780704706305e-07, + "loss": 0.7488, + "step": 1909 + }, + { + "epoch": 3.537037037037037, + "grad_norm": 1.9156954767277399, + "learning_rate": 1.9821498773472256e-07, + "loss": 0.7194, + "step": 1910 + }, + { + "epoch": 3.5388888888888888, + "grad_norm": 1.8031470161670127, + "learning_rate": 1.9821216621424237e-07, + "loss": 0.7232, + "step": 1911 + }, + { + "epoch": 3.5407407407407407, + "grad_norm": 1.5315746107696702, + "learning_rate": 1.9820934248568593e-07, + "loss": 0.6812, + "step": 1912 + }, + { + "epoch": 3.5425925925925927, + "grad_norm": 1.3025150919659598, + "learning_rate": 1.9820651654911676e-07, + "loss": 0.7151, + "step": 1913 + }, + { + "epoch": 3.5444444444444443, + "grad_norm": 1.6177111951062892, + "learning_rate": 1.9820368840459836e-07, + "loss": 0.7453, + "step": 1914 + }, + { + "epoch": 3.5462962962962963, + "grad_norm": 1.7753496097386425, + "learning_rate": 1.9820085805219433e-07, + "loss": 0.6847, + "step": 1915 + }, + { + "epoch": 3.5481481481481483, + "grad_norm": 1.4141943625649132, + "learning_rate": 1.981980254919683e-07, + "loss": 0.6573, + "step": 1916 + }, + { + "epoch": 3.55, + "grad_norm": 1.3753426273588385, + "learning_rate": 1.9819519072398395e-07, + "loss": 0.7115, + "step": 1917 + }, + { + "epoch": 3.551851851851852, + "grad_norm": 1.6651880352440698, + "learning_rate": 1.9819235374830505e-07, + "loss": 0.7062, + "step": 1918 + }, + { + "epoch": 3.553703703703704, + "grad_norm": 5.42353531434544, + "learning_rate": 1.9818951456499532e-07, + "loss": 0.7279, + "step": 1919 + }, + { + "epoch": 3.5555555555555554, + "grad_norm": 1.442599430680827, + "learning_rate": 1.9818667317411864e-07, + "loss": 0.6872, + "step": 1920 + }, + { + "epoch": 3.5574074074074074, + "grad_norm": 1.596581035572579, + "learning_rate": 1.9818382957573887e-07, + "loss": 0.6706, + "step": 1921 + }, + { + "epoch": 3.5592592592592593, + "grad_norm": 1.5801005818380645, + "learning_rate": 1.9818098376991995e-07, + "loss": 0.7154, + "step": 1922 + }, + { + "epoch": 3.561111111111111, + "grad_norm": 1.348122080031413, + "learning_rate": 1.9817813575672587e-07, + "loss": 0.6542, + "step": 1923 + }, + { + "epoch": 3.562962962962963, + "grad_norm": 1.5730064508346542, + "learning_rate": 1.981752855362206e-07, + "loss": 0.7175, + "step": 1924 + }, + { + "epoch": 3.564814814814815, + "grad_norm": 1.2527586950293443, + "learning_rate": 1.9817243310846833e-07, + "loss": 0.6598, + "step": 1925 + }, + { + "epoch": 3.5666666666666664, + "grad_norm": 1.8362728130364492, + "learning_rate": 1.981695784735331e-07, + "loss": 0.7132, + "step": 1926 + }, + { + "epoch": 3.5685185185185184, + "grad_norm": 1.436897812936575, + "learning_rate": 1.9816672163147913e-07, + "loss": 0.6979, + "step": 1927 + }, + { + "epoch": 3.5703703703703704, + "grad_norm": 1.3818273524498794, + "learning_rate": 1.9816386258237065e-07, + "loss": 0.6888, + "step": 1928 + }, + { + "epoch": 3.572222222222222, + "grad_norm": 1.8411447032643546, + "learning_rate": 1.981610013262719e-07, + "loss": 0.6785, + "step": 1929 + }, + { + "epoch": 3.574074074074074, + "grad_norm": 1.7121233462109091, + "learning_rate": 1.981581378632473e-07, + "loss": 0.7334, + "step": 1930 + }, + { + "epoch": 3.575925925925926, + "grad_norm": 1.561830720118593, + "learning_rate": 1.9815527219336113e-07, + "loss": 0.7199, + "step": 1931 + }, + { + "epoch": 3.5777777777777775, + "grad_norm": 3.4500124348032752, + "learning_rate": 1.9815240431667782e-07, + "loss": 0.7113, + "step": 1932 + }, + { + "epoch": 3.5796296296296295, + "grad_norm": 1.5484682542410215, + "learning_rate": 1.9814953423326193e-07, + "loss": 0.7341, + "step": 1933 + }, + { + "epoch": 3.5814814814814815, + "grad_norm": 1.4238796151975872, + "learning_rate": 1.9814666194317793e-07, + "loss": 0.6976, + "step": 1934 + }, + { + "epoch": 3.5833333333333335, + "grad_norm": 2.283827498050039, + "learning_rate": 1.9814378744649041e-07, + "loss": 0.6982, + "step": 1935 + }, + { + "epoch": 3.585185185185185, + "grad_norm": 1.4087371610510286, + "learning_rate": 1.9814091074326396e-07, + "loss": 0.7009, + "step": 1936 + }, + { + "epoch": 3.587037037037037, + "grad_norm": 1.3477722141793254, + "learning_rate": 1.9813803183356335e-07, + "loss": 0.6668, + "step": 1937 + }, + { + "epoch": 3.588888888888889, + "grad_norm": 1.8745696148294126, + "learning_rate": 1.9813515071745322e-07, + "loss": 0.6847, + "step": 1938 + }, + { + "epoch": 3.590740740740741, + "grad_norm": 1.8021310741794647, + "learning_rate": 1.9813226739499836e-07, + "loss": 0.7186, + "step": 1939 + }, + { + "epoch": 3.5925925925925926, + "grad_norm": 1.509633072480398, + "learning_rate": 1.981293818662636e-07, + "loss": 0.7159, + "step": 1940 + }, + { + "epoch": 3.5944444444444446, + "grad_norm": 1.4675348428612094, + "learning_rate": 1.9812649413131383e-07, + "loss": 0.692, + "step": 1941 + }, + { + "epoch": 3.5962962962962965, + "grad_norm": 1.4097829290374755, + "learning_rate": 1.9812360419021395e-07, + "loss": 0.6862, + "step": 1942 + }, + { + "epoch": 3.598148148148148, + "grad_norm": 1.5781638537942753, + "learning_rate": 1.9812071204302897e-07, + "loss": 0.7129, + "step": 1943 + }, + { + "epoch": 3.6, + "grad_norm": 1.5196045953144095, + "learning_rate": 1.981178176898239e-07, + "loss": 0.7054, + "step": 1944 + }, + { + "epoch": 3.601851851851852, + "grad_norm": 1.8758151840317552, + "learning_rate": 1.9811492113066378e-07, + "loss": 0.7297, + "step": 1945 + }, + { + "epoch": 3.6037037037037036, + "grad_norm": 1.5253808813624143, + "learning_rate": 1.9811202236561377e-07, + "loss": 0.6904, + "step": 1946 + }, + { + "epoch": 3.6055555555555556, + "grad_norm": 1.698065489781154, + "learning_rate": 1.9810912139473904e-07, + "loss": 0.6856, + "step": 1947 + }, + { + "epoch": 3.6074074074074076, + "grad_norm": 3.5386847166983766, + "learning_rate": 1.9810621821810475e-07, + "loss": 0.6707, + "step": 1948 + }, + { + "epoch": 3.609259259259259, + "grad_norm": 1.7630963583130548, + "learning_rate": 1.9810331283577628e-07, + "loss": 0.6897, + "step": 1949 + }, + { + "epoch": 3.611111111111111, + "grad_norm": 1.3857742277969796, + "learning_rate": 1.9810040524781888e-07, + "loss": 0.7382, + "step": 1950 + }, + { + "epoch": 3.612962962962963, + "grad_norm": 1.3734946341593384, + "learning_rate": 1.9809749545429791e-07, + "loss": 0.6744, + "step": 1951 + }, + { + "epoch": 3.6148148148148147, + "grad_norm": 2.303107556487947, + "learning_rate": 1.9809458345527883e-07, + "loss": 0.6553, + "step": 1952 + }, + { + "epoch": 3.6166666666666667, + "grad_norm": 1.4950940892456774, + "learning_rate": 1.980916692508271e-07, + "loss": 0.6463, + "step": 1953 + }, + { + "epoch": 3.6185185185185187, + "grad_norm": 1.7925375390372997, + "learning_rate": 1.9808875284100824e-07, + "loss": 0.6826, + "step": 1954 + }, + { + "epoch": 3.6203703703703702, + "grad_norm": 1.480093862909119, + "learning_rate": 1.9808583422588775e-07, + "loss": 0.692, + "step": 1955 + }, + { + "epoch": 3.6222222222222222, + "grad_norm": 1.6878061639490665, + "learning_rate": 1.9808291340553139e-07, + "loss": 0.7004, + "step": 1956 + }, + { + "epoch": 3.6240740740740742, + "grad_norm": 1.6328820596569467, + "learning_rate": 1.9807999038000468e-07, + "loss": 0.7131, + "step": 1957 + }, + { + "epoch": 3.6259259259259258, + "grad_norm": 1.4413178657128076, + "learning_rate": 1.9807706514937344e-07, + "loss": 0.6968, + "step": 1958 + }, + { + "epoch": 3.6277777777777778, + "grad_norm": 1.4861856575483776, + "learning_rate": 1.980741377137034e-07, + "loss": 0.7087, + "step": 1959 + }, + { + "epoch": 3.6296296296296298, + "grad_norm": 1.450398711422783, + "learning_rate": 1.980712080730604e-07, + "loss": 0.7298, + "step": 1960 + }, + { + "epoch": 3.6314814814814813, + "grad_norm": 1.7686534728701246, + "learning_rate": 1.9806827622751026e-07, + "loss": 0.6793, + "step": 1961 + }, + { + "epoch": 3.6333333333333333, + "grad_norm": 2.0503800147413265, + "learning_rate": 1.9806534217711892e-07, + "loss": 0.704, + "step": 1962 + }, + { + "epoch": 3.6351851851851853, + "grad_norm": 1.4919802579266688, + "learning_rate": 1.9806240592195235e-07, + "loss": 0.7447, + "step": 1963 + }, + { + "epoch": 3.637037037037037, + "grad_norm": 1.3158391307142547, + "learning_rate": 1.9805946746207655e-07, + "loss": 0.6951, + "step": 1964 + }, + { + "epoch": 3.638888888888889, + "grad_norm": 1.387093836355547, + "learning_rate": 1.9805652679755763e-07, + "loss": 0.6782, + "step": 1965 + }, + { + "epoch": 3.640740740740741, + "grad_norm": 1.58426465857652, + "learning_rate": 1.9805358392846164e-07, + "loss": 0.7507, + "step": 1966 + }, + { + "epoch": 3.6425925925925924, + "grad_norm": 1.4062073987628834, + "learning_rate": 1.9805063885485476e-07, + "loss": 0.6919, + "step": 1967 + }, + { + "epoch": 3.6444444444444444, + "grad_norm": 1.4713974122016362, + "learning_rate": 1.9804769157680325e-07, + "loss": 0.7027, + "step": 1968 + }, + { + "epoch": 3.6462962962962964, + "grad_norm": 1.4662203585245022, + "learning_rate": 1.9804474209437333e-07, + "loss": 0.7188, + "step": 1969 + }, + { + "epoch": 3.648148148148148, + "grad_norm": 1.3029363415877, + "learning_rate": 1.9804179040763132e-07, + "loss": 0.7262, + "step": 1970 + }, + { + "epoch": 3.65, + "grad_norm": 1.405182602453171, + "learning_rate": 1.9803883651664357e-07, + "loss": 0.7192, + "step": 1971 + }, + { + "epoch": 3.651851851851852, + "grad_norm": 1.5528130029290417, + "learning_rate": 1.9803588042147653e-07, + "loss": 0.7816, + "step": 1972 + }, + { + "epoch": 3.6537037037037035, + "grad_norm": 1.5805543499120351, + "learning_rate": 1.9803292212219662e-07, + "loss": 0.6829, + "step": 1973 + }, + { + "epoch": 3.6555555555555554, + "grad_norm": 1.5424612361575634, + "learning_rate": 1.980299616188704e-07, + "loss": 0.7299, + "step": 1974 + }, + { + "epoch": 3.6574074074074074, + "grad_norm": 1.4060507652410217, + "learning_rate": 1.9802699891156436e-07, + "loss": 0.6992, + "step": 1975 + }, + { + "epoch": 3.659259259259259, + "grad_norm": 1.7471079807195133, + "learning_rate": 1.9802403400034516e-07, + "loss": 0.6756, + "step": 1976 + }, + { + "epoch": 3.661111111111111, + "grad_norm": 3.0085525503831794, + "learning_rate": 1.9802106688527945e-07, + "loss": 0.6703, + "step": 1977 + }, + { + "epoch": 3.662962962962963, + "grad_norm": 2.1842143747557836, + "learning_rate": 1.9801809756643394e-07, + "loss": 0.6757, + "step": 1978 + }, + { + "epoch": 3.664814814814815, + "grad_norm": 1.634080689590214, + "learning_rate": 1.9801512604387535e-07, + "loss": 0.7206, + "step": 1979 + }, + { + "epoch": 3.6666666666666665, + "grad_norm": 1.56745649562272, + "learning_rate": 1.9801215231767056e-07, + "loss": 0.7439, + "step": 1980 + }, + { + "epoch": 3.6685185185185185, + "grad_norm": 1.3999167261626693, + "learning_rate": 1.9800917638788636e-07, + "loss": 0.7085, + "step": 1981 + }, + { + "epoch": 3.6703703703703705, + "grad_norm": 1.464422738842721, + "learning_rate": 1.980061982545897e-07, + "loss": 0.7237, + "step": 1982 + }, + { + "epoch": 3.6722222222222225, + "grad_norm": 1.520337377827549, + "learning_rate": 1.9800321791784754e-07, + "loss": 0.6891, + "step": 1983 + }, + { + "epoch": 3.674074074074074, + "grad_norm": 1.4504553429395952, + "learning_rate": 1.9800023537772686e-07, + "loss": 0.7088, + "step": 1984 + }, + { + "epoch": 3.675925925925926, + "grad_norm": 1.31576516299285, + "learning_rate": 1.9799725063429472e-07, + "loss": 0.7121, + "step": 1985 + }, + { + "epoch": 3.677777777777778, + "grad_norm": 1.647316184829491, + "learning_rate": 1.979942636876182e-07, + "loss": 0.6851, + "step": 1986 + }, + { + "epoch": 3.6796296296296296, + "grad_norm": 1.5264392525036654, + "learning_rate": 1.979912745377645e-07, + "loss": 0.7135, + "step": 1987 + }, + { + "epoch": 3.6814814814814816, + "grad_norm": 1.4004447779462403, + "learning_rate": 1.9798828318480084e-07, + "loss": 0.7534, + "step": 1988 + }, + { + "epoch": 3.6833333333333336, + "grad_norm": 1.421737883389671, + "learning_rate": 1.979852896287944e-07, + "loss": 0.6945, + "step": 1989 + }, + { + "epoch": 3.685185185185185, + "grad_norm": 1.3769733359853031, + "learning_rate": 1.9798229386981256e-07, + "loss": 0.6964, + "step": 1990 + }, + { + "epoch": 3.687037037037037, + "grad_norm": 1.380796306197519, + "learning_rate": 1.979792959079226e-07, + "loss": 0.6818, + "step": 1991 + }, + { + "epoch": 3.688888888888889, + "grad_norm": 1.507949694068586, + "learning_rate": 1.9797629574319199e-07, + "loss": 0.6516, + "step": 1992 + }, + { + "epoch": 3.6907407407407407, + "grad_norm": 1.3504008732820652, + "learning_rate": 1.9797329337568815e-07, + "loss": 0.6784, + "step": 1993 + }, + { + "epoch": 3.6925925925925926, + "grad_norm": 1.8646438247757418, + "learning_rate": 1.9797028880547857e-07, + "loss": 0.6611, + "step": 1994 + }, + { + "epoch": 3.6944444444444446, + "grad_norm": 1.5331222234717918, + "learning_rate": 1.9796728203263084e-07, + "loss": 0.7425, + "step": 1995 + }, + { + "epoch": 3.696296296296296, + "grad_norm": 1.4674305006650243, + "learning_rate": 1.979642730572125e-07, + "loss": 0.7261, + "step": 1996 + }, + { + "epoch": 3.698148148148148, + "grad_norm": 1.433639727653537, + "learning_rate": 1.9796126187929127e-07, + "loss": 0.7131, + "step": 1997 + }, + { + "epoch": 3.7, + "grad_norm": 1.4766776329393654, + "learning_rate": 1.9795824849893478e-07, + "loss": 0.7274, + "step": 1998 + }, + { + "epoch": 3.7018518518518517, + "grad_norm": 1.5498696619591807, + "learning_rate": 1.9795523291621084e-07, + "loss": 0.6378, + "step": 1999 + }, + { + "epoch": 3.7037037037037037, + "grad_norm": 2.056798683658681, + "learning_rate": 1.979522151311872e-07, + "loss": 0.7054, + "step": 2000 + }, + { + "epoch": 3.7055555555555557, + "grad_norm": 1.9168652757739209, + "learning_rate": 1.9794919514393174e-07, + "loss": 0.6955, + "step": 2001 + }, + { + "epoch": 3.7074074074074073, + "grad_norm": 1.3427405856092864, + "learning_rate": 1.9794617295451236e-07, + "loss": 0.6803, + "step": 2002 + }, + { + "epoch": 3.7092592592592593, + "grad_norm": 1.4538624910105042, + "learning_rate": 1.97943148562997e-07, + "loss": 0.7037, + "step": 2003 + }, + { + "epoch": 3.7111111111111112, + "grad_norm": 1.609870531030301, + "learning_rate": 1.9794012196945359e-07, + "loss": 0.6755, + "step": 2004 + }, + { + "epoch": 3.712962962962963, + "grad_norm": 1.3067156590234197, + "learning_rate": 1.979370931739503e-07, + "loss": 0.6655, + "step": 2005 + }, + { + "epoch": 3.714814814814815, + "grad_norm": 1.4877041837675509, + "learning_rate": 1.9793406217655514e-07, + "loss": 0.7153, + "step": 2006 + }, + { + "epoch": 3.716666666666667, + "grad_norm": 1.4331058594994008, + "learning_rate": 1.9793102897733628e-07, + "loss": 0.6593, + "step": 2007 + }, + { + "epoch": 3.7185185185185183, + "grad_norm": 1.4166386806150968, + "learning_rate": 1.979279935763619e-07, + "loss": 0.6745, + "step": 2008 + }, + { + "epoch": 3.7203703703703703, + "grad_norm": 1.5513582114294344, + "learning_rate": 1.9792495597370028e-07, + "loss": 0.6914, + "step": 2009 + }, + { + "epoch": 3.7222222222222223, + "grad_norm": 1.6670143178468493, + "learning_rate": 1.9792191616941967e-07, + "loss": 0.7328, + "step": 2010 + }, + { + "epoch": 3.724074074074074, + "grad_norm": 1.57401772482662, + "learning_rate": 1.9791887416358843e-07, + "loss": 0.7114, + "step": 2011 + }, + { + "epoch": 3.725925925925926, + "grad_norm": 1.521960752005683, + "learning_rate": 1.9791582995627497e-07, + "loss": 0.6835, + "step": 2012 + }, + { + "epoch": 3.727777777777778, + "grad_norm": 1.5065661637057683, + "learning_rate": 1.9791278354754773e-07, + "loss": 0.688, + "step": 2013 + }, + { + "epoch": 3.7296296296296294, + "grad_norm": 1.943834273685275, + "learning_rate": 1.9790973493747515e-07, + "loss": 0.7396, + "step": 2014 + }, + { + "epoch": 3.7314814814814814, + "grad_norm": 1.8792028376394398, + "learning_rate": 1.9790668412612584e-07, + "loss": 0.7555, + "step": 2015 + }, + { + "epoch": 3.7333333333333334, + "grad_norm": 1.4512496006177689, + "learning_rate": 1.9790363111356836e-07, + "loss": 0.7064, + "step": 2016 + }, + { + "epoch": 3.735185185185185, + "grad_norm": 1.521998551107538, + "learning_rate": 1.9790057589987135e-07, + "loss": 0.707, + "step": 2017 + }, + { + "epoch": 3.737037037037037, + "grad_norm": 2.4848097369119615, + "learning_rate": 1.9789751848510349e-07, + "loss": 0.6751, + "step": 2018 + }, + { + "epoch": 3.738888888888889, + "grad_norm": 1.5320525142519865, + "learning_rate": 1.978944588693335e-07, + "loss": 0.6554, + "step": 2019 + }, + { + "epoch": 3.7407407407407405, + "grad_norm": 1.5317323963504297, + "learning_rate": 1.9789139705263025e-07, + "loss": 0.7245, + "step": 2020 + }, + { + "epoch": 3.7425925925925925, + "grad_norm": 1.4848651542471398, + "learning_rate": 1.978883330350625e-07, + "loss": 0.6909, + "step": 2021 + }, + { + "epoch": 3.7444444444444445, + "grad_norm": 1.3011226506723446, + "learning_rate": 1.9788526681669918e-07, + "loss": 0.7162, + "step": 2022 + }, + { + "epoch": 3.7462962962962965, + "grad_norm": 2.0603608025329914, + "learning_rate": 1.9788219839760918e-07, + "loss": 0.6976, + "step": 2023 + }, + { + "epoch": 3.748148148148148, + "grad_norm": 1.5721552562851024, + "learning_rate": 1.9787912777786154e-07, + "loss": 0.693, + "step": 2024 + }, + { + "epoch": 3.75, + "grad_norm": 1.776457099525263, + "learning_rate": 1.9787605495752527e-07, + "loss": 0.6921, + "step": 2025 + }, + { + "epoch": 3.751851851851852, + "grad_norm": 1.4995937563146968, + "learning_rate": 1.9787297993666943e-07, + "loss": 0.68, + "step": 2026 + }, + { + "epoch": 3.7537037037037035, + "grad_norm": 1.5205568125459694, + "learning_rate": 1.9786990271536318e-07, + "loss": 0.7253, + "step": 2027 + }, + { + "epoch": 3.7555555555555555, + "grad_norm": 1.5091086882895302, + "learning_rate": 1.9786682329367575e-07, + "loss": 0.712, + "step": 2028 + }, + { + "epoch": 3.7574074074074075, + "grad_norm": 1.7870643678695348, + "learning_rate": 1.978637416716763e-07, + "loss": 0.6854, + "step": 2029 + }, + { + "epoch": 3.7592592592592595, + "grad_norm": 1.911121254209067, + "learning_rate": 1.9786065784943415e-07, + "loss": 0.696, + "step": 2030 + }, + { + "epoch": 3.761111111111111, + "grad_norm": 3.5116070925072425, + "learning_rate": 1.9785757182701862e-07, + "loss": 0.6973, + "step": 2031 + }, + { + "epoch": 3.762962962962963, + "grad_norm": 1.288370880172143, + "learning_rate": 1.9785448360449912e-07, + "loss": 0.717, + "step": 2032 + }, + { + "epoch": 3.764814814814815, + "grad_norm": 1.4654645664311308, + "learning_rate": 1.9785139318194504e-07, + "loss": 0.6982, + "step": 2033 + }, + { + "epoch": 3.7666666666666666, + "grad_norm": 1.4875349563135496, + "learning_rate": 1.978483005594259e-07, + "loss": 0.6417, + "step": 2034 + }, + { + "epoch": 3.7685185185185186, + "grad_norm": 1.4903407458896933, + "learning_rate": 1.978452057370112e-07, + "loss": 0.706, + "step": 2035 + }, + { + "epoch": 3.7703703703703706, + "grad_norm": 1.3498656727464402, + "learning_rate": 1.9784210871477053e-07, + "loss": 0.7015, + "step": 2036 + }, + { + "epoch": 3.772222222222222, + "grad_norm": 1.4561177158041902, + "learning_rate": 1.978390094927735e-07, + "loss": 0.7258, + "step": 2037 + }, + { + "epoch": 3.774074074074074, + "grad_norm": 1.5056345130887145, + "learning_rate": 1.9783590807108984e-07, + "loss": 0.6856, + "step": 2038 + }, + { + "epoch": 3.775925925925926, + "grad_norm": 1.3636144780062167, + "learning_rate": 1.9783280444978924e-07, + "loss": 0.6788, + "step": 2039 + }, + { + "epoch": 3.7777777777777777, + "grad_norm": 1.5763413064558116, + "learning_rate": 1.9782969862894148e-07, + "loss": 0.7254, + "step": 2040 + }, + { + "epoch": 3.7796296296296297, + "grad_norm": 1.5831828661719225, + "learning_rate": 1.9782659060861642e-07, + "loss": 0.7085, + "step": 2041 + }, + { + "epoch": 3.7814814814814817, + "grad_norm": 1.559301490312729, + "learning_rate": 1.9782348038888387e-07, + "loss": 0.7009, + "step": 2042 + }, + { + "epoch": 3.783333333333333, + "grad_norm": 1.9807510684118732, + "learning_rate": 1.9782036796981383e-07, + "loss": 0.7166, + "step": 2043 + }, + { + "epoch": 3.785185185185185, + "grad_norm": 1.9191446790891853, + "learning_rate": 1.9781725335147624e-07, + "loss": 0.6643, + "step": 2044 + }, + { + "epoch": 3.787037037037037, + "grad_norm": 3.1126646602583503, + "learning_rate": 1.978141365339411e-07, + "loss": 0.6856, + "step": 2045 + }, + { + "epoch": 3.7888888888888888, + "grad_norm": 2.337472437667415, + "learning_rate": 1.978110175172785e-07, + "loss": 0.7104, + "step": 2046 + }, + { + "epoch": 3.7907407407407407, + "grad_norm": 1.7197943963713567, + "learning_rate": 1.9780789630155862e-07, + "loss": 0.6728, + "step": 2047 + }, + { + "epoch": 3.7925925925925927, + "grad_norm": 1.6202242251905143, + "learning_rate": 1.9780477288685158e-07, + "loss": 0.6419, + "step": 2048 + }, + { + "epoch": 3.7944444444444443, + "grad_norm": 1.4939723533886449, + "learning_rate": 1.9780164727322758e-07, + "loss": 0.6997, + "step": 2049 + }, + { + "epoch": 3.7962962962962963, + "grad_norm": 1.600820817558738, + "learning_rate": 1.9779851946075695e-07, + "loss": 0.7045, + "step": 2050 + }, + { + "epoch": 3.7981481481481483, + "grad_norm": 1.4179458361924653, + "learning_rate": 1.9779538944951e-07, + "loss": 0.6433, + "step": 2051 + }, + { + "epoch": 3.8, + "grad_norm": 1.408718324432222, + "learning_rate": 1.9779225723955707e-07, + "loss": 0.7133, + "step": 2052 + }, + { + "epoch": 3.801851851851852, + "grad_norm": 1.705288026674248, + "learning_rate": 1.977891228309686e-07, + "loss": 0.7153, + "step": 2053 + }, + { + "epoch": 3.803703703703704, + "grad_norm": 1.5061348743499523, + "learning_rate": 1.9778598622381503e-07, + "loss": 0.7049, + "step": 2054 + }, + { + "epoch": 3.8055555555555554, + "grad_norm": 1.4265299231778683, + "learning_rate": 1.9778284741816692e-07, + "loss": 0.6546, + "step": 2055 + }, + { + "epoch": 3.8074074074074074, + "grad_norm": 1.3183671470338603, + "learning_rate": 1.9777970641409485e-07, + "loss": 0.7073, + "step": 2056 + }, + { + "epoch": 3.8092592592592593, + "grad_norm": 1.5032706601594712, + "learning_rate": 1.977765632116694e-07, + "loss": 0.7246, + "step": 2057 + }, + { + "epoch": 3.811111111111111, + "grad_norm": 1.3855457654491812, + "learning_rate": 1.9777341781096127e-07, + "loss": 0.734, + "step": 2058 + }, + { + "epoch": 3.812962962962963, + "grad_norm": 1.6538440986071663, + "learning_rate": 1.9777027021204112e-07, + "loss": 0.6826, + "step": 2059 + }, + { + "epoch": 3.814814814814815, + "grad_norm": 1.4233937776207124, + "learning_rate": 1.9776712041497977e-07, + "loss": 0.7046, + "step": 2060 + }, + { + "epoch": 3.8166666666666664, + "grad_norm": 2.088124984233603, + "learning_rate": 1.9776396841984804e-07, + "loss": 0.6707, + "step": 2061 + }, + { + "epoch": 3.8185185185185184, + "grad_norm": 1.5372764521498996, + "learning_rate": 1.9776081422671677e-07, + "loss": 0.7223, + "step": 2062 + }, + { + "epoch": 3.8203703703703704, + "grad_norm": 1.6259154880893407, + "learning_rate": 1.977576578356569e-07, + "loss": 0.731, + "step": 2063 + }, + { + "epoch": 3.822222222222222, + "grad_norm": 1.35635294757172, + "learning_rate": 1.9775449924673932e-07, + "loss": 0.6952, + "step": 2064 + }, + { + "epoch": 3.824074074074074, + "grad_norm": 1.412001533714791, + "learning_rate": 1.9775133846003515e-07, + "loss": 0.7235, + "step": 2065 + }, + { + "epoch": 3.825925925925926, + "grad_norm": 1.4062812757225207, + "learning_rate": 1.9774817547561538e-07, + "loss": 0.723, + "step": 2066 + }, + { + "epoch": 3.8277777777777775, + "grad_norm": 1.6023306834626792, + "learning_rate": 1.9774501029355116e-07, + "loss": 0.7002, + "step": 2067 + }, + { + "epoch": 3.8296296296296295, + "grad_norm": 1.4951150910164601, + "learning_rate": 1.9774184291391362e-07, + "loss": 0.704, + "step": 2068 + }, + { + "epoch": 3.8314814814814815, + "grad_norm": 1.5231815140564111, + "learning_rate": 1.97738673336774e-07, + "loss": 0.7131, + "step": 2069 + }, + { + "epoch": 3.8333333333333335, + "grad_norm": 1.5957682017314692, + "learning_rate": 1.9773550156220354e-07, + "loss": 0.6992, + "step": 2070 + }, + { + "epoch": 3.835185185185185, + "grad_norm": 2.0166025999711525, + "learning_rate": 1.9773232759027355e-07, + "loss": 0.6928, + "step": 2071 + }, + { + "epoch": 3.837037037037037, + "grad_norm": 1.571719310867269, + "learning_rate": 1.9772915142105542e-07, + "loss": 0.6729, + "step": 2072 + }, + { + "epoch": 3.838888888888889, + "grad_norm": 1.3956578120181538, + "learning_rate": 1.9772597305462054e-07, + "loss": 0.7294, + "step": 2073 + }, + { + "epoch": 3.840740740740741, + "grad_norm": 1.4021165919239085, + "learning_rate": 1.9772279249104037e-07, + "loss": 0.727, + "step": 2074 + }, + { + "epoch": 3.8425925925925926, + "grad_norm": 1.4928251728343354, + "learning_rate": 1.9771960973038638e-07, + "loss": 0.698, + "step": 2075 + }, + { + "epoch": 3.8444444444444446, + "grad_norm": 1.3185730537474556, + "learning_rate": 1.9771642477273018e-07, + "loss": 0.7006, + "step": 2076 + }, + { + "epoch": 3.8462962962962965, + "grad_norm": 1.2587927776882453, + "learning_rate": 1.9771323761814337e-07, + "loss": 0.7173, + "step": 2077 + }, + { + "epoch": 3.848148148148148, + "grad_norm": 1.4150570154501556, + "learning_rate": 1.9771004826669757e-07, + "loss": 0.6956, + "step": 2078 + }, + { + "epoch": 3.85, + "grad_norm": 1.3471200353892092, + "learning_rate": 1.9770685671846453e-07, + "loss": 0.7201, + "step": 2079 + }, + { + "epoch": 3.851851851851852, + "grad_norm": 1.6364455422132873, + "learning_rate": 1.9770366297351598e-07, + "loss": 0.7436, + "step": 2080 + }, + { + "epoch": 3.8537037037037036, + "grad_norm": 1.5439290186514478, + "learning_rate": 1.9770046703192372e-07, + "loss": 0.7223, + "step": 2081 + }, + { + "epoch": 3.8555555555555556, + "grad_norm": 1.3750221231896445, + "learning_rate": 1.9769726889375962e-07, + "loss": 0.6436, + "step": 2082 + }, + { + "epoch": 3.8574074074074076, + "grad_norm": 1.4648862148395114, + "learning_rate": 1.9769406855909556e-07, + "loss": 0.6973, + "step": 2083 + }, + { + "epoch": 3.859259259259259, + "grad_norm": 1.5109337836145946, + "learning_rate": 1.9769086602800353e-07, + "loss": 0.713, + "step": 2084 + }, + { + "epoch": 3.861111111111111, + "grad_norm": 1.369789548190783, + "learning_rate": 1.9768766130055548e-07, + "loss": 0.7304, + "step": 2085 + }, + { + "epoch": 3.862962962962963, + "grad_norm": 1.4954309097823089, + "learning_rate": 1.976844543768235e-07, + "loss": 0.6516, + "step": 2086 + }, + { + "epoch": 3.8648148148148147, + "grad_norm": 1.3887369520444244, + "learning_rate": 1.9768124525687968e-07, + "loss": 0.7136, + "step": 2087 + }, + { + "epoch": 3.8666666666666667, + "grad_norm": 1.4408895085668612, + "learning_rate": 1.9767803394079615e-07, + "loss": 0.7069, + "step": 2088 + }, + { + "epoch": 3.8685185185185187, + "grad_norm": 1.9529059632425914, + "learning_rate": 1.9767482042864514e-07, + "loss": 0.6825, + "step": 2089 + }, + { + "epoch": 3.8703703703703702, + "grad_norm": 1.9033095152552635, + "learning_rate": 1.976716047204989e-07, + "loss": 0.6625, + "step": 2090 + }, + { + "epoch": 3.8722222222222222, + "grad_norm": 1.4464385013475276, + "learning_rate": 1.976683868164297e-07, + "loss": 0.6889, + "step": 2091 + }, + { + "epoch": 3.8740740740740742, + "grad_norm": 1.7461217796006685, + "learning_rate": 1.976651667165099e-07, + "loss": 0.687, + "step": 2092 + }, + { + "epoch": 3.8759259259259258, + "grad_norm": 1.5172342151832177, + "learning_rate": 1.976619444208119e-07, + "loss": 0.7089, + "step": 2093 + }, + { + "epoch": 3.8777777777777778, + "grad_norm": 1.9882867582913393, + "learning_rate": 1.9765871992940814e-07, + "loss": 0.7023, + "step": 2094 + }, + { + "epoch": 3.8796296296296298, + "grad_norm": 1.3665506673574688, + "learning_rate": 1.976554932423711e-07, + "loss": 0.687, + "step": 2095 + }, + { + "epoch": 3.8814814814814813, + "grad_norm": 1.39428078518463, + "learning_rate": 1.976522643597734e-07, + "loss": 0.7245, + "step": 2096 + }, + { + "epoch": 3.8833333333333333, + "grad_norm": 1.6418300696502135, + "learning_rate": 1.9764903328168754e-07, + "loss": 0.6876, + "step": 2097 + }, + { + "epoch": 3.8851851851851853, + "grad_norm": 1.4648251915385586, + "learning_rate": 1.9764580000818621e-07, + "loss": 0.7176, + "step": 2098 + }, + { + "epoch": 3.887037037037037, + "grad_norm": 1.4366369843777507, + "learning_rate": 1.9764256453934206e-07, + "loss": 0.6957, + "step": 2099 + }, + { + "epoch": 3.888888888888889, + "grad_norm": 1.532200763932793, + "learning_rate": 1.9763932687522793e-07, + "loss": 0.707, + "step": 2100 + }, + { + "epoch": 3.890740740740741, + "grad_norm": 1.754310847101926, + "learning_rate": 1.9763608701591652e-07, + "loss": 0.6561, + "step": 2101 + }, + { + "epoch": 3.8925925925925924, + "grad_norm": 1.7106571761424132, + "learning_rate": 1.9763284496148067e-07, + "loss": 0.7514, + "step": 2102 + }, + { + "epoch": 3.8944444444444444, + "grad_norm": 1.33344946077013, + "learning_rate": 1.9762960071199333e-07, + "loss": 0.6714, + "step": 2103 + }, + { + "epoch": 3.8962962962962964, + "grad_norm": 1.3480036027435867, + "learning_rate": 1.9762635426752738e-07, + "loss": 0.6785, + "step": 2104 + }, + { + "epoch": 3.898148148148148, + "grad_norm": 1.535732902513717, + "learning_rate": 1.9762310562815588e-07, + "loss": 0.7003, + "step": 2105 + }, + { + "epoch": 3.9, + "grad_norm": 2.3342955607009253, + "learning_rate": 1.9761985479395177e-07, + "loss": 0.6553, + "step": 2106 + }, + { + "epoch": 3.901851851851852, + "grad_norm": 1.4494541848641362, + "learning_rate": 1.9761660176498825e-07, + "loss": 0.7365, + "step": 2107 + }, + { + "epoch": 3.9037037037037035, + "grad_norm": 1.5628178216199449, + "learning_rate": 1.9761334654133835e-07, + "loss": 0.7124, + "step": 2108 + }, + { + "epoch": 3.9055555555555554, + "grad_norm": 2.2697898769289284, + "learning_rate": 1.9761008912307533e-07, + "loss": 0.6823, + "step": 2109 + }, + { + "epoch": 3.9074074074074074, + "grad_norm": 1.781464621555349, + "learning_rate": 1.976068295102724e-07, + "loss": 0.665, + "step": 2110 + }, + { + "epoch": 3.909259259259259, + "grad_norm": 1.3942453117716922, + "learning_rate": 1.9760356770300285e-07, + "loss": 0.7116, + "step": 2111 + }, + { + "epoch": 3.911111111111111, + "grad_norm": 1.4152386658428937, + "learning_rate": 1.9760030370134e-07, + "loss": 0.6577, + "step": 2112 + }, + { + "epoch": 3.912962962962963, + "grad_norm": 1.4363132682118671, + "learning_rate": 1.9759703750535727e-07, + "loss": 0.7228, + "step": 2113 + }, + { + "epoch": 3.914814814814815, + "grad_norm": 1.856928142706703, + "learning_rate": 1.9759376911512805e-07, + "loss": 0.7051, + "step": 2114 + }, + { + "epoch": 3.9166666666666665, + "grad_norm": 1.5644776469398753, + "learning_rate": 1.9759049853072582e-07, + "loss": 0.6926, + "step": 2115 + }, + { + "epoch": 3.9185185185185185, + "grad_norm": 1.8180639546252566, + "learning_rate": 1.9758722575222416e-07, + "loss": 0.7115, + "step": 2116 + }, + { + "epoch": 3.9203703703703705, + "grad_norm": 1.5593164389046483, + "learning_rate": 1.9758395077969664e-07, + "loss": 0.7199, + "step": 2117 + }, + { + "epoch": 3.9222222222222225, + "grad_norm": 1.4475018495526304, + "learning_rate": 1.9758067361321681e-07, + "loss": 0.6925, + "step": 2118 + }, + { + "epoch": 3.924074074074074, + "grad_norm": 1.456726814484005, + "learning_rate": 1.9757739425285847e-07, + "loss": 0.6887, + "step": 2119 + }, + { + "epoch": 3.925925925925926, + "grad_norm": 1.45533907653281, + "learning_rate": 1.9757411269869527e-07, + "loss": 0.7021, + "step": 2120 + }, + { + "epoch": 3.927777777777778, + "grad_norm": 1.32284728498338, + "learning_rate": 1.97570828950801e-07, + "loss": 0.7028, + "step": 2121 + }, + { + "epoch": 3.9296296296296296, + "grad_norm": 1.7725585450558663, + "learning_rate": 1.975675430092495e-07, + "loss": 0.6658, + "step": 2122 + }, + { + "epoch": 3.9314814814814816, + "grad_norm": 1.4115818603619108, + "learning_rate": 1.9756425487411463e-07, + "loss": 0.6653, + "step": 2123 + }, + { + "epoch": 3.9333333333333336, + "grad_norm": 1.5111266002908923, + "learning_rate": 1.9756096454547038e-07, + "loss": 0.6809, + "step": 2124 + }, + { + "epoch": 3.935185185185185, + "grad_norm": 2.1484609200607245, + "learning_rate": 1.9755767202339063e-07, + "loss": 0.6965, + "step": 2125 + }, + { + "epoch": 3.937037037037037, + "grad_norm": 1.9629451950219259, + "learning_rate": 1.9755437730794946e-07, + "loss": 0.6781, + "step": 2126 + }, + { + "epoch": 3.938888888888889, + "grad_norm": 1.4073283972031392, + "learning_rate": 1.9755108039922094e-07, + "loss": 0.679, + "step": 2127 + }, + { + "epoch": 3.9407407407407407, + "grad_norm": 1.5044559613728987, + "learning_rate": 1.9754778129727917e-07, + "loss": 0.6975, + "step": 2128 + }, + { + "epoch": 3.9425925925925926, + "grad_norm": 1.4781379310574028, + "learning_rate": 1.9754448000219837e-07, + "loss": 0.6524, + "step": 2129 + }, + { + "epoch": 3.9444444444444446, + "grad_norm": 1.7175765794120874, + "learning_rate": 1.9754117651405271e-07, + "loss": 0.7085, + "step": 2130 + }, + { + "epoch": 3.946296296296296, + "grad_norm": 1.3991352643221722, + "learning_rate": 1.9753787083291652e-07, + "loss": 0.7439, + "step": 2131 + }, + { + "epoch": 3.948148148148148, + "grad_norm": 1.3898289272509385, + "learning_rate": 1.9753456295886403e-07, + "loss": 0.7004, + "step": 2132 + }, + { + "epoch": 3.95, + "grad_norm": 1.549627942835458, + "learning_rate": 1.975312528919697e-07, + "loss": 0.6927, + "step": 2133 + }, + { + "epoch": 3.9518518518518517, + "grad_norm": 1.4378032393098668, + "learning_rate": 1.975279406323079e-07, + "loss": 0.6969, + "step": 2134 + }, + { + "epoch": 3.9537037037037037, + "grad_norm": 1.3791988285978989, + "learning_rate": 1.975246261799531e-07, + "loss": 0.6965, + "step": 2135 + }, + { + "epoch": 3.9555555555555557, + "grad_norm": 1.481814109095302, + "learning_rate": 1.9752130953497986e-07, + "loss": 0.6934, + "step": 2136 + }, + { + "epoch": 3.9574074074074073, + "grad_norm": 1.5090948159860287, + "learning_rate": 1.975179906974627e-07, + "loss": 0.7053, + "step": 2137 + }, + { + "epoch": 3.9592592592592593, + "grad_norm": 1.4225543850617801, + "learning_rate": 1.9751466966747628e-07, + "loss": 0.7022, + "step": 2138 + }, + { + "epoch": 3.9611111111111112, + "grad_norm": 1.6541347671954156, + "learning_rate": 1.975113464450952e-07, + "loss": 0.6823, + "step": 2139 + }, + { + "epoch": 3.962962962962963, + "grad_norm": 1.6594650372192261, + "learning_rate": 1.975080210303943e-07, + "loss": 0.719, + "step": 2140 + }, + { + "epoch": 3.964814814814815, + "grad_norm": 1.9309095868089667, + "learning_rate": 1.9750469342344815e-07, + "loss": 0.7146, + "step": 2141 + }, + { + "epoch": 3.966666666666667, + "grad_norm": 1.6333450369742117, + "learning_rate": 1.9750136362433174e-07, + "loss": 0.6571, + "step": 2142 + }, + { + "epoch": 3.9685185185185183, + "grad_norm": 1.7329951546550835, + "learning_rate": 1.9749803163311985e-07, + "loss": 0.6623, + "step": 2143 + }, + { + "epoch": 3.9703703703703703, + "grad_norm": 1.5532786104958647, + "learning_rate": 1.974946974498874e-07, + "loss": 0.7035, + "step": 2144 + }, + { + "epoch": 3.9722222222222223, + "grad_norm": 1.6323745480348946, + "learning_rate": 1.9749136107470936e-07, + "loss": 0.6795, + "step": 2145 + }, + { + "epoch": 3.974074074074074, + "grad_norm": 1.3989398802253061, + "learning_rate": 1.9748802250766075e-07, + "loss": 0.6797, + "step": 2146 + }, + { + "epoch": 3.975925925925926, + "grad_norm": 1.7248875519213498, + "learning_rate": 1.974846817488166e-07, + "loss": 0.7201, + "step": 2147 + }, + { + "epoch": 3.977777777777778, + "grad_norm": 1.3932903335531583, + "learning_rate": 1.9748133879825206e-07, + "loss": 0.6729, + "step": 2148 + }, + { + "epoch": 3.9796296296296294, + "grad_norm": 1.5604056151476517, + "learning_rate": 1.9747799365604228e-07, + "loss": 0.6769, + "step": 2149 + }, + { + "epoch": 3.9814814814814814, + "grad_norm": 1.548387860802549, + "learning_rate": 1.9747464632226242e-07, + "loss": 0.7067, + "step": 2150 + }, + { + "epoch": 3.9833333333333334, + "grad_norm": 1.5999215115914402, + "learning_rate": 1.9747129679698783e-07, + "loss": 0.7391, + "step": 2151 + }, + { + "epoch": 3.985185185185185, + "grad_norm": 1.2913727032002287, + "learning_rate": 1.9746794508029368e-07, + "loss": 0.6915, + "step": 2152 + }, + { + "epoch": 3.987037037037037, + "grad_norm": 1.7733173594819152, + "learning_rate": 1.9746459117225548e-07, + "loss": 0.6658, + "step": 2153 + }, + { + "epoch": 3.988888888888889, + "grad_norm": 1.7798418460077214, + "learning_rate": 1.9746123507294852e-07, + "loss": 0.6677, + "step": 2154 + }, + { + "epoch": 3.9907407407407405, + "grad_norm": 1.5272738685494236, + "learning_rate": 1.974578767824483e-07, + "loss": 0.684, + "step": 2155 + }, + { + "epoch": 3.9925925925925925, + "grad_norm": 1.6031112127581029, + "learning_rate": 1.9745451630083032e-07, + "loss": 0.6936, + "step": 2156 + }, + { + "epoch": 3.9944444444444445, + "grad_norm": 1.5080577083063602, + "learning_rate": 1.9745115362817013e-07, + "loss": 0.6984, + "step": 2157 + }, + { + "epoch": 3.9962962962962965, + "grad_norm": 1.5029217673104314, + "learning_rate": 1.9744778876454333e-07, + "loss": 0.7014, + "step": 2158 + }, + { + "epoch": 3.998148148148148, + "grad_norm": 1.5298356743188082, + "learning_rate": 1.974444217100256e-07, + "loss": 0.7117, + "step": 2159 + }, + { + "epoch": 4.0, + "grad_norm": 1.6547944841668067, + "learning_rate": 1.974410524646926e-07, + "loss": 0.6649, + "step": 2160 + }, + { + "epoch": 4.0018518518518515, + "grad_norm": 1.4615931784732643, + "learning_rate": 1.9743768102862008e-07, + "loss": 0.7127, + "step": 2161 + }, + { + "epoch": 4.003703703703704, + "grad_norm": 1.6981235427227295, + "learning_rate": 1.9743430740188388e-07, + "loss": 0.7041, + "step": 2162 + }, + { + "epoch": 4.0055555555555555, + "grad_norm": 1.579588220028123, + "learning_rate": 1.9743093158455978e-07, + "loss": 0.681, + "step": 2163 + }, + { + "epoch": 4.007407407407407, + "grad_norm": 1.6718687893799213, + "learning_rate": 1.9742755357672375e-07, + "loss": 0.7043, + "step": 2164 + }, + { + "epoch": 4.0092592592592595, + "grad_norm": 1.5367334488998505, + "learning_rate": 1.9742417337845173e-07, + "loss": 0.7381, + "step": 2165 + }, + { + "epoch": 4.011111111111111, + "grad_norm": 1.5151069959415868, + "learning_rate": 1.9742079098981966e-07, + "loss": 0.6807, + "step": 2166 + }, + { + "epoch": 4.012962962962963, + "grad_norm": 4.656958498627862, + "learning_rate": 1.9741740641090364e-07, + "loss": 0.6555, + "step": 2167 + }, + { + "epoch": 4.014814814814815, + "grad_norm": 1.3947326597756573, + "learning_rate": 1.974140196417797e-07, + "loss": 0.7133, + "step": 2168 + }, + { + "epoch": 4.016666666666667, + "grad_norm": 1.4680308663458679, + "learning_rate": 1.9741063068252406e-07, + "loss": 0.6907, + "step": 2169 + }, + { + "epoch": 4.018518518518518, + "grad_norm": 1.5330549305361343, + "learning_rate": 1.974072395332129e-07, + "loss": 0.6882, + "step": 2170 + }, + { + "epoch": 4.020370370370371, + "grad_norm": 1.4826289867980418, + "learning_rate": 1.9740384619392238e-07, + "loss": 0.6796, + "step": 2171 + }, + { + "epoch": 4.022222222222222, + "grad_norm": 2.895970942377341, + "learning_rate": 1.9740045066472888e-07, + "loss": 0.6617, + "step": 2172 + }, + { + "epoch": 4.024074074074074, + "grad_norm": 1.715138901443854, + "learning_rate": 1.9739705294570872e-07, + "loss": 0.6921, + "step": 2173 + }, + { + "epoch": 4.025925925925926, + "grad_norm": 1.661284148376113, + "learning_rate": 1.9739365303693827e-07, + "loss": 0.677, + "step": 2174 + }, + { + "epoch": 4.027777777777778, + "grad_norm": 1.401716886576476, + "learning_rate": 1.97390250938494e-07, + "loss": 0.6531, + "step": 2175 + }, + { + "epoch": 4.029629629629629, + "grad_norm": 1.468490670459241, + "learning_rate": 1.9738684665045237e-07, + "loss": 0.7035, + "step": 2176 + }, + { + "epoch": 4.031481481481482, + "grad_norm": 1.5817868846639522, + "learning_rate": 1.9738344017288996e-07, + "loss": 0.6696, + "step": 2177 + }, + { + "epoch": 4.033333333333333, + "grad_norm": 1.671296048466896, + "learning_rate": 1.973800315058833e-07, + "loss": 0.6557, + "step": 2178 + }, + { + "epoch": 4.035185185185185, + "grad_norm": 1.452127251710725, + "learning_rate": 1.9737662064950905e-07, + "loss": 0.7101, + "step": 2179 + }, + { + "epoch": 4.037037037037037, + "grad_norm": 1.4707461133550865, + "learning_rate": 1.9737320760384387e-07, + "loss": 0.7459, + "step": 2180 + }, + { + "epoch": 4.038888888888889, + "grad_norm": 1.4109725872193846, + "learning_rate": 1.9736979236896457e-07, + "loss": 0.6851, + "step": 2181 + }, + { + "epoch": 4.040740740740741, + "grad_norm": 1.2909741955074516, + "learning_rate": 1.9736637494494787e-07, + "loss": 0.6944, + "step": 2182 + }, + { + "epoch": 4.042592592592593, + "grad_norm": 1.4539069373199944, + "learning_rate": 1.9736295533187062e-07, + "loss": 0.6853, + "step": 2183 + }, + { + "epoch": 4.044444444444444, + "grad_norm": 1.4381460541649138, + "learning_rate": 1.9735953352980967e-07, + "loss": 0.6961, + "step": 2184 + }, + { + "epoch": 4.046296296296297, + "grad_norm": 1.5695544992835817, + "learning_rate": 1.97356109538842e-07, + "loss": 0.7177, + "step": 2185 + }, + { + "epoch": 4.048148148148148, + "grad_norm": 1.6211469964245664, + "learning_rate": 1.973526833590446e-07, + "loss": 0.6877, + "step": 2186 + }, + { + "epoch": 4.05, + "grad_norm": 1.2867753143903538, + "learning_rate": 1.9734925499049444e-07, + "loss": 0.6489, + "step": 2187 + }, + { + "epoch": 4.051851851851852, + "grad_norm": 1.909640708661203, + "learning_rate": 1.9734582443326862e-07, + "loss": 0.6674, + "step": 2188 + }, + { + "epoch": 4.053703703703704, + "grad_norm": 2.1438291234846467, + "learning_rate": 1.9734239168744433e-07, + "loss": 0.7147, + "step": 2189 + }, + { + "epoch": 4.055555555555555, + "grad_norm": 2.0191059311475974, + "learning_rate": 1.9733895675309865e-07, + "loss": 0.6962, + "step": 2190 + }, + { + "epoch": 4.057407407407408, + "grad_norm": 1.3273764517973292, + "learning_rate": 1.973355196303089e-07, + "loss": 0.7137, + "step": 2191 + }, + { + "epoch": 4.059259259259259, + "grad_norm": 1.403757029300068, + "learning_rate": 1.9733208031915226e-07, + "loss": 0.6906, + "step": 2192 + }, + { + "epoch": 4.061111111111111, + "grad_norm": 1.3260719766271234, + "learning_rate": 1.9732863881970612e-07, + "loss": 0.679, + "step": 2193 + }, + { + "epoch": 4.062962962962963, + "grad_norm": 1.7334559671238368, + "learning_rate": 1.9732519513204785e-07, + "loss": 0.6547, + "step": 2194 + }, + { + "epoch": 4.064814814814815, + "grad_norm": 1.4306997267146155, + "learning_rate": 1.9732174925625485e-07, + "loss": 0.6502, + "step": 2195 + }, + { + "epoch": 4.066666666666666, + "grad_norm": 1.7511762120130396, + "learning_rate": 1.9731830119240462e-07, + "loss": 0.6647, + "step": 2196 + }, + { + "epoch": 4.068518518518519, + "grad_norm": 1.6338369370694825, + "learning_rate": 1.9731485094057466e-07, + "loss": 0.7323, + "step": 2197 + }, + { + "epoch": 4.07037037037037, + "grad_norm": 2.1227467076699456, + "learning_rate": 1.9731139850084254e-07, + "loss": 0.7332, + "step": 2198 + }, + { + "epoch": 4.072222222222222, + "grad_norm": 1.5372721559203275, + "learning_rate": 1.9730794387328587e-07, + "loss": 0.6605, + "step": 2199 + }, + { + "epoch": 4.074074074074074, + "grad_norm": 1.5656828327346766, + "learning_rate": 1.9730448705798237e-07, + "loss": 0.6717, + "step": 2200 + }, + { + "epoch": 4.075925925925926, + "grad_norm": 2.1763961935398024, + "learning_rate": 1.9730102805500973e-07, + "loss": 0.6928, + "step": 2201 + }, + { + "epoch": 4.0777777777777775, + "grad_norm": 1.5406319524000287, + "learning_rate": 1.972975668644457e-07, + "loss": 0.7133, + "step": 2202 + }, + { + "epoch": 4.07962962962963, + "grad_norm": 1.5334006906455309, + "learning_rate": 1.9729410348636811e-07, + "loss": 0.7402, + "step": 2203 + }, + { + "epoch": 4.0814814814814815, + "grad_norm": 1.3507887777965895, + "learning_rate": 1.972906379208548e-07, + "loss": 0.6864, + "step": 2204 + }, + { + "epoch": 4.083333333333333, + "grad_norm": 1.4709371853705544, + "learning_rate": 1.9728717016798378e-07, + "loss": 0.728, + "step": 2205 + }, + { + "epoch": 4.0851851851851855, + "grad_norm": 1.6176567857381396, + "learning_rate": 1.9728370022783286e-07, + "loss": 0.6588, + "step": 2206 + }, + { + "epoch": 4.087037037037037, + "grad_norm": 1.4503358353816915, + "learning_rate": 1.9728022810048022e-07, + "loss": 0.6808, + "step": 2207 + }, + { + "epoch": 4.088888888888889, + "grad_norm": 1.3604464627227404, + "learning_rate": 1.972767537860038e-07, + "loss": 0.7081, + "step": 2208 + }, + { + "epoch": 4.090740740740741, + "grad_norm": 1.4538034743653498, + "learning_rate": 1.9727327728448175e-07, + "loss": 0.6807, + "step": 2209 + }, + { + "epoch": 4.092592592592593, + "grad_norm": 1.4474918488198876, + "learning_rate": 1.9726979859599226e-07, + "loss": 0.696, + "step": 2210 + }, + { + "epoch": 4.094444444444444, + "grad_norm": 1.6549450936720331, + "learning_rate": 1.9726631772061352e-07, + "loss": 0.6814, + "step": 2211 + }, + { + "epoch": 4.0962962962962965, + "grad_norm": 1.3512251985750467, + "learning_rate": 1.9726283465842377e-07, + "loss": 0.7127, + "step": 2212 + }, + { + "epoch": 4.098148148148148, + "grad_norm": 1.327772691807005, + "learning_rate": 1.9725934940950136e-07, + "loss": 0.691, + "step": 2213 + }, + { + "epoch": 4.1, + "grad_norm": 1.7885523371542662, + "learning_rate": 1.972558619739246e-07, + "loss": 0.6471, + "step": 2214 + }, + { + "epoch": 4.101851851851852, + "grad_norm": 1.5540435558714403, + "learning_rate": 1.9725237235177193e-07, + "loss": 0.7084, + "step": 2215 + }, + { + "epoch": 4.103703703703704, + "grad_norm": 1.6317095980122032, + "learning_rate": 1.9724888054312177e-07, + "loss": 0.692, + "step": 2216 + }, + { + "epoch": 4.105555555555555, + "grad_norm": 1.5196913536493002, + "learning_rate": 1.9724538654805267e-07, + "loss": 0.6981, + "step": 2217 + }, + { + "epoch": 4.107407407407408, + "grad_norm": 1.390276381124955, + "learning_rate": 1.9724189036664317e-07, + "loss": 0.6412, + "step": 2218 + }, + { + "epoch": 4.109259259259259, + "grad_norm": 1.619629918950123, + "learning_rate": 1.972383919989719e-07, + "loss": 0.6927, + "step": 2219 + }, + { + "epoch": 4.111111111111111, + "grad_norm": 1.3732543235182741, + "learning_rate": 1.972348914451174e-07, + "loss": 0.6737, + "step": 2220 + }, + { + "epoch": 4.112962962962963, + "grad_norm": 1.380189554354143, + "learning_rate": 1.972313887051585e-07, + "loss": 0.6825, + "step": 2221 + }, + { + "epoch": 4.114814814814815, + "grad_norm": 1.490360712284871, + "learning_rate": 1.9722788377917393e-07, + "loss": 0.7127, + "step": 2222 + }, + { + "epoch": 4.116666666666666, + "grad_norm": 1.5537586492854871, + "learning_rate": 1.9722437666724243e-07, + "loss": 0.7089, + "step": 2223 + }, + { + "epoch": 4.118518518518519, + "grad_norm": 1.4089764960492903, + "learning_rate": 1.972208673694429e-07, + "loss": 0.7038, + "step": 2224 + }, + { + "epoch": 4.12037037037037, + "grad_norm": 1.6411428446735306, + "learning_rate": 1.9721735588585422e-07, + "loss": 0.6955, + "step": 2225 + }, + { + "epoch": 4.122222222222222, + "grad_norm": 1.4940716163272492, + "learning_rate": 1.9721384221655535e-07, + "loss": 0.6954, + "step": 2226 + }, + { + "epoch": 4.124074074074074, + "grad_norm": 2.131987017236616, + "learning_rate": 1.9721032636162528e-07, + "loss": 0.6622, + "step": 2227 + }, + { + "epoch": 4.125925925925926, + "grad_norm": 1.4513611427607525, + "learning_rate": 1.9720680832114303e-07, + "loss": 0.7035, + "step": 2228 + }, + { + "epoch": 4.127777777777778, + "grad_norm": 1.4567607267379143, + "learning_rate": 1.9720328809518777e-07, + "loss": 0.6919, + "step": 2229 + }, + { + "epoch": 4.12962962962963, + "grad_norm": 1.6848523295755233, + "learning_rate": 1.9719976568383854e-07, + "loss": 0.7328, + "step": 2230 + }, + { + "epoch": 4.131481481481481, + "grad_norm": 1.4987194252598426, + "learning_rate": 1.9719624108717458e-07, + "loss": 0.6528, + "step": 2231 + }, + { + "epoch": 4.133333333333334, + "grad_norm": 1.4437971921970107, + "learning_rate": 1.9719271430527518e-07, + "loss": 0.6745, + "step": 2232 + }, + { + "epoch": 4.135185185185185, + "grad_norm": 1.5568687632915494, + "learning_rate": 1.9718918533821958e-07, + "loss": 0.6705, + "step": 2233 + }, + { + "epoch": 4.137037037037037, + "grad_norm": 1.4253078383756228, + "learning_rate": 1.9718565418608712e-07, + "loss": 0.651, + "step": 2234 + }, + { + "epoch": 4.138888888888889, + "grad_norm": 1.3987986571385564, + "learning_rate": 1.9718212084895723e-07, + "loss": 0.616, + "step": 2235 + }, + { + "epoch": 4.140740740740741, + "grad_norm": 1.5825004287965694, + "learning_rate": 1.9717858532690927e-07, + "loss": 0.702, + "step": 2236 + }, + { + "epoch": 4.142592592592592, + "grad_norm": 2.9471120870232492, + "learning_rate": 1.9717504762002283e-07, + "loss": 0.7082, + "step": 2237 + }, + { + "epoch": 4.144444444444445, + "grad_norm": 1.4925028987116182, + "learning_rate": 1.9717150772837737e-07, + "loss": 0.656, + "step": 2238 + }, + { + "epoch": 4.146296296296296, + "grad_norm": 1.2629375992874476, + "learning_rate": 1.9716796565205252e-07, + "loss": 0.6757, + "step": 2239 + }, + { + "epoch": 4.148148148148148, + "grad_norm": 1.456162169204436, + "learning_rate": 1.971644213911279e-07, + "loss": 0.6712, + "step": 2240 + }, + { + "epoch": 4.15, + "grad_norm": 2.0523464747360287, + "learning_rate": 1.9716087494568314e-07, + "loss": 0.7397, + "step": 2241 + }, + { + "epoch": 4.151851851851852, + "grad_norm": 1.6850372818349062, + "learning_rate": 1.971573263157981e-07, + "loss": 0.7001, + "step": 2242 + }, + { + "epoch": 4.1537037037037035, + "grad_norm": 1.3856317262839872, + "learning_rate": 1.9715377550155243e-07, + "loss": 0.6192, + "step": 2243 + }, + { + "epoch": 4.155555555555556, + "grad_norm": 2.027055331601776, + "learning_rate": 1.9715022250302605e-07, + "loss": 0.6701, + "step": 2244 + }, + { + "epoch": 4.157407407407407, + "grad_norm": 1.4407227930368782, + "learning_rate": 1.971466673202988e-07, + "loss": 0.6737, + "step": 2245 + }, + { + "epoch": 4.159259259259259, + "grad_norm": 1.369143827499616, + "learning_rate": 1.9714310995345063e-07, + "loss": 0.7217, + "step": 2246 + }, + { + "epoch": 4.161111111111111, + "grad_norm": 1.5866031852302096, + "learning_rate": 1.9713955040256153e-07, + "loss": 0.6896, + "step": 2247 + }, + { + "epoch": 4.162962962962963, + "grad_norm": 2.3982661927193547, + "learning_rate": 1.9713598866771148e-07, + "loss": 0.6616, + "step": 2248 + }, + { + "epoch": 4.1648148148148145, + "grad_norm": 1.5310348420037856, + "learning_rate": 1.971324247489806e-07, + "loss": 0.6302, + "step": 2249 + }, + { + "epoch": 4.166666666666667, + "grad_norm": 1.3852294639169036, + "learning_rate": 1.9712885864644899e-07, + "loss": 0.7009, + "step": 2250 + }, + { + "epoch": 4.1685185185185185, + "grad_norm": 1.949168338344836, + "learning_rate": 1.9712529036019685e-07, + "loss": 0.6668, + "step": 2251 + }, + { + "epoch": 4.17037037037037, + "grad_norm": 1.403538975222234, + "learning_rate": 1.971217198903044e-07, + "loss": 0.6814, + "step": 2252 + }, + { + "epoch": 4.1722222222222225, + "grad_norm": 1.4063020007965452, + "learning_rate": 1.971181472368519e-07, + "loss": 0.6849, + "step": 2253 + }, + { + "epoch": 4.174074074074074, + "grad_norm": 1.502607814919187, + "learning_rate": 1.9711457239991968e-07, + "loss": 0.6965, + "step": 2254 + }, + { + "epoch": 4.175925925925926, + "grad_norm": 1.4549386071020158, + "learning_rate": 1.9711099537958816e-07, + "loss": 0.6769, + "step": 2255 + }, + { + "epoch": 4.177777777777778, + "grad_norm": 1.6408425489752825, + "learning_rate": 1.9710741617593766e-07, + "loss": 0.6872, + "step": 2256 + }, + { + "epoch": 4.17962962962963, + "grad_norm": 1.3878875237278199, + "learning_rate": 1.971038347890487e-07, + "loss": 0.7287, + "step": 2257 + }, + { + "epoch": 4.181481481481481, + "grad_norm": 1.4814270605062365, + "learning_rate": 1.9710025121900182e-07, + "loss": 0.6812, + "step": 2258 + }, + { + "epoch": 4.183333333333334, + "grad_norm": 1.6718393882272573, + "learning_rate": 1.9709666546587758e-07, + "loss": 0.703, + "step": 2259 + }, + { + "epoch": 4.185185185185185, + "grad_norm": 1.5794060442929199, + "learning_rate": 1.9709307752975658e-07, + "loss": 0.6779, + "step": 2260 + }, + { + "epoch": 4.187037037037037, + "grad_norm": 1.4340291266180283, + "learning_rate": 1.9708948741071952e-07, + "loss": 0.7084, + "step": 2261 + }, + { + "epoch": 4.188888888888889, + "grad_norm": 1.4022833261570278, + "learning_rate": 1.9708589510884708e-07, + "loss": 0.6374, + "step": 2262 + }, + { + "epoch": 4.190740740740741, + "grad_norm": 1.6988093889320932, + "learning_rate": 1.9708230062422002e-07, + "loss": 0.6267, + "step": 2263 + }, + { + "epoch": 4.192592592592592, + "grad_norm": 2.7673642160659644, + "learning_rate": 1.9707870395691916e-07, + "loss": 0.7112, + "step": 2264 + }, + { + "epoch": 4.194444444444445, + "grad_norm": 1.4308799160643817, + "learning_rate": 1.970751051070254e-07, + "loss": 0.7114, + "step": 2265 + }, + { + "epoch": 4.196296296296296, + "grad_norm": 1.479304820147132, + "learning_rate": 1.9707150407461964e-07, + "loss": 0.6767, + "step": 2266 + }, + { + "epoch": 4.198148148148148, + "grad_norm": 1.581126447377077, + "learning_rate": 1.970679008597828e-07, + "loss": 0.6829, + "step": 2267 + }, + { + "epoch": 4.2, + "grad_norm": 1.6758079381412818, + "learning_rate": 1.970642954625959e-07, + "loss": 0.7111, + "step": 2268 + }, + { + "epoch": 4.201851851851852, + "grad_norm": 1.7290307164664305, + "learning_rate": 1.9706068788314005e-07, + "loss": 0.6948, + "step": 2269 + }, + { + "epoch": 4.203703703703703, + "grad_norm": 2.5065337782890453, + "learning_rate": 1.970570781214963e-07, + "loss": 0.7169, + "step": 2270 + }, + { + "epoch": 4.205555555555556, + "grad_norm": 1.4876006766849879, + "learning_rate": 1.9705346617774582e-07, + "loss": 0.7442, + "step": 2271 + }, + { + "epoch": 4.207407407407407, + "grad_norm": 1.6868008505828358, + "learning_rate": 1.9704985205196987e-07, + "loss": 0.7037, + "step": 2272 + }, + { + "epoch": 4.209259259259259, + "grad_norm": 1.7897595947542868, + "learning_rate": 1.9704623574424964e-07, + "loss": 0.7647, + "step": 2273 + }, + { + "epoch": 4.211111111111111, + "grad_norm": 1.4062866460885806, + "learning_rate": 1.9704261725466648e-07, + "loss": 0.7046, + "step": 2274 + }, + { + "epoch": 4.212962962962963, + "grad_norm": 2.4129085814698557, + "learning_rate": 1.9703899658330168e-07, + "loss": 0.7162, + "step": 2275 + }, + { + "epoch": 4.214814814814815, + "grad_norm": 1.9701160169061898, + "learning_rate": 1.9703537373023668e-07, + "loss": 0.6995, + "step": 2276 + }, + { + "epoch": 4.216666666666667, + "grad_norm": 1.3513334182313512, + "learning_rate": 1.9703174869555296e-07, + "loss": 0.7123, + "step": 2277 + }, + { + "epoch": 4.218518518518518, + "grad_norm": 1.7334622199352077, + "learning_rate": 1.97028121479332e-07, + "loss": 0.6776, + "step": 2278 + }, + { + "epoch": 4.220370370370371, + "grad_norm": 2.191188979558819, + "learning_rate": 1.9702449208165532e-07, + "loss": 0.7168, + "step": 2279 + }, + { + "epoch": 4.222222222222222, + "grad_norm": 1.50460988072586, + "learning_rate": 1.9702086050260454e-07, + "loss": 0.7014, + "step": 2280 + }, + { + "epoch": 4.224074074074074, + "grad_norm": 1.6722983569725687, + "learning_rate": 1.9701722674226135e-07, + "loss": 0.6265, + "step": 2281 + }, + { + "epoch": 4.225925925925926, + "grad_norm": 1.5421617775623717, + "learning_rate": 1.9701359080070738e-07, + "loss": 0.6733, + "step": 2282 + }, + { + "epoch": 4.227777777777778, + "grad_norm": 1.3824095268494505, + "learning_rate": 1.9700995267802444e-07, + "loss": 0.68, + "step": 2283 + }, + { + "epoch": 4.229629629629629, + "grad_norm": 1.415004352752854, + "learning_rate": 1.9700631237429426e-07, + "loss": 0.7003, + "step": 2284 + }, + { + "epoch": 4.231481481481482, + "grad_norm": 1.56431904802866, + "learning_rate": 1.970026698895987e-07, + "loss": 0.6722, + "step": 2285 + }, + { + "epoch": 4.233333333333333, + "grad_norm": 1.5869417264614845, + "learning_rate": 1.969990252240197e-07, + "loss": 0.6537, + "step": 2286 + }, + { + "epoch": 4.235185185185185, + "grad_norm": 1.4883632494014178, + "learning_rate": 1.9699537837763912e-07, + "loss": 0.6926, + "step": 2287 + }, + { + "epoch": 4.237037037037037, + "grad_norm": 1.4891195908400743, + "learning_rate": 1.9699172935053902e-07, + "loss": 0.6835, + "step": 2288 + }, + { + "epoch": 4.238888888888889, + "grad_norm": 1.3910281952779788, + "learning_rate": 1.9698807814280145e-07, + "loss": 0.6574, + "step": 2289 + }, + { + "epoch": 4.2407407407407405, + "grad_norm": 1.6819731681986096, + "learning_rate": 1.9698442475450843e-07, + "loss": 0.6688, + "step": 2290 + }, + { + "epoch": 4.242592592592593, + "grad_norm": 1.5872465106139797, + "learning_rate": 1.9698076918574214e-07, + "loss": 0.6432, + "step": 2291 + }, + { + "epoch": 4.2444444444444445, + "grad_norm": 1.4686013989363231, + "learning_rate": 1.9697711143658477e-07, + "loss": 0.7397, + "step": 2292 + }, + { + "epoch": 4.246296296296296, + "grad_norm": 2.4159677052324504, + "learning_rate": 1.9697345150711854e-07, + "loss": 0.6663, + "step": 2293 + }, + { + "epoch": 4.2481481481481485, + "grad_norm": 4.980204864651794, + "learning_rate": 1.9696978939742577e-07, + "loss": 0.675, + "step": 2294 + }, + { + "epoch": 4.25, + "grad_norm": 1.4060698365820736, + "learning_rate": 1.9696612510758874e-07, + "loss": 0.7257, + "step": 2295 + }, + { + "epoch": 4.2518518518518515, + "grad_norm": 1.4089277172815022, + "learning_rate": 1.9696245863768987e-07, + "loss": 0.6574, + "step": 2296 + }, + { + "epoch": 4.253703703703704, + "grad_norm": 1.6803499585189763, + "learning_rate": 1.9695878998781158e-07, + "loss": 0.6952, + "step": 2297 + }, + { + "epoch": 4.2555555555555555, + "grad_norm": 1.5430778345737401, + "learning_rate": 1.9695511915803638e-07, + "loss": 0.6854, + "step": 2298 + }, + { + "epoch": 4.257407407407407, + "grad_norm": 1.5882620877329388, + "learning_rate": 1.9695144614844676e-07, + "loss": 0.6717, + "step": 2299 + }, + { + "epoch": 4.2592592592592595, + "grad_norm": 1.4311850445392638, + "learning_rate": 1.9694777095912533e-07, + "loss": 0.6974, + "step": 2300 + }, + { + "epoch": 4.261111111111111, + "grad_norm": 1.3753003535916093, + "learning_rate": 1.9694409359015465e-07, + "loss": 0.6978, + "step": 2301 + }, + { + "epoch": 4.262962962962963, + "grad_norm": 1.6863316782220699, + "learning_rate": 1.969404140416175e-07, + "loss": 0.727, + "step": 2302 + }, + { + "epoch": 4.264814814814815, + "grad_norm": 1.456873810713547, + "learning_rate": 1.9693673231359652e-07, + "loss": 0.6628, + "step": 2303 + }, + { + "epoch": 4.266666666666667, + "grad_norm": 1.6201139473330586, + "learning_rate": 1.9693304840617456e-07, + "loss": 0.6928, + "step": 2304 + }, + { + "epoch": 4.268518518518518, + "grad_norm": 1.4839274081238842, + "learning_rate": 1.9692936231943438e-07, + "loss": 0.721, + "step": 2305 + }, + { + "epoch": 4.270370370370371, + "grad_norm": 1.3895999291296606, + "learning_rate": 1.9692567405345888e-07, + "loss": 0.7082, + "step": 2306 + }, + { + "epoch": 4.272222222222222, + "grad_norm": 1.3761618374147762, + "learning_rate": 1.9692198360833098e-07, + "loss": 0.6738, + "step": 2307 + }, + { + "epoch": 4.274074074074074, + "grad_norm": 1.692909362625955, + "learning_rate": 1.9691829098413368e-07, + "loss": 0.7318, + "step": 2308 + }, + { + "epoch": 4.275925925925926, + "grad_norm": 2.2654075167340655, + "learning_rate": 1.9691459618094996e-07, + "loss": 0.7193, + "step": 2309 + }, + { + "epoch": 4.277777777777778, + "grad_norm": 1.4717886113152736, + "learning_rate": 1.969108991988629e-07, + "loss": 0.7267, + "step": 2310 + }, + { + "epoch": 4.279629629629629, + "grad_norm": 1.8551498631186467, + "learning_rate": 1.969072000379556e-07, + "loss": 0.6772, + "step": 2311 + }, + { + "epoch": 4.281481481481482, + "grad_norm": 1.31833327253509, + "learning_rate": 1.9690349869831125e-07, + "loss": 0.7001, + "step": 2312 + }, + { + "epoch": 4.283333333333333, + "grad_norm": 1.4921738146631278, + "learning_rate": 1.968997951800131e-07, + "loss": 0.6843, + "step": 2313 + }, + { + "epoch": 4.285185185185185, + "grad_norm": 1.4869023237836838, + "learning_rate": 1.9689608948314433e-07, + "loss": 0.7066, + "step": 2314 + }, + { + "epoch": 4.287037037037037, + "grad_norm": 1.7499097730414759, + "learning_rate": 1.9689238160778833e-07, + "loss": 0.6995, + "step": 2315 + }, + { + "epoch": 4.288888888888889, + "grad_norm": 1.5476176483612694, + "learning_rate": 1.9688867155402843e-07, + "loss": 0.7139, + "step": 2316 + }, + { + "epoch": 4.290740740740741, + "grad_norm": 1.6543335984688297, + "learning_rate": 1.9688495932194805e-07, + "loss": 0.6553, + "step": 2317 + }, + { + "epoch": 4.292592592592593, + "grad_norm": 3.4176797230006306, + "learning_rate": 1.9688124491163066e-07, + "loss": 0.6984, + "step": 2318 + }, + { + "epoch": 4.294444444444444, + "grad_norm": 1.4890040045378254, + "learning_rate": 1.9687752832315973e-07, + "loss": 0.7351, + "step": 2319 + }, + { + "epoch": 4.296296296296296, + "grad_norm": 1.4150723142801551, + "learning_rate": 1.9687380955661888e-07, + "loss": 0.6748, + "step": 2320 + }, + { + "epoch": 4.298148148148148, + "grad_norm": 1.3442858635744965, + "learning_rate": 1.9687008861209167e-07, + "loss": 0.7559, + "step": 2321 + }, + { + "epoch": 4.3, + "grad_norm": 1.5278787688267936, + "learning_rate": 1.9686636548966176e-07, + "loss": 0.7076, + "step": 2322 + }, + { + "epoch": 4.301851851851852, + "grad_norm": 1.3617132550771704, + "learning_rate": 1.968626401894129e-07, + "loss": 0.626, + "step": 2323 + }, + { + "epoch": 4.303703703703704, + "grad_norm": 3.2242961736469318, + "learning_rate": 1.968589127114288e-07, + "loss": 0.6766, + "step": 2324 + }, + { + "epoch": 4.305555555555555, + "grad_norm": 1.7128904080471745, + "learning_rate": 1.9685518305579327e-07, + "loss": 0.668, + "step": 2325 + }, + { + "epoch": 4.307407407407408, + "grad_norm": 1.6107532461968255, + "learning_rate": 1.9685145122259016e-07, + "loss": 0.6854, + "step": 2326 + }, + { + "epoch": 4.309259259259259, + "grad_norm": 1.5890173961707887, + "learning_rate": 1.9684771721190342e-07, + "loss": 0.6377, + "step": 2327 + }, + { + "epoch": 4.311111111111111, + "grad_norm": 1.323674044479915, + "learning_rate": 1.9684398102381692e-07, + "loss": 0.7093, + "step": 2328 + }, + { + "epoch": 4.312962962962963, + "grad_norm": 3.675952489110242, + "learning_rate": 1.9684024265841474e-07, + "loss": 0.6596, + "step": 2329 + }, + { + "epoch": 4.314814814814815, + "grad_norm": 2.0034725577657424, + "learning_rate": 1.9683650211578088e-07, + "loss": 0.6319, + "step": 2330 + }, + { + "epoch": 4.316666666666666, + "grad_norm": 1.5845396526248812, + "learning_rate": 1.9683275939599942e-07, + "loss": 0.7091, + "step": 2331 + }, + { + "epoch": 4.318518518518519, + "grad_norm": 1.4877926277331357, + "learning_rate": 1.9682901449915455e-07, + "loss": 0.6735, + "step": 2332 + }, + { + "epoch": 4.32037037037037, + "grad_norm": 1.4784240311293557, + "learning_rate": 1.9682526742533045e-07, + "loss": 0.6897, + "step": 2333 + }, + { + "epoch": 4.322222222222222, + "grad_norm": 1.3816827709237887, + "learning_rate": 1.9682151817461138e-07, + "loss": 0.6699, + "step": 2334 + }, + { + "epoch": 4.324074074074074, + "grad_norm": 1.5334864373471044, + "learning_rate": 1.968177667470816e-07, + "loss": 0.6998, + "step": 2335 + }, + { + "epoch": 4.325925925925926, + "grad_norm": 1.5550643485124138, + "learning_rate": 1.9681401314282548e-07, + "loss": 0.7168, + "step": 2336 + }, + { + "epoch": 4.3277777777777775, + "grad_norm": 1.706395438568234, + "learning_rate": 1.9681025736192742e-07, + "loss": 0.6919, + "step": 2337 + }, + { + "epoch": 4.32962962962963, + "grad_norm": 1.5992700670608937, + "learning_rate": 1.9680649940447182e-07, + "loss": 0.6784, + "step": 2338 + }, + { + "epoch": 4.3314814814814815, + "grad_norm": 1.6981882215391304, + "learning_rate": 1.968027392705432e-07, + "loss": 0.6916, + "step": 2339 + }, + { + "epoch": 4.333333333333333, + "grad_norm": 1.4934205886203187, + "learning_rate": 1.9679897696022607e-07, + "loss": 0.7427, + "step": 2340 + }, + { + "epoch": 4.3351851851851855, + "grad_norm": 1.4002997648779245, + "learning_rate": 1.9679521247360504e-07, + "loss": 0.6595, + "step": 2341 + }, + { + "epoch": 4.337037037037037, + "grad_norm": 2.047164484611014, + "learning_rate": 1.9679144581076477e-07, + "loss": 0.6527, + "step": 2342 + }, + { + "epoch": 4.338888888888889, + "grad_norm": 1.4305475391234135, + "learning_rate": 1.9678767697178988e-07, + "loss": 0.6839, + "step": 2343 + }, + { + "epoch": 4.340740740740741, + "grad_norm": 1.5038768583705346, + "learning_rate": 1.9678390595676516e-07, + "loss": 0.6979, + "step": 2344 + }, + { + "epoch": 4.342592592592593, + "grad_norm": 1.8953918709655337, + "learning_rate": 1.9678013276577538e-07, + "loss": 0.6268, + "step": 2345 + }, + { + "epoch": 4.344444444444444, + "grad_norm": 1.890318725555, + "learning_rate": 1.9677635739890534e-07, + "loss": 0.6675, + "step": 2346 + }, + { + "epoch": 4.3462962962962965, + "grad_norm": 2.031671680129892, + "learning_rate": 1.9677257985623995e-07, + "loss": 0.7113, + "step": 2347 + }, + { + "epoch": 4.348148148148148, + "grad_norm": 1.599502066015545, + "learning_rate": 1.9676880013786416e-07, + "loss": 0.6859, + "step": 2348 + }, + { + "epoch": 4.35, + "grad_norm": 1.5070218514048113, + "learning_rate": 1.9676501824386292e-07, + "loss": 0.6969, + "step": 2349 + }, + { + "epoch": 4.351851851851852, + "grad_norm": 1.4977923178284103, + "learning_rate": 1.9676123417432123e-07, + "loss": 0.6923, + "step": 2350 + }, + { + "epoch": 4.353703703703704, + "grad_norm": 1.3998476686446415, + "learning_rate": 1.9675744792932424e-07, + "loss": 0.7088, + "step": 2351 + }, + { + "epoch": 4.355555555555555, + "grad_norm": 1.6322499275784297, + "learning_rate": 1.96753659508957e-07, + "loss": 0.6873, + "step": 2352 + }, + { + "epoch": 4.357407407407408, + "grad_norm": 1.5679297419678981, + "learning_rate": 1.967498689133047e-07, + "loss": 0.7184, + "step": 2353 + }, + { + "epoch": 4.359259259259259, + "grad_norm": 1.686332798631044, + "learning_rate": 1.967460761424526e-07, + "loss": 0.6904, + "step": 2354 + }, + { + "epoch": 4.361111111111111, + "grad_norm": 1.490096183947135, + "learning_rate": 1.9674228119648596e-07, + "loss": 0.6771, + "step": 2355 + }, + { + "epoch": 4.362962962962963, + "grad_norm": 1.558047028061556, + "learning_rate": 1.9673848407549007e-07, + "loss": 0.6961, + "step": 2356 + }, + { + "epoch": 4.364814814814815, + "grad_norm": 1.7441474155082488, + "learning_rate": 1.9673468477955032e-07, + "loss": 0.6923, + "step": 2357 + }, + { + "epoch": 4.366666666666666, + "grad_norm": 1.7117247371163942, + "learning_rate": 1.9673088330875215e-07, + "loss": 0.6831, + "step": 2358 + }, + { + "epoch": 4.368518518518519, + "grad_norm": 1.5542845877712854, + "learning_rate": 1.9672707966318099e-07, + "loss": 0.6569, + "step": 2359 + }, + { + "epoch": 4.37037037037037, + "grad_norm": 1.5844531527933063, + "learning_rate": 1.9672327384292237e-07, + "loss": 0.6756, + "step": 2360 + }, + { + "epoch": 4.372222222222222, + "grad_norm": 2.1076030936767762, + "learning_rate": 1.9671946584806188e-07, + "loss": 0.6679, + "step": 2361 + }, + { + "epoch": 4.374074074074074, + "grad_norm": 1.3780602984216406, + "learning_rate": 1.967156556786851e-07, + "loss": 0.6928, + "step": 2362 + }, + { + "epoch": 4.375925925925926, + "grad_norm": 1.7865773329558863, + "learning_rate": 1.9671184333487769e-07, + "loss": 0.7067, + "step": 2363 + }, + { + "epoch": 4.377777777777778, + "grad_norm": 1.6991029406573863, + "learning_rate": 1.9670802881672538e-07, + "loss": 0.6782, + "step": 2364 + }, + { + "epoch": 4.37962962962963, + "grad_norm": 1.7640894888355967, + "learning_rate": 1.9670421212431396e-07, + "loss": 0.6872, + "step": 2365 + }, + { + "epoch": 4.381481481481481, + "grad_norm": 1.6820068347164712, + "learning_rate": 1.9670039325772918e-07, + "loss": 0.6815, + "step": 2366 + }, + { + "epoch": 4.383333333333334, + "grad_norm": 1.50621941463131, + "learning_rate": 1.9669657221705692e-07, + "loss": 0.6503, + "step": 2367 + }, + { + "epoch": 4.385185185185185, + "grad_norm": 1.5467105191025625, + "learning_rate": 1.966927490023831e-07, + "loss": 0.6681, + "step": 2368 + }, + { + "epoch": 4.387037037037037, + "grad_norm": 2.07453642348771, + "learning_rate": 1.9668892361379364e-07, + "loss": 0.696, + "step": 2369 + }, + { + "epoch": 4.388888888888889, + "grad_norm": 1.4732809889821752, + "learning_rate": 1.9668509605137463e-07, + "loss": 0.6973, + "step": 2370 + }, + { + "epoch": 4.390740740740741, + "grad_norm": 1.3780632992932822, + "learning_rate": 1.9668126631521203e-07, + "loss": 0.7168, + "step": 2371 + }, + { + "epoch": 4.392592592592592, + "grad_norm": 3.092274387794323, + "learning_rate": 1.96677434405392e-07, + "loss": 0.749, + "step": 2372 + }, + { + "epoch": 4.394444444444445, + "grad_norm": 1.5498985181523233, + "learning_rate": 1.9667360032200064e-07, + "loss": 0.6788, + "step": 2373 + }, + { + "epoch": 4.396296296296296, + "grad_norm": 1.5431896323403844, + "learning_rate": 1.966697640651242e-07, + "loss": 0.6823, + "step": 2374 + }, + { + "epoch": 4.398148148148148, + "grad_norm": 1.3448848916045388, + "learning_rate": 1.9666592563484893e-07, + "loss": 0.6743, + "step": 2375 + }, + { + "epoch": 4.4, + "grad_norm": 1.6670249929742256, + "learning_rate": 1.966620850312611e-07, + "loss": 0.695, + "step": 2376 + }, + { + "epoch": 4.401851851851852, + "grad_norm": 1.4748960946861083, + "learning_rate": 1.9665824225444707e-07, + "loss": 0.7143, + "step": 2377 + }, + { + "epoch": 4.4037037037037035, + "grad_norm": 1.4431618700190834, + "learning_rate": 1.9665439730449325e-07, + "loss": 0.7015, + "step": 2378 + }, + { + "epoch": 4.405555555555556, + "grad_norm": 1.5125610840472394, + "learning_rate": 1.9665055018148603e-07, + "loss": 0.6946, + "step": 2379 + }, + { + "epoch": 4.407407407407407, + "grad_norm": 1.4962978890567262, + "learning_rate": 1.9664670088551198e-07, + "loss": 0.6597, + "step": 2380 + }, + { + "epoch": 4.409259259259259, + "grad_norm": 1.66247046530707, + "learning_rate": 1.966428494166576e-07, + "loss": 0.7001, + "step": 2381 + }, + { + "epoch": 4.411111111111111, + "grad_norm": 1.6837769840799373, + "learning_rate": 1.9663899577500947e-07, + "loss": 0.635, + "step": 2382 + }, + { + "epoch": 4.412962962962963, + "grad_norm": 1.51486886127756, + "learning_rate": 1.9663513996065425e-07, + "loss": 0.7089, + "step": 2383 + }, + { + "epoch": 4.4148148148148145, + "grad_norm": 1.3122943212886047, + "learning_rate": 1.9663128197367864e-07, + "loss": 0.6537, + "step": 2384 + }, + { + "epoch": 4.416666666666667, + "grad_norm": 2.6176071359062556, + "learning_rate": 1.9662742181416936e-07, + "loss": 0.7263, + "step": 2385 + }, + { + "epoch": 4.4185185185185185, + "grad_norm": 1.9221524390865365, + "learning_rate": 1.966235594822132e-07, + "loss": 0.6741, + "step": 2386 + }, + { + "epoch": 4.42037037037037, + "grad_norm": 1.554543842041304, + "learning_rate": 1.96619694977897e-07, + "loss": 0.7177, + "step": 2387 + }, + { + "epoch": 4.4222222222222225, + "grad_norm": 1.520867093901909, + "learning_rate": 1.9661582830130762e-07, + "loss": 0.6715, + "step": 2388 + }, + { + "epoch": 4.424074074074074, + "grad_norm": 1.3633850987431106, + "learning_rate": 1.9661195945253203e-07, + "loss": 0.6953, + "step": 2389 + }, + { + "epoch": 4.425925925925926, + "grad_norm": 1.621114779937641, + "learning_rate": 1.966080884316572e-07, + "loss": 0.668, + "step": 2390 + }, + { + "epoch": 4.427777777777778, + "grad_norm": 1.6283779267691725, + "learning_rate": 1.966042152387702e-07, + "loss": 0.664, + "step": 2391 + }, + { + "epoch": 4.42962962962963, + "grad_norm": 1.7732568278325749, + "learning_rate": 1.9660033987395802e-07, + "loss": 0.7128, + "step": 2392 + }, + { + "epoch": 4.431481481481481, + "grad_norm": 2.005239505823147, + "learning_rate": 1.9659646233730782e-07, + "loss": 0.6819, + "step": 2393 + }, + { + "epoch": 4.433333333333334, + "grad_norm": 1.4806031086417617, + "learning_rate": 1.965925826289068e-07, + "loss": 0.7168, + "step": 2394 + }, + { + "epoch": 4.435185185185185, + "grad_norm": 2.7011701540418165, + "learning_rate": 1.9658870074884223e-07, + "loss": 0.7213, + "step": 2395 + }, + { + "epoch": 4.437037037037037, + "grad_norm": 1.6615366133170555, + "learning_rate": 1.965848166972013e-07, + "loss": 0.6156, + "step": 2396 + }, + { + "epoch": 4.438888888888889, + "grad_norm": 1.3741393099098673, + "learning_rate": 1.9658093047407136e-07, + "loss": 0.7057, + "step": 2397 + }, + { + "epoch": 4.440740740740741, + "grad_norm": 1.5457289951528703, + "learning_rate": 1.9657704207953983e-07, + "loss": 0.6635, + "step": 2398 + }, + { + "epoch": 4.442592592592592, + "grad_norm": 1.781929065650584, + "learning_rate": 1.9657315151369408e-07, + "loss": 0.6263, + "step": 2399 + }, + { + "epoch": 4.444444444444445, + "grad_norm": 1.6173288177732108, + "learning_rate": 1.965692587766216e-07, + "loss": 0.6296, + "step": 2400 + }, + { + "epoch": 4.446296296296296, + "grad_norm": 1.6761171449830787, + "learning_rate": 1.9656536386840987e-07, + "loss": 0.7233, + "step": 2401 + }, + { + "epoch": 4.448148148148148, + "grad_norm": 1.3702406511156924, + "learning_rate": 1.965614667891465e-07, + "loss": 0.6626, + "step": 2402 + }, + { + "epoch": 4.45, + "grad_norm": 1.6119917256295062, + "learning_rate": 1.9655756753891915e-07, + "loss": 0.703, + "step": 2403 + }, + { + "epoch": 4.451851851851852, + "grad_norm": 1.5402894568752385, + "learning_rate": 1.9655366611781538e-07, + "loss": 0.691, + "step": 2404 + }, + { + "epoch": 4.453703703703704, + "grad_norm": 1.54251161451274, + "learning_rate": 1.96549762525923e-07, + "loss": 0.7081, + "step": 2405 + }, + { + "epoch": 4.455555555555556, + "grad_norm": 1.5856002467827401, + "learning_rate": 1.965458567633297e-07, + "loss": 0.7057, + "step": 2406 + }, + { + "epoch": 4.457407407407407, + "grad_norm": 1.5166706394239964, + "learning_rate": 1.9654194883012335e-07, + "loss": 0.639, + "step": 2407 + }, + { + "epoch": 4.459259259259259, + "grad_norm": 1.4020674626160166, + "learning_rate": 1.9653803872639175e-07, + "loss": 0.6922, + "step": 2408 + }, + { + "epoch": 4.461111111111111, + "grad_norm": 1.6831364244895552, + "learning_rate": 1.9653412645222288e-07, + "loss": 0.707, + "step": 2409 + }, + { + "epoch": 4.462962962962963, + "grad_norm": 1.4110691733242287, + "learning_rate": 1.9653021200770467e-07, + "loss": 0.6761, + "step": 2410 + }, + { + "epoch": 4.464814814814815, + "grad_norm": 1.7237659498575442, + "learning_rate": 1.9652629539292512e-07, + "loss": 0.6972, + "step": 2411 + }, + { + "epoch": 4.466666666666667, + "grad_norm": 1.4932202921523328, + "learning_rate": 1.9652237660797225e-07, + "loss": 0.6749, + "step": 2412 + }, + { + "epoch": 4.468518518518518, + "grad_norm": 1.6116619910054157, + "learning_rate": 1.9651845565293423e-07, + "loss": 0.6347, + "step": 2413 + }, + { + "epoch": 4.470370370370371, + "grad_norm": 1.6772409168750315, + "learning_rate": 1.965145325278992e-07, + "loss": 0.7371, + "step": 2414 + }, + { + "epoch": 4.472222222222222, + "grad_norm": 1.5878064543705943, + "learning_rate": 1.9651060723295535e-07, + "loss": 0.6431, + "step": 2415 + }, + { + "epoch": 4.474074074074074, + "grad_norm": 1.3830616204809014, + "learning_rate": 1.965066797681909e-07, + "loss": 0.6777, + "step": 2416 + }, + { + "epoch": 4.475925925925926, + "grad_norm": 1.7002639460923434, + "learning_rate": 1.9650275013369421e-07, + "loss": 0.7034, + "step": 2417 + }, + { + "epoch": 4.477777777777778, + "grad_norm": 2.038646674254075, + "learning_rate": 1.964988183295536e-07, + "loss": 0.6578, + "step": 2418 + }, + { + "epoch": 4.479629629629629, + "grad_norm": 1.7307788126039625, + "learning_rate": 1.9649488435585743e-07, + "loss": 0.6944, + "step": 2419 + }, + { + "epoch": 4.481481481481482, + "grad_norm": 1.5922951431882066, + "learning_rate": 1.9649094821269424e-07, + "loss": 0.6489, + "step": 2420 + }, + { + "epoch": 4.483333333333333, + "grad_norm": 1.3216260980940011, + "learning_rate": 1.964870099001524e-07, + "loss": 0.6504, + "step": 2421 + }, + { + "epoch": 4.485185185185185, + "grad_norm": 1.4984329468265338, + "learning_rate": 1.9648306941832055e-07, + "loss": 0.6495, + "step": 2422 + }, + { + "epoch": 4.487037037037037, + "grad_norm": 1.5446642960928283, + "learning_rate": 1.9647912676728729e-07, + "loss": 0.7122, + "step": 2423 + }, + { + "epoch": 4.488888888888889, + "grad_norm": 2.543488707428639, + "learning_rate": 1.9647518194714122e-07, + "loss": 0.719, + "step": 2424 + }, + { + "epoch": 4.4907407407407405, + "grad_norm": 2.060011821320633, + "learning_rate": 1.96471234957971e-07, + "loss": 0.6642, + "step": 2425 + }, + { + "epoch": 4.492592592592593, + "grad_norm": 1.587620858457138, + "learning_rate": 1.9646728579986542e-07, + "loss": 0.6551, + "step": 2426 + }, + { + "epoch": 4.4944444444444445, + "grad_norm": 1.6116222103706, + "learning_rate": 1.964633344729133e-07, + "loss": 0.6835, + "step": 2427 + }, + { + "epoch": 4.496296296296296, + "grad_norm": 1.4478623102876287, + "learning_rate": 1.9645938097720336e-07, + "loss": 0.6453, + "step": 2428 + }, + { + "epoch": 4.4981481481481485, + "grad_norm": 1.6068415350856775, + "learning_rate": 1.964554253128246e-07, + "loss": 0.7145, + "step": 2429 + }, + { + "epoch": 4.5, + "grad_norm": 1.4256365709305527, + "learning_rate": 1.9645146747986589e-07, + "loss": 0.6522, + "step": 2430 + }, + { + "epoch": 4.5018518518518515, + "grad_norm": 1.5353351608677912, + "learning_rate": 1.9644750747841622e-07, + "loss": 0.7083, + "step": 2431 + }, + { + "epoch": 4.503703703703704, + "grad_norm": 1.7544991379226202, + "learning_rate": 1.9644354530856465e-07, + "loss": 0.7151, + "step": 2432 + }, + { + "epoch": 4.5055555555555555, + "grad_norm": 1.5087673143829372, + "learning_rate": 1.9643958097040024e-07, + "loss": 0.6671, + "step": 2433 + }, + { + "epoch": 4.507407407407407, + "grad_norm": 1.356781894671453, + "learning_rate": 1.9643561446401217e-07, + "loss": 0.684, + "step": 2434 + }, + { + "epoch": 4.5092592592592595, + "grad_norm": 1.5841396202681788, + "learning_rate": 1.964316457894895e-07, + "loss": 0.6815, + "step": 2435 + }, + { + "epoch": 4.511111111111111, + "grad_norm": 2.159185886689002, + "learning_rate": 1.9642767494692158e-07, + "loss": 0.7013, + "step": 2436 + }, + { + "epoch": 4.512962962962963, + "grad_norm": 1.6160052270632668, + "learning_rate": 1.964237019363976e-07, + "loss": 0.6552, + "step": 2437 + }, + { + "epoch": 4.514814814814815, + "grad_norm": 1.4558896411201991, + "learning_rate": 1.9641972675800695e-07, + "loss": 0.6823, + "step": 2438 + }, + { + "epoch": 4.516666666666667, + "grad_norm": 1.695200750044827, + "learning_rate": 1.9641574941183893e-07, + "loss": 0.7192, + "step": 2439 + }, + { + "epoch": 4.518518518518518, + "grad_norm": 1.7358590588009553, + "learning_rate": 1.9641176989798302e-07, + "loss": 0.7085, + "step": 2440 + }, + { + "epoch": 4.520370370370371, + "grad_norm": 1.5119955027098257, + "learning_rate": 1.964077882165287e-07, + "loss": 0.6718, + "step": 2441 + }, + { + "epoch": 4.522222222222222, + "grad_norm": 1.6872179603981665, + "learning_rate": 1.9640380436756544e-07, + "loss": 0.6804, + "step": 2442 + }, + { + "epoch": 4.524074074074074, + "grad_norm": 1.7023356501940974, + "learning_rate": 1.963998183511828e-07, + "loss": 0.7048, + "step": 2443 + }, + { + "epoch": 4.525925925925926, + "grad_norm": 1.864878210703174, + "learning_rate": 1.9639583016747048e-07, + "loss": 0.7453, + "step": 2444 + }, + { + "epoch": 4.527777777777778, + "grad_norm": 1.7407177542937717, + "learning_rate": 1.9639183981651808e-07, + "loss": 0.6842, + "step": 2445 + }, + { + "epoch": 4.52962962962963, + "grad_norm": 1.7883034287823847, + "learning_rate": 1.963878472984153e-07, + "loss": 0.6785, + "step": 2446 + }, + { + "epoch": 4.531481481481482, + "grad_norm": 1.5591214579620831, + "learning_rate": 1.9638385261325192e-07, + "loss": 0.6729, + "step": 2447 + }, + { + "epoch": 4.533333333333333, + "grad_norm": 1.6080240317581804, + "learning_rate": 1.9637985576111778e-07, + "loss": 0.685, + "step": 2448 + }, + { + "epoch": 4.535185185185185, + "grad_norm": 1.9957364959103105, + "learning_rate": 1.9637585674210274e-07, + "loss": 0.661, + "step": 2449 + }, + { + "epoch": 4.537037037037037, + "grad_norm": 1.5215072428715826, + "learning_rate": 1.9637185555629663e-07, + "loss": 0.6713, + "step": 2450 + }, + { + "epoch": 4.538888888888889, + "grad_norm": 1.643521573232913, + "learning_rate": 1.963678522037895e-07, + "loss": 0.6935, + "step": 2451 + }, + { + "epoch": 4.540740740740741, + "grad_norm": 1.4458233435771315, + "learning_rate": 1.963638466846713e-07, + "loss": 0.7033, + "step": 2452 + }, + { + "epoch": 4.542592592592593, + "grad_norm": 1.4529875669676566, + "learning_rate": 1.9635983899903215e-07, + "loss": 0.7, + "step": 2453 + }, + { + "epoch": 4.544444444444444, + "grad_norm": 1.535621846755058, + "learning_rate": 1.963558291469621e-07, + "loss": 0.7377, + "step": 2454 + }, + { + "epoch": 4.546296296296296, + "grad_norm": 1.3917112409585297, + "learning_rate": 1.963518171285513e-07, + "loss": 0.6688, + "step": 2455 + }, + { + "epoch": 4.548148148148148, + "grad_norm": 2.145977150141421, + "learning_rate": 1.9634780294388995e-07, + "loss": 0.6811, + "step": 2456 + }, + { + "epoch": 4.55, + "grad_norm": 1.634622934844824, + "learning_rate": 1.9634378659306832e-07, + "loss": 0.6966, + "step": 2457 + }, + { + "epoch": 4.551851851851852, + "grad_norm": 1.4789292012956052, + "learning_rate": 1.963397680761767e-07, + "loss": 0.6946, + "step": 2458 + }, + { + "epoch": 4.553703703703704, + "grad_norm": 1.4459577812881277, + "learning_rate": 1.9633574739330546e-07, + "loss": 0.671, + "step": 2459 + }, + { + "epoch": 4.555555555555555, + "grad_norm": 1.3421470003354312, + "learning_rate": 1.9633172454454495e-07, + "loss": 0.7119, + "step": 2460 + }, + { + "epoch": 4.557407407407408, + "grad_norm": 1.4777860469181126, + "learning_rate": 1.9632769952998566e-07, + "loss": 0.7256, + "step": 2461 + }, + { + "epoch": 4.559259259259259, + "grad_norm": 1.3293392775929562, + "learning_rate": 1.9632367234971802e-07, + "loss": 0.7326, + "step": 2462 + }, + { + "epoch": 4.561111111111111, + "grad_norm": 1.7657205146975035, + "learning_rate": 1.9631964300383267e-07, + "loss": 0.7225, + "step": 2463 + }, + { + "epoch": 4.562962962962963, + "grad_norm": 1.6070353156946582, + "learning_rate": 1.963156114924201e-07, + "loss": 0.6556, + "step": 2464 + }, + { + "epoch": 4.564814814814815, + "grad_norm": 1.8722185653369976, + "learning_rate": 1.96311577815571e-07, + "loss": 0.6969, + "step": 2465 + }, + { + "epoch": 4.566666666666666, + "grad_norm": 1.5852800887906722, + "learning_rate": 1.9630754197337608e-07, + "loss": 0.7098, + "step": 2466 + }, + { + "epoch": 4.568518518518519, + "grad_norm": 2.5997157514137337, + "learning_rate": 1.9630350396592604e-07, + "loss": 0.6959, + "step": 2467 + }, + { + "epoch": 4.57037037037037, + "grad_norm": 1.4523933697811655, + "learning_rate": 1.9629946379331167e-07, + "loss": 0.6616, + "step": 2468 + }, + { + "epoch": 4.572222222222222, + "grad_norm": 1.5856203419210637, + "learning_rate": 1.962954214556238e-07, + "loss": 0.6888, + "step": 2469 + }, + { + "epoch": 4.574074074074074, + "grad_norm": 1.610935755469287, + "learning_rate": 1.9629137695295334e-07, + "loss": 0.7206, + "step": 2470 + }, + { + "epoch": 4.575925925925926, + "grad_norm": 1.6159209436172026, + "learning_rate": 1.962873302853912e-07, + "loss": 0.6918, + "step": 2471 + }, + { + "epoch": 4.5777777777777775, + "grad_norm": 1.4612138613264, + "learning_rate": 1.9628328145302834e-07, + "loss": 0.6529, + "step": 2472 + }, + { + "epoch": 4.57962962962963, + "grad_norm": 1.6277979455846525, + "learning_rate": 1.9627923045595582e-07, + "loss": 0.654, + "step": 2473 + }, + { + "epoch": 4.5814814814814815, + "grad_norm": 1.4594466889751476, + "learning_rate": 1.9627517729426476e-07, + "loss": 0.6605, + "step": 2474 + }, + { + "epoch": 4.583333333333333, + "grad_norm": 1.5986337471335546, + "learning_rate": 1.9627112196804618e-07, + "loss": 0.6503, + "step": 2475 + }, + { + "epoch": 4.5851851851851855, + "grad_norm": 1.7207611157671368, + "learning_rate": 1.962670644773913e-07, + "loss": 0.6656, + "step": 2476 + }, + { + "epoch": 4.587037037037037, + "grad_norm": 1.7472845925595935, + "learning_rate": 1.9626300482239139e-07, + "loss": 0.615, + "step": 2477 + }, + { + "epoch": 4.588888888888889, + "grad_norm": 1.5655867417505582, + "learning_rate": 1.9625894300313766e-07, + "loss": 0.683, + "step": 2478 + }, + { + "epoch": 4.590740740740741, + "grad_norm": 1.4673176152861458, + "learning_rate": 1.9625487901972146e-07, + "loss": 0.6685, + "step": 2479 + }, + { + "epoch": 4.592592592592593, + "grad_norm": 1.3277572844281333, + "learning_rate": 1.9625081287223417e-07, + "loss": 0.6612, + "step": 2480 + }, + { + "epoch": 4.594444444444444, + "grad_norm": 1.5876659714430752, + "learning_rate": 1.9624674456076717e-07, + "loss": 0.7101, + "step": 2481 + }, + { + "epoch": 4.5962962962962965, + "grad_norm": 1.6105766675469035, + "learning_rate": 1.9624267408541197e-07, + "loss": 0.6791, + "step": 2482 + }, + { + "epoch": 4.598148148148148, + "grad_norm": 1.6155759236205214, + "learning_rate": 1.9623860144626007e-07, + "loss": 0.716, + "step": 2483 + }, + { + "epoch": 4.6, + "grad_norm": 1.3439939259020257, + "learning_rate": 1.9623452664340303e-07, + "loss": 0.6676, + "step": 2484 + }, + { + "epoch": 4.601851851851852, + "grad_norm": 1.4679818582936264, + "learning_rate": 1.9623044967693242e-07, + "loss": 0.6811, + "step": 2485 + }, + { + "epoch": 4.603703703703704, + "grad_norm": 1.559172559240203, + "learning_rate": 1.9622637054694e-07, + "loss": 0.6472, + "step": 2486 + }, + { + "epoch": 4.605555555555555, + "grad_norm": 1.6036210917023734, + "learning_rate": 1.9622228925351737e-07, + "loss": 0.6962, + "step": 2487 + }, + { + "epoch": 4.607407407407408, + "grad_norm": 1.3561314436215313, + "learning_rate": 1.9621820579675637e-07, + "loss": 0.637, + "step": 2488 + }, + { + "epoch": 4.609259259259259, + "grad_norm": 1.3972570028452092, + "learning_rate": 1.9621412017674877e-07, + "loss": 0.7311, + "step": 2489 + }, + { + "epoch": 4.611111111111111, + "grad_norm": 1.3740122124621938, + "learning_rate": 1.9621003239358648e-07, + "loss": 0.7166, + "step": 2490 + }, + { + "epoch": 4.612962962962963, + "grad_norm": 1.689510697769527, + "learning_rate": 1.962059424473613e-07, + "loss": 0.7339, + "step": 2491 + }, + { + "epoch": 4.614814814814815, + "grad_norm": 1.4247349983545494, + "learning_rate": 1.962018503381653e-07, + "loss": 0.6459, + "step": 2492 + }, + { + "epoch": 4.616666666666667, + "grad_norm": 1.6077183637468233, + "learning_rate": 1.9619775606609036e-07, + "loss": 0.689, + "step": 2493 + }, + { + "epoch": 4.618518518518519, + "grad_norm": 1.621950037448575, + "learning_rate": 1.9619365963122862e-07, + "loss": 0.6728, + "step": 2494 + }, + { + "epoch": 4.62037037037037, + "grad_norm": 1.588792512223338, + "learning_rate": 1.9618956103367216e-07, + "loss": 0.6673, + "step": 2495 + }, + { + "epoch": 4.622222222222222, + "grad_norm": 1.3558640415594747, + "learning_rate": 1.9618546027351311e-07, + "loss": 0.6748, + "step": 2496 + }, + { + "epoch": 4.624074074074074, + "grad_norm": 1.4596811440019657, + "learning_rate": 1.961813573508437e-07, + "loss": 0.6683, + "step": 2497 + }, + { + "epoch": 4.625925925925926, + "grad_norm": 1.5338412161859356, + "learning_rate": 1.9617725226575615e-07, + "loss": 0.6908, + "step": 2498 + }, + { + "epoch": 4.627777777777778, + "grad_norm": 1.610252367001902, + "learning_rate": 1.9617314501834273e-07, + "loss": 0.6915, + "step": 2499 + }, + { + "epoch": 4.62962962962963, + "grad_norm": 1.6332266311468415, + "learning_rate": 1.9616903560869583e-07, + "loss": 0.7113, + "step": 2500 + }, + { + "epoch": 4.631481481481481, + "grad_norm": 1.6114485279437283, + "learning_rate": 1.961649240369078e-07, + "loss": 0.6711, + "step": 2501 + }, + { + "epoch": 4.633333333333333, + "grad_norm": 1.406623520026227, + "learning_rate": 1.961608103030711e-07, + "loss": 0.6527, + "step": 2502 + }, + { + "epoch": 4.635185185185185, + "grad_norm": 1.9476433823561836, + "learning_rate": 1.9615669440727825e-07, + "loss": 0.677, + "step": 2503 + }, + { + "epoch": 4.637037037037037, + "grad_norm": 1.451848427975165, + "learning_rate": 1.961525763496217e-07, + "loss": 0.6561, + "step": 2504 + }, + { + "epoch": 4.638888888888889, + "grad_norm": 1.4690683263411852, + "learning_rate": 1.9614845613019413e-07, + "loss": 0.6891, + "step": 2505 + }, + { + "epoch": 4.640740740740741, + "grad_norm": 1.473266870790206, + "learning_rate": 1.961443337490881e-07, + "loss": 0.6572, + "step": 2506 + }, + { + "epoch": 4.642592592592592, + "grad_norm": 1.3814093429351664, + "learning_rate": 1.9614020920639633e-07, + "loss": 0.6416, + "step": 2507 + }, + { + "epoch": 4.644444444444445, + "grad_norm": 1.4335432358850984, + "learning_rate": 1.9613608250221156e-07, + "loss": 0.6912, + "step": 2508 + }, + { + "epoch": 4.646296296296296, + "grad_norm": 1.54130799663505, + "learning_rate": 1.9613195363662656e-07, + "loss": 0.7293, + "step": 2509 + }, + { + "epoch": 4.648148148148148, + "grad_norm": 1.4140725259659122, + "learning_rate": 1.9612782260973414e-07, + "loss": 0.6871, + "step": 2510 + }, + { + "epoch": 4.65, + "grad_norm": 1.3933266209765902, + "learning_rate": 1.9612368942162717e-07, + "loss": 0.6884, + "step": 2511 + }, + { + "epoch": 4.651851851851852, + "grad_norm": 1.6801208131532885, + "learning_rate": 1.961195540723986e-07, + "loss": 0.6966, + "step": 2512 + }, + { + "epoch": 4.6537037037037035, + "grad_norm": 1.5616389800348724, + "learning_rate": 1.9611541656214142e-07, + "loss": 0.7646, + "step": 2513 + }, + { + "epoch": 4.655555555555556, + "grad_norm": 1.6018904201339035, + "learning_rate": 1.961112768909486e-07, + "loss": 0.7127, + "step": 2514 + }, + { + "epoch": 4.657407407407407, + "grad_norm": 1.536136422995422, + "learning_rate": 1.9610713505891327e-07, + "loss": 0.661, + "step": 2515 + }, + { + "epoch": 4.659259259259259, + "grad_norm": 1.2916127592865736, + "learning_rate": 1.9610299106612848e-07, + "loss": 0.6235, + "step": 2516 + }, + { + "epoch": 4.661111111111111, + "grad_norm": 1.4296912572085654, + "learning_rate": 1.9609884491268748e-07, + "loss": 0.716, + "step": 2517 + }, + { + "epoch": 4.662962962962963, + "grad_norm": 1.3975080778540827, + "learning_rate": 1.9609469659868344e-07, + "loss": 0.6644, + "step": 2518 + }, + { + "epoch": 4.6648148148148145, + "grad_norm": 1.5320929309325129, + "learning_rate": 1.9609054612420963e-07, + "loss": 0.6219, + "step": 2519 + }, + { + "epoch": 4.666666666666667, + "grad_norm": 1.5315430129556165, + "learning_rate": 1.9608639348935936e-07, + "loss": 0.7452, + "step": 2520 + }, + { + "epoch": 4.6685185185185185, + "grad_norm": 1.7254608249682457, + "learning_rate": 1.9608223869422597e-07, + "loss": 0.7178, + "step": 2521 + }, + { + "epoch": 4.67037037037037, + "grad_norm": 1.504428069918678, + "learning_rate": 1.9607808173890294e-07, + "loss": 0.696, + "step": 2522 + }, + { + "epoch": 4.6722222222222225, + "grad_norm": 1.5322461300382144, + "learning_rate": 1.960739226234837e-07, + "loss": 0.7157, + "step": 2523 + }, + { + "epoch": 4.674074074074074, + "grad_norm": 1.5900607180789752, + "learning_rate": 1.960697613480617e-07, + "loss": 0.6546, + "step": 2524 + }, + { + "epoch": 4.675925925925926, + "grad_norm": 1.5930014641807357, + "learning_rate": 1.960655979127306e-07, + "loss": 0.7054, + "step": 2525 + }, + { + "epoch": 4.677777777777778, + "grad_norm": 1.5777043461019717, + "learning_rate": 1.960614323175839e-07, + "loss": 0.745, + "step": 2526 + }, + { + "epoch": 4.67962962962963, + "grad_norm": 1.672902950794046, + "learning_rate": 1.960572645627153e-07, + "loss": 0.6302, + "step": 2527 + }, + { + "epoch": 4.681481481481481, + "grad_norm": 1.7492836158072649, + "learning_rate": 1.9605309464821852e-07, + "loss": 0.6689, + "step": 2528 + }, + { + "epoch": 4.683333333333334, + "grad_norm": 1.4531047783738247, + "learning_rate": 1.960489225741873e-07, + "loss": 0.6562, + "step": 2529 + }, + { + "epoch": 4.685185185185185, + "grad_norm": 1.7522197282750123, + "learning_rate": 1.9604474834071542e-07, + "loss": 0.642, + "step": 2530 + }, + { + "epoch": 4.687037037037037, + "grad_norm": 1.9035753566690667, + "learning_rate": 1.9604057194789676e-07, + "loss": 0.7133, + "step": 2531 + }, + { + "epoch": 4.688888888888889, + "grad_norm": 2.249287265532964, + "learning_rate": 1.960363933958252e-07, + "loss": 0.7063, + "step": 2532 + }, + { + "epoch": 4.690740740740741, + "grad_norm": 2.000160447310834, + "learning_rate": 1.9603221268459468e-07, + "loss": 0.7022, + "step": 2533 + }, + { + "epoch": 4.692592592592592, + "grad_norm": 1.5689898123125854, + "learning_rate": 1.960280298142992e-07, + "loss": 0.6588, + "step": 2534 + }, + { + "epoch": 4.694444444444445, + "grad_norm": 1.4839313249873542, + "learning_rate": 1.9602384478503283e-07, + "loss": 0.6588, + "step": 2535 + }, + { + "epoch": 4.696296296296296, + "grad_norm": 1.404681618289382, + "learning_rate": 1.960196575968896e-07, + "loss": 0.6782, + "step": 2536 + }, + { + "epoch": 4.698148148148148, + "grad_norm": 1.5440557451402022, + "learning_rate": 1.9601546824996369e-07, + "loss": 0.7025, + "step": 2537 + }, + { + "epoch": 4.7, + "grad_norm": 1.5209307620580186, + "learning_rate": 1.9601127674434928e-07, + "loss": 0.7049, + "step": 2538 + }, + { + "epoch": 4.701851851851852, + "grad_norm": 1.4882912604297536, + "learning_rate": 1.9600708308014057e-07, + "loss": 0.7011, + "step": 2539 + }, + { + "epoch": 4.703703703703704, + "grad_norm": 1.376471744192453, + "learning_rate": 1.9600288725743192e-07, + "loss": 0.7226, + "step": 2540 + }, + { + "epoch": 4.705555555555556, + "grad_norm": 1.395341320189781, + "learning_rate": 1.9599868927631765e-07, + "loss": 0.6449, + "step": 2541 + }, + { + "epoch": 4.707407407407407, + "grad_norm": 1.523134352032048, + "learning_rate": 1.9599448913689204e-07, + "loss": 0.7144, + "step": 2542 + }, + { + "epoch": 4.709259259259259, + "grad_norm": 2.0961339458867516, + "learning_rate": 1.9599028683924964e-07, + "loss": 0.703, + "step": 2543 + }, + { + "epoch": 4.711111111111111, + "grad_norm": 1.3300624922406747, + "learning_rate": 1.959860823834849e-07, + "loss": 0.6339, + "step": 2544 + }, + { + "epoch": 4.712962962962963, + "grad_norm": 1.302116166969961, + "learning_rate": 1.9598187576969234e-07, + "loss": 0.7167, + "step": 2545 + }, + { + "epoch": 4.714814814814815, + "grad_norm": 1.4664818998034495, + "learning_rate": 1.959776669979665e-07, + "loss": 0.7121, + "step": 2546 + }, + { + "epoch": 4.716666666666667, + "grad_norm": 1.4684768506851666, + "learning_rate": 1.9597345606840206e-07, + "loss": 0.6771, + "step": 2547 + }, + { + "epoch": 4.718518518518518, + "grad_norm": 1.3619799791310647, + "learning_rate": 1.9596924298109365e-07, + "loss": 0.6889, + "step": 2548 + }, + { + "epoch": 4.72037037037037, + "grad_norm": 1.9874509884656524, + "learning_rate": 1.9596502773613602e-07, + "loss": 0.6676, + "step": 2549 + }, + { + "epoch": 4.722222222222222, + "grad_norm": 1.4666539531597036, + "learning_rate": 1.9596081033362393e-07, + "loss": 0.7081, + "step": 2550 + }, + { + "epoch": 4.724074074074074, + "grad_norm": 1.4591212775113622, + "learning_rate": 1.959565907736522e-07, + "loss": 0.6821, + "step": 2551 + }, + { + "epoch": 4.725925925925926, + "grad_norm": 1.6105275042909308, + "learning_rate": 1.9595236905631567e-07, + "loss": 0.6451, + "step": 2552 + }, + { + "epoch": 4.727777777777778, + "grad_norm": 1.5352243609040608, + "learning_rate": 1.959481451817093e-07, + "loss": 0.6831, + "step": 2553 + }, + { + "epoch": 4.729629629629629, + "grad_norm": 6.049561837551794, + "learning_rate": 1.9594391914992803e-07, + "loss": 0.6992, + "step": 2554 + }, + { + "epoch": 4.731481481481482, + "grad_norm": 1.6117055707567023, + "learning_rate": 1.9593969096106688e-07, + "loss": 0.6822, + "step": 2555 + }, + { + "epoch": 4.733333333333333, + "grad_norm": 1.6100100616769308, + "learning_rate": 1.959354606152209e-07, + "loss": 0.7079, + "step": 2556 + }, + { + "epoch": 4.735185185185185, + "grad_norm": 1.3680011324512609, + "learning_rate": 1.9593122811248522e-07, + "loss": 0.633, + "step": 2557 + }, + { + "epoch": 4.737037037037037, + "grad_norm": 1.533417437300952, + "learning_rate": 1.95926993452955e-07, + "loss": 0.6672, + "step": 2558 + }, + { + "epoch": 4.738888888888889, + "grad_norm": 4.351594857068372, + "learning_rate": 1.9592275663672543e-07, + "loss": 0.6345, + "step": 2559 + }, + { + "epoch": 4.7407407407407405, + "grad_norm": 1.5497296381413772, + "learning_rate": 1.9591851766389174e-07, + "loss": 0.7001, + "step": 2560 + }, + { + "epoch": 4.742592592592593, + "grad_norm": 1.3588472471020898, + "learning_rate": 1.959142765345493e-07, + "loss": 0.6969, + "step": 2561 + }, + { + "epoch": 4.7444444444444445, + "grad_norm": 2.144818587608991, + "learning_rate": 1.9591003324879338e-07, + "loss": 0.6832, + "step": 2562 + }, + { + "epoch": 4.746296296296296, + "grad_norm": 1.3072838602291186, + "learning_rate": 1.9590578780671945e-07, + "loss": 0.6666, + "step": 2563 + }, + { + "epoch": 4.7481481481481485, + "grad_norm": 1.3331818712867731, + "learning_rate": 1.9590154020842294e-07, + "loss": 0.6936, + "step": 2564 + }, + { + "epoch": 4.75, + "grad_norm": 1.699866272058813, + "learning_rate": 1.9589729045399933e-07, + "loss": 0.7237, + "step": 2565 + }, + { + "epoch": 4.7518518518518515, + "grad_norm": 1.4264088677582054, + "learning_rate": 1.9589303854354416e-07, + "loss": 0.7232, + "step": 2566 + }, + { + "epoch": 4.753703703703704, + "grad_norm": 1.5549432734879838, + "learning_rate": 1.9588878447715306e-07, + "loss": 0.7117, + "step": 2567 + }, + { + "epoch": 4.7555555555555555, + "grad_norm": 1.4472904339605825, + "learning_rate": 1.9588452825492168e-07, + "loss": 0.7106, + "step": 2568 + }, + { + "epoch": 4.757407407407407, + "grad_norm": 1.5394126519144269, + "learning_rate": 1.9588026987694565e-07, + "loss": 0.6925, + "step": 2569 + }, + { + "epoch": 4.7592592592592595, + "grad_norm": 2.971944118039773, + "learning_rate": 1.9587600934332075e-07, + "loss": 0.6522, + "step": 2570 + }, + { + "epoch": 4.761111111111111, + "grad_norm": 1.3407780155578763, + "learning_rate": 1.9587174665414282e-07, + "loss": 0.6632, + "step": 2571 + }, + { + "epoch": 4.762962962962963, + "grad_norm": 1.6055322955615947, + "learning_rate": 1.9586748180950758e-07, + "loss": 0.6814, + "step": 2572 + }, + { + "epoch": 4.764814814814815, + "grad_norm": 1.4152320630850797, + "learning_rate": 1.95863214809511e-07, + "loss": 0.6733, + "step": 2573 + }, + { + "epoch": 4.766666666666667, + "grad_norm": 1.4929362577876002, + "learning_rate": 1.9585894565424898e-07, + "loss": 0.6939, + "step": 2574 + }, + { + "epoch": 4.768518518518518, + "grad_norm": 1.460521891075413, + "learning_rate": 1.9585467434381751e-07, + "loss": 0.6585, + "step": 2575 + }, + { + "epoch": 4.770370370370371, + "grad_norm": 1.461293730109857, + "learning_rate": 1.9585040087831265e-07, + "loss": 0.6135, + "step": 2576 + }, + { + "epoch": 4.772222222222222, + "grad_norm": 2.3813280478966883, + "learning_rate": 1.9584612525783042e-07, + "loss": 0.6652, + "step": 2577 + }, + { + "epoch": 4.774074074074074, + "grad_norm": 1.8412587455644527, + "learning_rate": 1.95841847482467e-07, + "loss": 0.6484, + "step": 2578 + }, + { + "epoch": 4.775925925925926, + "grad_norm": 1.4172860772977958, + "learning_rate": 1.9583756755231853e-07, + "loss": 0.6529, + "step": 2579 + }, + { + "epoch": 4.777777777777778, + "grad_norm": 1.5530412211210471, + "learning_rate": 1.9583328546748125e-07, + "loss": 0.6268, + "step": 2580 + }, + { + "epoch": 4.77962962962963, + "grad_norm": 1.535534316611214, + "learning_rate": 1.9582900122805145e-07, + "loss": 0.6641, + "step": 2581 + }, + { + "epoch": 4.781481481481482, + "grad_norm": 2.1748549856854567, + "learning_rate": 1.9582471483412543e-07, + "loss": 0.655, + "step": 2582 + }, + { + "epoch": 4.783333333333333, + "grad_norm": 1.6841573333989606, + "learning_rate": 1.9582042628579956e-07, + "loss": 0.6864, + "step": 2583 + }, + { + "epoch": 4.785185185185185, + "grad_norm": 1.5311357546678117, + "learning_rate": 1.9581613558317027e-07, + "loss": 0.6467, + "step": 2584 + }, + { + "epoch": 4.787037037037037, + "grad_norm": 2.8057207953903953, + "learning_rate": 1.95811842726334e-07, + "loss": 0.6969, + "step": 2585 + }, + { + "epoch": 4.788888888888889, + "grad_norm": 1.718009726212732, + "learning_rate": 1.9580754771538732e-07, + "loss": 0.6807, + "step": 2586 + }, + { + "epoch": 4.790740740740741, + "grad_norm": 1.4090495628613642, + "learning_rate": 1.9580325055042674e-07, + "loss": 0.7133, + "step": 2587 + }, + { + "epoch": 4.792592592592593, + "grad_norm": 2.0765659934498637, + "learning_rate": 1.9579895123154887e-07, + "loss": 0.7212, + "step": 2588 + }, + { + "epoch": 4.794444444444444, + "grad_norm": 1.5506387328654636, + "learning_rate": 1.9579464975885042e-07, + "loss": 0.6659, + "step": 2589 + }, + { + "epoch": 4.796296296296296, + "grad_norm": 1.3343560966769126, + "learning_rate": 1.9579034613242804e-07, + "loss": 0.6534, + "step": 2590 + }, + { + "epoch": 4.798148148148148, + "grad_norm": 1.5716720780290347, + "learning_rate": 1.9578604035237856e-07, + "loss": 0.691, + "step": 2591 + }, + { + "epoch": 4.8, + "grad_norm": 1.3554224563661759, + "learning_rate": 1.957817324187987e-07, + "loss": 0.6639, + "step": 2592 + }, + { + "epoch": 4.801851851851852, + "grad_norm": 1.28732952185562, + "learning_rate": 1.9577742233178536e-07, + "loss": 0.6832, + "step": 2593 + }, + { + "epoch": 4.803703703703704, + "grad_norm": 1.5309043659866546, + "learning_rate": 1.9577311009143548e-07, + "loss": 0.6947, + "step": 2594 + }, + { + "epoch": 4.805555555555555, + "grad_norm": 1.453191398942573, + "learning_rate": 1.9576879569784592e-07, + "loss": 0.6971, + "step": 2595 + }, + { + "epoch": 4.807407407407408, + "grad_norm": 1.4143822486270237, + "learning_rate": 1.9576447915111377e-07, + "loss": 0.6854, + "step": 2596 + }, + { + "epoch": 4.809259259259259, + "grad_norm": 1.5755946097336484, + "learning_rate": 1.95760160451336e-07, + "loss": 0.6974, + "step": 2597 + }, + { + "epoch": 4.811111111111111, + "grad_norm": 1.4725985076973205, + "learning_rate": 1.9575583959860975e-07, + "loss": 0.6929, + "step": 2598 + }, + { + "epoch": 4.812962962962963, + "grad_norm": 2.002995867864467, + "learning_rate": 1.9575151659303218e-07, + "loss": 0.7016, + "step": 2599 + }, + { + "epoch": 4.814814814814815, + "grad_norm": 1.8102165871063058, + "learning_rate": 1.9574719143470042e-07, + "loss": 0.6662, + "step": 2600 + }, + { + "epoch": 4.816666666666666, + "grad_norm": 1.621177989958412, + "learning_rate": 1.9574286412371177e-07, + "loss": 0.7199, + "step": 2601 + }, + { + "epoch": 4.818518518518519, + "grad_norm": 1.5148548287176322, + "learning_rate": 1.9573853466016354e-07, + "loss": 0.6733, + "step": 2602 + }, + { + "epoch": 4.82037037037037, + "grad_norm": 1.8660690254127292, + "learning_rate": 1.95734203044153e-07, + "loss": 0.6594, + "step": 2603 + }, + { + "epoch": 4.822222222222222, + "grad_norm": 1.9403288070156381, + "learning_rate": 1.9572986927577754e-07, + "loss": 0.6994, + "step": 2604 + }, + { + "epoch": 4.824074074074074, + "grad_norm": 1.5574392648529414, + "learning_rate": 1.9572553335513465e-07, + "loss": 0.6442, + "step": 2605 + }, + { + "epoch": 4.825925925925926, + "grad_norm": 1.9651121053242995, + "learning_rate": 1.9572119528232175e-07, + "loss": 0.6684, + "step": 2606 + }, + { + "epoch": 4.8277777777777775, + "grad_norm": 1.4836938331881797, + "learning_rate": 1.957168550574364e-07, + "loss": 0.6564, + "step": 2607 + }, + { + "epoch": 4.82962962962963, + "grad_norm": 2.074361031300887, + "learning_rate": 1.9571251268057626e-07, + "loss": 0.7036, + "step": 2608 + }, + { + "epoch": 4.8314814814814815, + "grad_norm": 1.3913488668062026, + "learning_rate": 1.957081681518388e-07, + "loss": 0.6491, + "step": 2609 + }, + { + "epoch": 4.833333333333333, + "grad_norm": 1.6698805886455452, + "learning_rate": 1.9570382147132185e-07, + "loss": 0.6576, + "step": 2610 + }, + { + "epoch": 4.8351851851851855, + "grad_norm": 1.6692744102885133, + "learning_rate": 1.9569947263912302e-07, + "loss": 0.6918, + "step": 2611 + }, + { + "epoch": 4.837037037037037, + "grad_norm": 2.250237785541846, + "learning_rate": 1.9569512165534015e-07, + "loss": 0.6858, + "step": 2612 + }, + { + "epoch": 4.838888888888889, + "grad_norm": 2.0258299823379997, + "learning_rate": 1.9569076852007104e-07, + "loss": 0.6792, + "step": 2613 + }, + { + "epoch": 4.840740740740741, + "grad_norm": 1.3166064995721136, + "learning_rate": 1.9568641323341355e-07, + "loss": 0.6862, + "step": 2614 + }, + { + "epoch": 4.842592592592593, + "grad_norm": 1.6378534252343364, + "learning_rate": 1.9568205579546561e-07, + "loss": 0.6476, + "step": 2615 + }, + { + "epoch": 4.844444444444444, + "grad_norm": 1.7411619440573904, + "learning_rate": 1.956776962063252e-07, + "loss": 0.6766, + "step": 2616 + }, + { + "epoch": 4.8462962962962965, + "grad_norm": 1.4902231341231, + "learning_rate": 1.956733344660903e-07, + "loss": 0.698, + "step": 2617 + }, + { + "epoch": 4.848148148148148, + "grad_norm": 1.6826732544785905, + "learning_rate": 1.95668970574859e-07, + "loss": 0.6925, + "step": 2618 + }, + { + "epoch": 4.85, + "grad_norm": 1.6132862339391547, + "learning_rate": 1.9566460453272943e-07, + "loss": 0.6426, + "step": 2619 + }, + { + "epoch": 4.851851851851852, + "grad_norm": 1.509499724322264, + "learning_rate": 1.9566023633979974e-07, + "loss": 0.6677, + "step": 2620 + }, + { + "epoch": 4.853703703703704, + "grad_norm": 1.4286323318046328, + "learning_rate": 1.956558659961681e-07, + "loss": 0.7127, + "step": 2621 + }, + { + "epoch": 4.855555555555555, + "grad_norm": 1.4604074391120105, + "learning_rate": 1.956514935019328e-07, + "loss": 0.7076, + "step": 2622 + }, + { + "epoch": 4.857407407407408, + "grad_norm": 1.3116592330715366, + "learning_rate": 1.9564711885719214e-07, + "loss": 0.6899, + "step": 2623 + }, + { + "epoch": 4.859259259259259, + "grad_norm": 1.4730802042284514, + "learning_rate": 1.9564274206204447e-07, + "loss": 0.6729, + "step": 2624 + }, + { + "epoch": 4.861111111111111, + "grad_norm": 1.4718604988630335, + "learning_rate": 1.9563836311658822e-07, + "loss": 0.7056, + "step": 2625 + }, + { + "epoch": 4.862962962962963, + "grad_norm": 1.4359698303780324, + "learning_rate": 1.956339820209218e-07, + "loss": 0.6728, + "step": 2626 + }, + { + "epoch": 4.864814814814815, + "grad_norm": 1.6351348722401102, + "learning_rate": 1.9562959877514374e-07, + "loss": 0.7104, + "step": 2627 + }, + { + "epoch": 4.866666666666667, + "grad_norm": 1.5378185101318458, + "learning_rate": 1.9562521337935253e-07, + "loss": 0.6885, + "step": 2628 + }, + { + "epoch": 4.868518518518519, + "grad_norm": 4.065624679685904, + "learning_rate": 1.9562082583364686e-07, + "loss": 0.6934, + "step": 2629 + }, + { + "epoch": 4.87037037037037, + "grad_norm": 1.5330305515150862, + "learning_rate": 1.956164361381253e-07, + "loss": 0.6545, + "step": 2630 + }, + { + "epoch": 4.872222222222222, + "grad_norm": 1.5792460026013166, + "learning_rate": 1.9561204429288657e-07, + "loss": 0.6585, + "step": 2631 + }, + { + "epoch": 4.874074074074074, + "grad_norm": 1.6852842892947653, + "learning_rate": 1.956076502980294e-07, + "loss": 0.6856, + "step": 2632 + }, + { + "epoch": 4.875925925925926, + "grad_norm": 1.446362265985485, + "learning_rate": 1.9560325415365257e-07, + "loss": 0.6691, + "step": 2633 + }, + { + "epoch": 4.877777777777778, + "grad_norm": 1.478944105102043, + "learning_rate": 1.9559885585985497e-07, + "loss": 0.6462, + "step": 2634 + }, + { + "epoch": 4.87962962962963, + "grad_norm": 1.3951419744936036, + "learning_rate": 1.9559445541673543e-07, + "loss": 0.6808, + "step": 2635 + }, + { + "epoch": 4.881481481481481, + "grad_norm": 1.754978855300922, + "learning_rate": 1.9559005282439292e-07, + "loss": 0.6429, + "step": 2636 + }, + { + "epoch": 4.883333333333333, + "grad_norm": 1.449261360830208, + "learning_rate": 1.9558564808292636e-07, + "loss": 0.6344, + "step": 2637 + }, + { + "epoch": 4.885185185185185, + "grad_norm": 1.5792897252697125, + "learning_rate": 1.9558124119243486e-07, + "loss": 0.6766, + "step": 2638 + }, + { + "epoch": 4.887037037037037, + "grad_norm": 1.7367672539313557, + "learning_rate": 1.9557683215301743e-07, + "loss": 0.66, + "step": 2639 + }, + { + "epoch": 4.888888888888889, + "grad_norm": 1.5862960703696574, + "learning_rate": 1.9557242096477327e-07, + "loss": 0.6994, + "step": 2640 + }, + { + "epoch": 4.890740740740741, + "grad_norm": 1.6922835335513207, + "learning_rate": 1.955680076278015e-07, + "loss": 0.694, + "step": 2641 + }, + { + "epoch": 4.892592592592592, + "grad_norm": 1.82944800147624, + "learning_rate": 1.9556359214220137e-07, + "loss": 0.705, + "step": 2642 + }, + { + "epoch": 4.894444444444445, + "grad_norm": 1.621980112509395, + "learning_rate": 1.955591745080721e-07, + "loss": 0.6569, + "step": 2643 + }, + { + "epoch": 4.896296296296296, + "grad_norm": 3.320650322721205, + "learning_rate": 1.9555475472551311e-07, + "loss": 0.6734, + "step": 2644 + }, + { + "epoch": 4.898148148148148, + "grad_norm": 1.5335604450726346, + "learning_rate": 1.9555033279462367e-07, + "loss": 0.6422, + "step": 2645 + }, + { + "epoch": 4.9, + "grad_norm": 1.3712288995221626, + "learning_rate": 1.9554590871550327e-07, + "loss": 0.6917, + "step": 2646 + }, + { + "epoch": 4.901851851851852, + "grad_norm": 1.349011385329623, + "learning_rate": 1.955414824882513e-07, + "loss": 0.6918, + "step": 2647 + }, + { + "epoch": 4.9037037037037035, + "grad_norm": 1.2876496306287544, + "learning_rate": 1.9553705411296735e-07, + "loss": 0.7059, + "step": 2648 + }, + { + "epoch": 4.905555555555556, + "grad_norm": 1.480916643859869, + "learning_rate": 1.9553262358975094e-07, + "loss": 0.643, + "step": 2649 + }, + { + "epoch": 4.907407407407407, + "grad_norm": 1.565604936070093, + "learning_rate": 1.955281909187017e-07, + "loss": 0.7096, + "step": 2650 + }, + { + "epoch": 4.909259259259259, + "grad_norm": 1.4676536126054007, + "learning_rate": 1.9552375609991927e-07, + "loss": 0.6893, + "step": 2651 + }, + { + "epoch": 4.911111111111111, + "grad_norm": 1.3380304793967999, + "learning_rate": 1.9551931913350334e-07, + "loss": 0.6505, + "step": 2652 + }, + { + "epoch": 4.912962962962963, + "grad_norm": 1.5615450651318896, + "learning_rate": 1.9551488001955373e-07, + "loss": 0.6665, + "step": 2653 + }, + { + "epoch": 4.9148148148148145, + "grad_norm": 1.7988459133966521, + "learning_rate": 1.9551043875817017e-07, + "loss": 0.709, + "step": 2654 + }, + { + "epoch": 4.916666666666667, + "grad_norm": 1.3876556704870973, + "learning_rate": 1.9550599534945259e-07, + "loss": 0.6868, + "step": 2655 + }, + { + "epoch": 4.9185185185185185, + "grad_norm": 1.5617115311718706, + "learning_rate": 1.955015497935008e-07, + "loss": 0.6962, + "step": 2656 + }, + { + "epoch": 4.92037037037037, + "grad_norm": 1.8783542449190045, + "learning_rate": 1.9549710209041478e-07, + "loss": 0.6585, + "step": 2657 + }, + { + "epoch": 4.9222222222222225, + "grad_norm": 1.434083253111606, + "learning_rate": 1.9549265224029456e-07, + "loss": 0.663, + "step": 2658 + }, + { + "epoch": 4.924074074074074, + "grad_norm": 1.4424141133200097, + "learning_rate": 1.9548820024324017e-07, + "loss": 0.6924, + "step": 2659 + }, + { + "epoch": 4.925925925925926, + "grad_norm": 1.5150646349713737, + "learning_rate": 1.954837460993517e-07, + "loss": 0.685, + "step": 2660 + }, + { + "epoch": 4.927777777777778, + "grad_norm": 1.927555098185982, + "learning_rate": 1.9547928980872929e-07, + "loss": 0.7084, + "step": 2661 + }, + { + "epoch": 4.92962962962963, + "grad_norm": 1.6846270214958476, + "learning_rate": 1.9547483137147312e-07, + "loss": 0.6817, + "step": 2662 + }, + { + "epoch": 4.931481481481481, + "grad_norm": 1.5251097022343931, + "learning_rate": 1.9547037078768342e-07, + "loss": 0.6996, + "step": 2663 + }, + { + "epoch": 4.933333333333334, + "grad_norm": 2.3336481867084107, + "learning_rate": 1.9546590805746051e-07, + "loss": 0.6445, + "step": 2664 + }, + { + "epoch": 4.935185185185185, + "grad_norm": 1.4766742961198567, + "learning_rate": 1.954614431809047e-07, + "loss": 0.6804, + "step": 2665 + }, + { + "epoch": 4.937037037037037, + "grad_norm": 1.7973668305011095, + "learning_rate": 1.9545697615811637e-07, + "loss": 0.6712, + "step": 2666 + }, + { + "epoch": 4.938888888888889, + "grad_norm": 1.4628137784980484, + "learning_rate": 1.9545250698919597e-07, + "loss": 0.7261, + "step": 2667 + }, + { + "epoch": 4.940740740740741, + "grad_norm": 1.4439564852227373, + "learning_rate": 1.9544803567424398e-07, + "loss": 0.6986, + "step": 2668 + }, + { + "epoch": 4.942592592592592, + "grad_norm": 1.6770071853074, + "learning_rate": 1.954435622133609e-07, + "loss": 0.7004, + "step": 2669 + }, + { + "epoch": 4.944444444444445, + "grad_norm": 1.4701736895259265, + "learning_rate": 1.9543908660664733e-07, + "loss": 0.6944, + "step": 2670 + }, + { + "epoch": 4.946296296296296, + "grad_norm": 1.6089138863677834, + "learning_rate": 1.9543460885420386e-07, + "loss": 0.6573, + "step": 2671 + }, + { + "epoch": 4.948148148148148, + "grad_norm": 1.4658356954218936, + "learning_rate": 1.954301289561312e-07, + "loss": 0.6548, + "step": 2672 + }, + { + "epoch": 4.95, + "grad_norm": 2.055443747313343, + "learning_rate": 1.9542564691253007e-07, + "loss": 0.7419, + "step": 2673 + }, + { + "epoch": 4.951851851851852, + "grad_norm": 1.407779024281975, + "learning_rate": 1.954211627235012e-07, + "loss": 0.6733, + "step": 2674 + }, + { + "epoch": 4.953703703703704, + "grad_norm": 1.737115123653284, + "learning_rate": 1.9541667638914546e-07, + "loss": 0.6543, + "step": 2675 + }, + { + "epoch": 4.955555555555556, + "grad_norm": 1.4803068485974478, + "learning_rate": 1.9541218790956367e-07, + "loss": 0.7033, + "step": 2676 + }, + { + "epoch": 4.957407407407407, + "grad_norm": 1.5013824918581729, + "learning_rate": 1.9540769728485674e-07, + "loss": 0.7117, + "step": 2677 + }, + { + "epoch": 4.959259259259259, + "grad_norm": 1.3156214150479348, + "learning_rate": 1.9540320451512567e-07, + "loss": 0.6867, + "step": 2678 + }, + { + "epoch": 4.961111111111111, + "grad_norm": 1.4407186739701892, + "learning_rate": 1.953987096004715e-07, + "loss": 0.6437, + "step": 2679 + }, + { + "epoch": 4.962962962962963, + "grad_norm": 1.5999497591386125, + "learning_rate": 1.9539421254099516e-07, + "loss": 0.7113, + "step": 2680 + }, + { + "epoch": 4.964814814814815, + "grad_norm": 1.4142931520112962, + "learning_rate": 1.9538971333679787e-07, + "loss": 0.7224, + "step": 2681 + }, + { + "epoch": 4.966666666666667, + "grad_norm": 1.3331528856433439, + "learning_rate": 1.9538521198798076e-07, + "loss": 0.701, + "step": 2682 + }, + { + "epoch": 4.968518518518518, + "grad_norm": 1.6423688070666986, + "learning_rate": 1.95380708494645e-07, + "loss": 0.6539, + "step": 2683 + }, + { + "epoch": 4.97037037037037, + "grad_norm": 1.3822785224406957, + "learning_rate": 1.953762028568919e-07, + "loss": 0.6442, + "step": 2684 + }, + { + "epoch": 4.972222222222222, + "grad_norm": 1.4700579032771612, + "learning_rate": 1.9537169507482267e-07, + "loss": 0.7062, + "step": 2685 + }, + { + "epoch": 4.974074074074074, + "grad_norm": 1.7922159796105215, + "learning_rate": 1.9536718514853873e-07, + "loss": 0.6698, + "step": 2686 + }, + { + "epoch": 4.975925925925926, + "grad_norm": 1.4652734939706755, + "learning_rate": 1.9536267307814147e-07, + "loss": 0.6257, + "step": 2687 + }, + { + "epoch": 4.977777777777778, + "grad_norm": 1.8360566393794944, + "learning_rate": 1.953581588637323e-07, + "loss": 0.709, + "step": 2688 + }, + { + "epoch": 4.979629629629629, + "grad_norm": 1.7996465469168061, + "learning_rate": 1.9535364250541274e-07, + "loss": 0.6548, + "step": 2689 + }, + { + "epoch": 4.981481481481482, + "grad_norm": 1.5514553935132824, + "learning_rate": 1.9534912400328433e-07, + "loss": 0.6584, + "step": 2690 + }, + { + "epoch": 4.983333333333333, + "grad_norm": 1.5941852534699035, + "learning_rate": 1.9534460335744864e-07, + "loss": 0.697, + "step": 2691 + }, + { + "epoch": 4.985185185185185, + "grad_norm": 1.5088265875512994, + "learning_rate": 1.953400805680073e-07, + "loss": 0.6538, + "step": 2692 + }, + { + "epoch": 4.987037037037037, + "grad_norm": 1.7594699460268615, + "learning_rate": 1.9533555563506202e-07, + "loss": 0.7083, + "step": 2693 + }, + { + "epoch": 4.988888888888889, + "grad_norm": 2.3528463802717994, + "learning_rate": 1.9533102855871454e-07, + "loss": 0.6566, + "step": 2694 + }, + { + "epoch": 4.9907407407407405, + "grad_norm": 1.6394322103731769, + "learning_rate": 1.953264993390666e-07, + "loss": 0.6996, + "step": 2695 + }, + { + "epoch": 4.992592592592593, + "grad_norm": 1.5742827726219282, + "learning_rate": 1.9532196797622005e-07, + "loss": 0.6638, + "step": 2696 + }, + { + "epoch": 4.9944444444444445, + "grad_norm": 1.4852115191028312, + "learning_rate": 1.9531743447027678e-07, + "loss": 0.6553, + "step": 2697 + }, + { + "epoch": 4.996296296296296, + "grad_norm": 1.7456692094689956, + "learning_rate": 1.953128988213387e-07, + "loss": 0.6787, + "step": 2698 + }, + { + "epoch": 4.9981481481481485, + "grad_norm": 1.4942473152661144, + "learning_rate": 1.9530836102950777e-07, + "loss": 0.6967, + "step": 2699 + }, + { + "epoch": 5.0, + "grad_norm": 1.5898572034515874, + "learning_rate": 1.9530382109488609e-07, + "loss": 0.6387, + "step": 2700 + }, + { + "epoch": 5.0018518518518515, + "grad_norm": 1.8622424932831836, + "learning_rate": 1.9529927901757563e-07, + "loss": 0.6643, + "step": 2701 + }, + { + "epoch": 5.003703703703704, + "grad_norm": 1.7352849687357825, + "learning_rate": 1.9529473479767855e-07, + "loss": 0.6157, + "step": 2702 + }, + { + "epoch": 5.0055555555555555, + "grad_norm": 1.538745971169652, + "learning_rate": 1.95290188435297e-07, + "loss": 0.6936, + "step": 2703 + }, + { + "epoch": 5.007407407407407, + "grad_norm": 1.5123732026287093, + "learning_rate": 1.9528563993053325e-07, + "loss": 0.6456, + "step": 2704 + }, + { + "epoch": 5.0092592592592595, + "grad_norm": 1.7103623939784156, + "learning_rate": 1.9528108928348948e-07, + "loss": 0.671, + "step": 2705 + }, + { + "epoch": 5.011111111111111, + "grad_norm": 2.2441659113562724, + "learning_rate": 1.9527653649426807e-07, + "loss": 0.6841, + "step": 2706 + }, + { + "epoch": 5.012962962962963, + "grad_norm": 2.013139402638732, + "learning_rate": 1.9527198156297133e-07, + "loss": 0.6311, + "step": 2707 + }, + { + "epoch": 5.014814814814815, + "grad_norm": 1.4998899794777623, + "learning_rate": 1.952674244897017e-07, + "loss": 0.6905, + "step": 2708 + }, + { + "epoch": 5.016666666666667, + "grad_norm": 1.58898736856241, + "learning_rate": 1.9526286527456163e-07, + "loss": 0.6762, + "step": 2709 + }, + { + "epoch": 5.018518518518518, + "grad_norm": 1.4504511132858373, + "learning_rate": 1.9525830391765363e-07, + "loss": 0.6571, + "step": 2710 + }, + { + "epoch": 5.020370370370371, + "grad_norm": 1.941280591241212, + "learning_rate": 1.952537404190802e-07, + "loss": 0.6777, + "step": 2711 + }, + { + "epoch": 5.022222222222222, + "grad_norm": 1.5970643986501012, + "learning_rate": 1.9524917477894402e-07, + "loss": 0.6477, + "step": 2712 + }, + { + "epoch": 5.024074074074074, + "grad_norm": 3.47689999741679, + "learning_rate": 1.9524460699734768e-07, + "loss": 0.6798, + "step": 2713 + }, + { + "epoch": 5.025925925925926, + "grad_norm": 1.4961824294695965, + "learning_rate": 1.952400370743939e-07, + "loss": 0.6993, + "step": 2714 + }, + { + "epoch": 5.027777777777778, + "grad_norm": 1.4592406550752999, + "learning_rate": 1.9523546501018542e-07, + "loss": 0.6578, + "step": 2715 + }, + { + "epoch": 5.029629629629629, + "grad_norm": 1.5687043212999647, + "learning_rate": 1.95230890804825e-07, + "loss": 0.7169, + "step": 2716 + }, + { + "epoch": 5.031481481481482, + "grad_norm": 1.4784765478469364, + "learning_rate": 1.9522631445841559e-07, + "loss": 0.711, + "step": 2717 + }, + { + "epoch": 5.033333333333333, + "grad_norm": 1.4771210302151068, + "learning_rate": 1.9522173597105995e-07, + "loss": 0.7465, + "step": 2718 + }, + { + "epoch": 5.035185185185185, + "grad_norm": 2.294481126889156, + "learning_rate": 1.9521715534286104e-07, + "loss": 0.6801, + "step": 2719 + }, + { + "epoch": 5.037037037037037, + "grad_norm": 1.4992248393087069, + "learning_rate": 1.9521257257392192e-07, + "loss": 0.6614, + "step": 2720 + }, + { + "epoch": 5.038888888888889, + "grad_norm": 1.652181246645023, + "learning_rate": 1.9520798766434553e-07, + "loss": 0.6695, + "step": 2721 + }, + { + "epoch": 5.040740740740741, + "grad_norm": 1.5520728647514515, + "learning_rate": 1.9520340061423505e-07, + "loss": 0.6692, + "step": 2722 + }, + { + "epoch": 5.042592592592593, + "grad_norm": 1.642697153147436, + "learning_rate": 1.9519881142369353e-07, + "loss": 0.7678, + "step": 2723 + }, + { + "epoch": 5.044444444444444, + "grad_norm": 1.8413925747381803, + "learning_rate": 1.9519422009282417e-07, + "loss": 0.6902, + "step": 2724 + }, + { + "epoch": 5.046296296296297, + "grad_norm": 1.558458270045277, + "learning_rate": 1.9518962662173019e-07, + "loss": 0.644, + "step": 2725 + }, + { + "epoch": 5.048148148148148, + "grad_norm": 1.506289546349195, + "learning_rate": 1.951850310105149e-07, + "loss": 0.692, + "step": 2726 + }, + { + "epoch": 5.05, + "grad_norm": 1.742068937604263, + "learning_rate": 1.9518043325928156e-07, + "loss": 0.6747, + "step": 2727 + }, + { + "epoch": 5.051851851851852, + "grad_norm": 1.6323158604015076, + "learning_rate": 1.951758333681336e-07, + "loss": 0.718, + "step": 2728 + }, + { + "epoch": 5.053703703703704, + "grad_norm": 1.792627807649043, + "learning_rate": 1.951712313371744e-07, + "loss": 0.6745, + "step": 2729 + }, + { + "epoch": 5.055555555555555, + "grad_norm": 1.8264906424021514, + "learning_rate": 1.9516662716650742e-07, + "loss": 0.7001, + "step": 2730 + }, + { + "epoch": 5.057407407407408, + "grad_norm": 1.6061210899553822, + "learning_rate": 1.951620208562362e-07, + "loss": 0.6726, + "step": 2731 + }, + { + "epoch": 5.059259259259259, + "grad_norm": 1.4979117063166532, + "learning_rate": 1.9515741240646431e-07, + "loss": 0.6287, + "step": 2732 + }, + { + "epoch": 5.061111111111111, + "grad_norm": 2.266713071108368, + "learning_rate": 1.9515280181729536e-07, + "loss": 0.6875, + "step": 2733 + }, + { + "epoch": 5.062962962962963, + "grad_norm": 1.517888437424714, + "learning_rate": 1.95148189088833e-07, + "loss": 0.6352, + "step": 2734 + }, + { + "epoch": 5.064814814814815, + "grad_norm": 1.5294411305699602, + "learning_rate": 1.9514357422118088e-07, + "loss": 0.6371, + "step": 2735 + }, + { + "epoch": 5.066666666666666, + "grad_norm": 1.4703201295182762, + "learning_rate": 1.9513895721444284e-07, + "loss": 0.6878, + "step": 2736 + }, + { + "epoch": 5.068518518518519, + "grad_norm": 1.9091504529435155, + "learning_rate": 1.9513433806872266e-07, + "loss": 0.6675, + "step": 2737 + }, + { + "epoch": 5.07037037037037, + "grad_norm": 1.535683250640905, + "learning_rate": 1.9512971678412416e-07, + "loss": 0.6354, + "step": 2738 + }, + { + "epoch": 5.072222222222222, + "grad_norm": 1.3864964391489514, + "learning_rate": 1.9512509336075124e-07, + "loss": 0.6704, + "step": 2739 + }, + { + "epoch": 5.074074074074074, + "grad_norm": 1.7633170844137405, + "learning_rate": 1.951204677987079e-07, + "loss": 0.6622, + "step": 2740 + }, + { + "epoch": 5.075925925925926, + "grad_norm": 1.5562407546239936, + "learning_rate": 1.9511584009809808e-07, + "loss": 0.635, + "step": 2741 + }, + { + "epoch": 5.0777777777777775, + "grad_norm": 2.5331213736466687, + "learning_rate": 1.9511121025902585e-07, + "loss": 0.6841, + "step": 2742 + }, + { + "epoch": 5.07962962962963, + "grad_norm": 1.8154606730327534, + "learning_rate": 1.951065782815953e-07, + "loss": 0.6569, + "step": 2743 + }, + { + "epoch": 5.0814814814814815, + "grad_norm": 1.5684281231840984, + "learning_rate": 1.9510194416591057e-07, + "loss": 0.6702, + "step": 2744 + }, + { + "epoch": 5.083333333333333, + "grad_norm": 1.440241814918552, + "learning_rate": 1.950973079120758e-07, + "loss": 0.6678, + "step": 2745 + }, + { + "epoch": 5.0851851851851855, + "grad_norm": 1.5155773370058676, + "learning_rate": 1.950926695201953e-07, + "loss": 0.6527, + "step": 2746 + }, + { + "epoch": 5.087037037037037, + "grad_norm": 2.3661705015095427, + "learning_rate": 1.9508802899037334e-07, + "loss": 0.6836, + "step": 2747 + }, + { + "epoch": 5.088888888888889, + "grad_norm": 1.5130410631065652, + "learning_rate": 1.950833863227142e-07, + "loss": 0.6919, + "step": 2748 + }, + { + "epoch": 5.090740740740741, + "grad_norm": 1.6021419486188553, + "learning_rate": 1.9507874151732234e-07, + "loss": 0.7322, + "step": 2749 + }, + { + "epoch": 5.092592592592593, + "grad_norm": 2.0235869546472274, + "learning_rate": 1.950740945743021e-07, + "loss": 0.7066, + "step": 2750 + }, + { + "epoch": 5.094444444444444, + "grad_norm": 1.8970633116807134, + "learning_rate": 1.95069445493758e-07, + "loss": 0.667, + "step": 2751 + }, + { + "epoch": 5.0962962962962965, + "grad_norm": 1.6674380557253778, + "learning_rate": 1.9506479427579454e-07, + "loss": 0.7001, + "step": 2752 + }, + { + "epoch": 5.098148148148148, + "grad_norm": 1.430605359305374, + "learning_rate": 1.9506014092051637e-07, + "loss": 0.682, + "step": 2753 + }, + { + "epoch": 5.1, + "grad_norm": 1.4845448059912651, + "learning_rate": 1.9505548542802802e-07, + "loss": 0.6929, + "step": 2754 + }, + { + "epoch": 5.101851851851852, + "grad_norm": 1.4359422915676057, + "learning_rate": 1.950508277984342e-07, + "loss": 0.7202, + "step": 2755 + }, + { + "epoch": 5.103703703703704, + "grad_norm": 1.5038846098969982, + "learning_rate": 1.950461680318396e-07, + "loss": 0.6521, + "step": 2756 + }, + { + "epoch": 5.105555555555555, + "grad_norm": 1.5236839483000373, + "learning_rate": 1.9504150612834899e-07, + "loss": 0.6454, + "step": 2757 + }, + { + "epoch": 5.107407407407408, + "grad_norm": 1.46801847253954, + "learning_rate": 1.9503684208806722e-07, + "loss": 0.7031, + "step": 2758 + }, + { + "epoch": 5.109259259259259, + "grad_norm": 1.3871955504440538, + "learning_rate": 1.950321759110991e-07, + "loss": 0.7016, + "step": 2759 + }, + { + "epoch": 5.111111111111111, + "grad_norm": 1.4539471871797784, + "learning_rate": 1.950275075975496e-07, + "loss": 0.6648, + "step": 2760 + }, + { + "epoch": 5.112962962962963, + "grad_norm": 1.615297523484419, + "learning_rate": 1.950228371475236e-07, + "loss": 0.6567, + "step": 2761 + }, + { + "epoch": 5.114814814814815, + "grad_norm": 1.474463806413928, + "learning_rate": 1.9501816456112618e-07, + "loss": 0.6704, + "step": 2762 + }, + { + "epoch": 5.116666666666666, + "grad_norm": 2.012067588383928, + "learning_rate": 1.9501348983846232e-07, + "loss": 0.6597, + "step": 2763 + }, + { + "epoch": 5.118518518518519, + "grad_norm": 1.5565270378519152, + "learning_rate": 1.9500881297963718e-07, + "loss": 0.708, + "step": 2764 + }, + { + "epoch": 5.12037037037037, + "grad_norm": 1.7055643886980758, + "learning_rate": 1.9500413398475587e-07, + "loss": 0.6618, + "step": 2765 + }, + { + "epoch": 5.122222222222222, + "grad_norm": 1.5020234475682968, + "learning_rate": 1.949994528539236e-07, + "loss": 0.6551, + "step": 2766 + }, + { + "epoch": 5.124074074074074, + "grad_norm": 1.5178357896780215, + "learning_rate": 1.9499476958724563e-07, + "loss": 0.6813, + "step": 2767 + }, + { + "epoch": 5.125925925925926, + "grad_norm": 1.4922108144510435, + "learning_rate": 1.9499008418482724e-07, + "loss": 0.6298, + "step": 2768 + }, + { + "epoch": 5.127777777777778, + "grad_norm": 1.6501580057463796, + "learning_rate": 1.9498539664677374e-07, + "loss": 0.7192, + "step": 2769 + }, + { + "epoch": 5.12962962962963, + "grad_norm": 1.823136672728746, + "learning_rate": 1.9498070697319057e-07, + "loss": 0.669, + "step": 2770 + }, + { + "epoch": 5.131481481481481, + "grad_norm": 1.6124739557201277, + "learning_rate": 1.9497601516418314e-07, + "loss": 0.6677, + "step": 2771 + }, + { + "epoch": 5.133333333333334, + "grad_norm": 1.6127603035957159, + "learning_rate": 1.9497132121985694e-07, + "loss": 0.6689, + "step": 2772 + }, + { + "epoch": 5.135185185185185, + "grad_norm": 1.459456301463151, + "learning_rate": 1.949666251403175e-07, + "loss": 0.6022, + "step": 2773 + }, + { + "epoch": 5.137037037037037, + "grad_norm": 1.7149872704751348, + "learning_rate": 1.949619269256704e-07, + "loss": 0.6667, + "step": 2774 + }, + { + "epoch": 5.138888888888889, + "grad_norm": 1.4993614782027365, + "learning_rate": 1.9495722657602128e-07, + "loss": 0.6647, + "step": 2775 + }, + { + "epoch": 5.140740740740741, + "grad_norm": 1.460664552811264, + "learning_rate": 1.9495252409147578e-07, + "loss": 0.6561, + "step": 2776 + }, + { + "epoch": 5.142592592592592, + "grad_norm": 1.4331919020484565, + "learning_rate": 1.9494781947213967e-07, + "loss": 0.6601, + "step": 2777 + }, + { + "epoch": 5.144444444444445, + "grad_norm": 1.3402165234324948, + "learning_rate": 1.949431127181187e-07, + "loss": 0.6605, + "step": 2778 + }, + { + "epoch": 5.146296296296296, + "grad_norm": 1.4228384370319394, + "learning_rate": 1.9493840382951866e-07, + "loss": 0.674, + "step": 2779 + }, + { + "epoch": 5.148148148148148, + "grad_norm": 1.4340539119694766, + "learning_rate": 1.9493369280644552e-07, + "loss": 0.7383, + "step": 2780 + }, + { + "epoch": 5.15, + "grad_norm": 1.479137910591616, + "learning_rate": 1.949289796490051e-07, + "loss": 0.7143, + "step": 2781 + }, + { + "epoch": 5.151851851851852, + "grad_norm": 1.376411319707628, + "learning_rate": 1.949242643573034e-07, + "loss": 0.6821, + "step": 2782 + }, + { + "epoch": 5.1537037037037035, + "grad_norm": 1.578345934052031, + "learning_rate": 1.9491954693144642e-07, + "loss": 0.6961, + "step": 2783 + }, + { + "epoch": 5.155555555555556, + "grad_norm": 1.2859267621520243, + "learning_rate": 1.9491482737154018e-07, + "loss": 0.6728, + "step": 2784 + }, + { + "epoch": 5.157407407407407, + "grad_norm": 1.4219097751728964, + "learning_rate": 1.9491010567769092e-07, + "loss": 0.6543, + "step": 2785 + }, + { + "epoch": 5.159259259259259, + "grad_norm": 1.9218915234661444, + "learning_rate": 1.9490538185000465e-07, + "loss": 0.6895, + "step": 2786 + }, + { + "epoch": 5.161111111111111, + "grad_norm": 2.6321554666272706, + "learning_rate": 1.949006558885877e-07, + "loss": 0.6907, + "step": 2787 + }, + { + "epoch": 5.162962962962963, + "grad_norm": 1.6683796734423295, + "learning_rate": 1.948959277935462e-07, + "loss": 0.6757, + "step": 2788 + }, + { + "epoch": 5.1648148148148145, + "grad_norm": 3.164475198559798, + "learning_rate": 1.9489119756498653e-07, + "loss": 0.7005, + "step": 2789 + }, + { + "epoch": 5.166666666666667, + "grad_norm": 1.3947709393826655, + "learning_rate": 1.94886465203015e-07, + "loss": 0.6643, + "step": 2790 + }, + { + "epoch": 5.1685185185185185, + "grad_norm": 1.407047543237564, + "learning_rate": 1.9488173070773805e-07, + "loss": 0.7031, + "step": 2791 + }, + { + "epoch": 5.17037037037037, + "grad_norm": 1.724143373380897, + "learning_rate": 1.948769940792621e-07, + "loss": 0.7153, + "step": 2792 + }, + { + "epoch": 5.1722222222222225, + "grad_norm": 1.6040764749658936, + "learning_rate": 1.9487225531769363e-07, + "loss": 0.6709, + "step": 2793 + }, + { + "epoch": 5.174074074074074, + "grad_norm": 1.4574123916461896, + "learning_rate": 1.948675144231392e-07, + "loss": 0.6957, + "step": 2794 + }, + { + "epoch": 5.175925925925926, + "grad_norm": 2.11707214898699, + "learning_rate": 1.948627713957054e-07, + "loss": 0.6914, + "step": 2795 + }, + { + "epoch": 5.177777777777778, + "grad_norm": 1.293401916482147, + "learning_rate": 1.948580262354988e-07, + "loss": 0.6873, + "step": 2796 + }, + { + "epoch": 5.17962962962963, + "grad_norm": 1.216865363540106, + "learning_rate": 1.948532789426262e-07, + "loss": 0.6683, + "step": 2797 + }, + { + "epoch": 5.181481481481481, + "grad_norm": 1.5937996438551358, + "learning_rate": 1.9484852951719425e-07, + "loss": 0.611, + "step": 2798 + }, + { + "epoch": 5.183333333333334, + "grad_norm": 1.7804555710582373, + "learning_rate": 1.9484377795930976e-07, + "loss": 0.7335, + "step": 2799 + }, + { + "epoch": 5.185185185185185, + "grad_norm": 1.6367260097845768, + "learning_rate": 1.9483902426907952e-07, + "loss": 0.7165, + "step": 2800 + }, + { + "epoch": 5.187037037037037, + "grad_norm": 1.9573853835410369, + "learning_rate": 1.9483426844661049e-07, + "loss": 0.642, + "step": 2801 + }, + { + "epoch": 5.188888888888889, + "grad_norm": 1.2969647223491638, + "learning_rate": 1.9482951049200952e-07, + "loss": 0.7149, + "step": 2802 + }, + { + "epoch": 5.190740740740741, + "grad_norm": 1.4353197914430602, + "learning_rate": 1.9482475040538358e-07, + "loss": 0.6546, + "step": 2803 + }, + { + "epoch": 5.192592592592592, + "grad_norm": 1.3875427752324039, + "learning_rate": 1.9481998818683972e-07, + "loss": 0.6609, + "step": 2804 + }, + { + "epoch": 5.194444444444445, + "grad_norm": 1.766264695603854, + "learning_rate": 1.9481522383648498e-07, + "loss": 0.6881, + "step": 2805 + }, + { + "epoch": 5.196296296296296, + "grad_norm": 1.6005034787564454, + "learning_rate": 1.948104573544265e-07, + "loss": 0.6429, + "step": 2806 + }, + { + "epoch": 5.198148148148148, + "grad_norm": 2.1307872920913824, + "learning_rate": 1.9480568874077146e-07, + "loss": 0.7005, + "step": 2807 + }, + { + "epoch": 5.2, + "grad_norm": 1.5677577449065927, + "learning_rate": 1.9480091799562703e-07, + "loss": 0.6611, + "step": 2808 + }, + { + "epoch": 5.201851851851852, + "grad_norm": 3.864663361425221, + "learning_rate": 1.9479614511910052e-07, + "loss": 0.6512, + "step": 2809 + }, + { + "epoch": 5.203703703703703, + "grad_norm": 1.7471123822403607, + "learning_rate": 1.9479137011129918e-07, + "loss": 0.7004, + "step": 2810 + }, + { + "epoch": 5.205555555555556, + "grad_norm": 1.5362430901262045, + "learning_rate": 1.9478659297233039e-07, + "loss": 0.6252, + "step": 2811 + }, + { + "epoch": 5.207407407407407, + "grad_norm": 1.477713954296486, + "learning_rate": 1.9478181370230153e-07, + "loss": 0.7105, + "step": 2812 + }, + { + "epoch": 5.209259259259259, + "grad_norm": 1.7323259525957793, + "learning_rate": 1.9477703230132012e-07, + "loss": 0.6576, + "step": 2813 + }, + { + "epoch": 5.211111111111111, + "grad_norm": 1.768586985527259, + "learning_rate": 1.9477224876949358e-07, + "loss": 0.6646, + "step": 2814 + }, + { + "epoch": 5.212962962962963, + "grad_norm": 1.5020262897631673, + "learning_rate": 1.9476746310692952e-07, + "loss": 0.7046, + "step": 2815 + }, + { + "epoch": 5.214814814814815, + "grad_norm": 2.3730736753357586, + "learning_rate": 1.9476267531373547e-07, + "loss": 0.6391, + "step": 2816 + }, + { + "epoch": 5.216666666666667, + "grad_norm": 1.7314570247601258, + "learning_rate": 1.9475788539001914e-07, + "loss": 0.6438, + "step": 2817 + }, + { + "epoch": 5.218518518518518, + "grad_norm": 1.484058269346073, + "learning_rate": 1.9475309333588815e-07, + "loss": 0.6714, + "step": 2818 + }, + { + "epoch": 5.220370370370371, + "grad_norm": 1.6954949408772848, + "learning_rate": 1.9474829915145033e-07, + "loss": 0.6579, + "step": 2819 + }, + { + "epoch": 5.222222222222222, + "grad_norm": 1.8071272843783646, + "learning_rate": 1.9474350283681337e-07, + "loss": 0.6646, + "step": 2820 + }, + { + "epoch": 5.224074074074074, + "grad_norm": 1.5677656925350874, + "learning_rate": 1.9473870439208514e-07, + "loss": 0.6783, + "step": 2821 + }, + { + "epoch": 5.225925925925926, + "grad_norm": 1.5609470747266903, + "learning_rate": 1.9473390381737355e-07, + "loss": 0.6711, + "step": 2822 + }, + { + "epoch": 5.227777777777778, + "grad_norm": 1.9426868122481387, + "learning_rate": 1.947291011127865e-07, + "loss": 0.6641, + "step": 2823 + }, + { + "epoch": 5.229629629629629, + "grad_norm": 1.4698191919185866, + "learning_rate": 1.9472429627843202e-07, + "loss": 0.7239, + "step": 2824 + }, + { + "epoch": 5.231481481481482, + "grad_norm": 1.4336401393811464, + "learning_rate": 1.9471948931441803e-07, + "loss": 0.6024, + "step": 2825 + }, + { + "epoch": 5.233333333333333, + "grad_norm": 1.7134130567823629, + "learning_rate": 1.947146802208527e-07, + "loss": 0.6364, + "step": 2826 + }, + { + "epoch": 5.235185185185185, + "grad_norm": 1.8008859986136447, + "learning_rate": 1.947098689978441e-07, + "loss": 0.631, + "step": 2827 + }, + { + "epoch": 5.237037037037037, + "grad_norm": 1.3728328385954804, + "learning_rate": 1.9470505564550045e-07, + "loss": 0.6792, + "step": 2828 + }, + { + "epoch": 5.238888888888889, + "grad_norm": 2.083104488316351, + "learning_rate": 1.947002401639299e-07, + "loss": 0.6897, + "step": 2829 + }, + { + "epoch": 5.2407407407407405, + "grad_norm": 1.7336016394842522, + "learning_rate": 1.9469542255324077e-07, + "loss": 0.6883, + "step": 2830 + }, + { + "epoch": 5.242592592592593, + "grad_norm": 1.4560158522070599, + "learning_rate": 1.9469060281354134e-07, + "loss": 0.6773, + "step": 2831 + }, + { + "epoch": 5.2444444444444445, + "grad_norm": 1.739413226781611, + "learning_rate": 1.9468578094493997e-07, + "loss": 0.6408, + "step": 2832 + }, + { + "epoch": 5.246296296296296, + "grad_norm": 1.5705140806790685, + "learning_rate": 1.946809569475451e-07, + "loss": 0.6874, + "step": 2833 + }, + { + "epoch": 5.2481481481481485, + "grad_norm": 1.7579573129028792, + "learning_rate": 1.9467613082146516e-07, + "loss": 0.6819, + "step": 2834 + }, + { + "epoch": 5.25, + "grad_norm": 1.5817169405323315, + "learning_rate": 1.9467130256680866e-07, + "loss": 0.6829, + "step": 2835 + }, + { + "epoch": 5.2518518518518515, + "grad_norm": 2.244142383353803, + "learning_rate": 1.9466647218368417e-07, + "loss": 0.6884, + "step": 2836 + }, + { + "epoch": 5.253703703703704, + "grad_norm": 1.6363122054091424, + "learning_rate": 1.9466163967220028e-07, + "loss": 0.7068, + "step": 2837 + }, + { + "epoch": 5.2555555555555555, + "grad_norm": 1.6144176275240842, + "learning_rate": 1.946568050324656e-07, + "loss": 0.716, + "step": 2838 + }, + { + "epoch": 5.257407407407407, + "grad_norm": 1.3674793886344967, + "learning_rate": 1.946519682645889e-07, + "loss": 0.6745, + "step": 2839 + }, + { + "epoch": 5.2592592592592595, + "grad_norm": 1.3938275450288142, + "learning_rate": 1.9464712936867885e-07, + "loss": 0.6904, + "step": 2840 + }, + { + "epoch": 5.261111111111111, + "grad_norm": 1.4049026712184929, + "learning_rate": 1.9464228834484428e-07, + "loss": 0.7159, + "step": 2841 + }, + { + "epoch": 5.262962962962963, + "grad_norm": 1.5771338785707232, + "learning_rate": 1.9463744519319402e-07, + "loss": 0.6726, + "step": 2842 + }, + { + "epoch": 5.264814814814815, + "grad_norm": 2.9542976555884404, + "learning_rate": 1.94632599913837e-07, + "loss": 0.6779, + "step": 2843 + }, + { + "epoch": 5.266666666666667, + "grad_norm": 1.3404559966669218, + "learning_rate": 1.9462775250688205e-07, + "loss": 0.6808, + "step": 2844 + }, + { + "epoch": 5.268518518518518, + "grad_norm": 1.3822389079831359, + "learning_rate": 1.9462290297243826e-07, + "loss": 0.715, + "step": 2845 + }, + { + "epoch": 5.270370370370371, + "grad_norm": 1.3824869495723504, + "learning_rate": 1.9461805131061463e-07, + "loss": 0.6578, + "step": 2846 + }, + { + "epoch": 5.272222222222222, + "grad_norm": 1.4505436950571033, + "learning_rate": 1.9461319752152022e-07, + "loss": 0.6757, + "step": 2847 + }, + { + "epoch": 5.274074074074074, + "grad_norm": 1.3475952155856883, + "learning_rate": 1.9460834160526413e-07, + "loss": 0.6874, + "step": 2848 + }, + { + "epoch": 5.275925925925926, + "grad_norm": 1.4689952822308374, + "learning_rate": 1.946034835619556e-07, + "loss": 0.6566, + "step": 2849 + }, + { + "epoch": 5.277777777777778, + "grad_norm": 1.7297581775706106, + "learning_rate": 1.945986233917038e-07, + "loss": 0.6957, + "step": 2850 + }, + { + "epoch": 5.279629629629629, + "grad_norm": 1.659480326424263, + "learning_rate": 1.9459376109461804e-07, + "loss": 0.7062, + "step": 2851 + }, + { + "epoch": 5.281481481481482, + "grad_norm": 1.4682629536535758, + "learning_rate": 1.9458889667080763e-07, + "loss": 0.6386, + "step": 2852 + }, + { + "epoch": 5.283333333333333, + "grad_norm": 1.5439377986448244, + "learning_rate": 1.945840301203819e-07, + "loss": 0.6881, + "step": 2853 + }, + { + "epoch": 5.285185185185185, + "grad_norm": 1.3735687875444373, + "learning_rate": 1.945791614434503e-07, + "loss": 0.6853, + "step": 2854 + }, + { + "epoch": 5.287037037037037, + "grad_norm": 1.5970107063251036, + "learning_rate": 1.9457429064012225e-07, + "loss": 0.6729, + "step": 2855 + }, + { + "epoch": 5.288888888888889, + "grad_norm": 1.5541939396683784, + "learning_rate": 1.9456941771050733e-07, + "loss": 0.7082, + "step": 2856 + }, + { + "epoch": 5.290740740740741, + "grad_norm": 1.4020851123494547, + "learning_rate": 1.9456454265471502e-07, + "loss": 0.6781, + "step": 2857 + }, + { + "epoch": 5.292592592592593, + "grad_norm": 1.6694653739531418, + "learning_rate": 1.9455966547285497e-07, + "loss": 0.6626, + "step": 2858 + }, + { + "epoch": 5.294444444444444, + "grad_norm": 1.577576229774534, + "learning_rate": 1.9455478616503683e-07, + "loss": 0.6745, + "step": 2859 + }, + { + "epoch": 5.296296296296296, + "grad_norm": 1.8847629395762089, + "learning_rate": 1.9454990473137026e-07, + "loss": 0.6914, + "step": 2860 + }, + { + "epoch": 5.298148148148148, + "grad_norm": 1.503833375461945, + "learning_rate": 1.9454502117196508e-07, + "loss": 0.6536, + "step": 2861 + }, + { + "epoch": 5.3, + "grad_norm": 1.358700152789321, + "learning_rate": 1.94540135486931e-07, + "loss": 0.6216, + "step": 2862 + }, + { + "epoch": 5.301851851851852, + "grad_norm": 1.4050809930502495, + "learning_rate": 1.9453524767637792e-07, + "loss": 0.6893, + "step": 2863 + }, + { + "epoch": 5.303703703703704, + "grad_norm": 1.8220424979567904, + "learning_rate": 1.9453035774041573e-07, + "loss": 0.6506, + "step": 2864 + }, + { + "epoch": 5.305555555555555, + "grad_norm": 1.401604670019065, + "learning_rate": 1.9452546567915435e-07, + "loss": 0.6716, + "step": 2865 + }, + { + "epoch": 5.307407407407408, + "grad_norm": 1.6118042619586435, + "learning_rate": 1.9452057149270377e-07, + "loss": 0.6495, + "step": 2866 + }, + { + "epoch": 5.309259259259259, + "grad_norm": 1.5542989499933388, + "learning_rate": 1.9451567518117405e-07, + "loss": 0.6912, + "step": 2867 + }, + { + "epoch": 5.311111111111111, + "grad_norm": 1.5381582135417735, + "learning_rate": 1.9451077674467523e-07, + "loss": 0.6593, + "step": 2868 + }, + { + "epoch": 5.312962962962963, + "grad_norm": 1.5042316006003649, + "learning_rate": 1.9450587618331747e-07, + "loss": 0.6998, + "step": 2869 + }, + { + "epoch": 5.314814814814815, + "grad_norm": 1.5120526154181062, + "learning_rate": 1.9450097349721094e-07, + "loss": 0.6759, + "step": 2870 + }, + { + "epoch": 5.316666666666666, + "grad_norm": 1.51614906203917, + "learning_rate": 1.9449606868646587e-07, + "loss": 0.7016, + "step": 2871 + }, + { + "epoch": 5.318518518518519, + "grad_norm": 1.428776845741125, + "learning_rate": 1.944911617511925e-07, + "loss": 0.6558, + "step": 2872 + }, + { + "epoch": 5.32037037037037, + "grad_norm": 1.6854191845790605, + "learning_rate": 1.944862526915012e-07, + "loss": 0.6837, + "step": 2873 + }, + { + "epoch": 5.322222222222222, + "grad_norm": 1.7341208906467187, + "learning_rate": 1.9448134150750233e-07, + "loss": 0.6367, + "step": 2874 + }, + { + "epoch": 5.324074074074074, + "grad_norm": 1.5528532968680224, + "learning_rate": 1.944764281993063e-07, + "loss": 0.6569, + "step": 2875 + }, + { + "epoch": 5.325925925925926, + "grad_norm": 1.3061477004565407, + "learning_rate": 1.9447151276702354e-07, + "loss": 0.6999, + "step": 2876 + }, + { + "epoch": 5.3277777777777775, + "grad_norm": 1.4335630851649692, + "learning_rate": 1.9446659521076462e-07, + "loss": 0.7029, + "step": 2877 + }, + { + "epoch": 5.32962962962963, + "grad_norm": 2.954479068314854, + "learning_rate": 1.9446167553064004e-07, + "loss": 0.6866, + "step": 2878 + }, + { + "epoch": 5.3314814814814815, + "grad_norm": 1.5355574747257594, + "learning_rate": 1.9445675372676047e-07, + "loss": 0.6518, + "step": 2879 + }, + { + "epoch": 5.333333333333333, + "grad_norm": 1.7808925484710638, + "learning_rate": 1.9445182979923653e-07, + "loss": 0.667, + "step": 2880 + }, + { + "epoch": 5.3351851851851855, + "grad_norm": 1.5361518503757934, + "learning_rate": 1.9444690374817892e-07, + "loss": 0.6729, + "step": 2881 + }, + { + "epoch": 5.337037037037037, + "grad_norm": 1.5419605479996106, + "learning_rate": 1.9444197557369838e-07, + "loss": 0.6257, + "step": 2882 + }, + { + "epoch": 5.338888888888889, + "grad_norm": 1.9984880547190327, + "learning_rate": 1.9443704527590578e-07, + "loss": 0.682, + "step": 2883 + }, + { + "epoch": 5.340740740740741, + "grad_norm": 1.4725431370509865, + "learning_rate": 1.9443211285491188e-07, + "loss": 0.6995, + "step": 2884 + }, + { + "epoch": 5.342592592592593, + "grad_norm": 1.8200180770753691, + "learning_rate": 1.944271783108276e-07, + "loss": 0.7545, + "step": 2885 + }, + { + "epoch": 5.344444444444444, + "grad_norm": 1.8547336472576708, + "learning_rate": 1.9442224164376393e-07, + "loss": 0.6793, + "step": 2886 + }, + { + "epoch": 5.3462962962962965, + "grad_norm": 1.6449937348992367, + "learning_rate": 1.9441730285383179e-07, + "loss": 0.6754, + "step": 2887 + }, + { + "epoch": 5.348148148148148, + "grad_norm": 1.476239413484283, + "learning_rate": 1.9441236194114225e-07, + "loss": 0.686, + "step": 2888 + }, + { + "epoch": 5.35, + "grad_norm": 1.564057599915203, + "learning_rate": 1.944074189058064e-07, + "loss": 0.7177, + "step": 2889 + }, + { + "epoch": 5.351851851851852, + "grad_norm": 1.6232443065500868, + "learning_rate": 1.9440247374793538e-07, + "loss": 0.6862, + "step": 2890 + }, + { + "epoch": 5.353703703703704, + "grad_norm": 1.6742158305027384, + "learning_rate": 1.9439752646764034e-07, + "loss": 0.6966, + "step": 2891 + }, + { + "epoch": 5.355555555555555, + "grad_norm": 1.3510746479521587, + "learning_rate": 1.9439257706503253e-07, + "loss": 0.6462, + "step": 2892 + }, + { + "epoch": 5.357407407407408, + "grad_norm": 1.483323814899246, + "learning_rate": 1.943876255402232e-07, + "loss": 0.7131, + "step": 2893 + }, + { + "epoch": 5.359259259259259, + "grad_norm": 1.9853380402379108, + "learning_rate": 1.9438267189332375e-07, + "loss": 0.6562, + "step": 2894 + }, + { + "epoch": 5.361111111111111, + "grad_norm": 1.5641062492327031, + "learning_rate": 1.9437771612444544e-07, + "loss": 0.6914, + "step": 2895 + }, + { + "epoch": 5.362962962962963, + "grad_norm": 1.5793841290274553, + "learning_rate": 1.9437275823369976e-07, + "loss": 0.7162, + "step": 2896 + }, + { + "epoch": 5.364814814814815, + "grad_norm": 5.693226334836632, + "learning_rate": 1.943677982211982e-07, + "loss": 0.6706, + "step": 2897 + }, + { + "epoch": 5.366666666666666, + "grad_norm": 1.7172205682507975, + "learning_rate": 1.943628360870522e-07, + "loss": 0.6726, + "step": 2898 + }, + { + "epoch": 5.368518518518519, + "grad_norm": 1.7366661289176768, + "learning_rate": 1.9435787183137335e-07, + "loss": 0.6228, + "step": 2899 + }, + { + "epoch": 5.37037037037037, + "grad_norm": 1.4233524879901576, + "learning_rate": 1.9435290545427327e-07, + "loss": 0.6165, + "step": 2900 + }, + { + "epoch": 5.372222222222222, + "grad_norm": 1.586204624208911, + "learning_rate": 1.9434793695586365e-07, + "loss": 0.6723, + "step": 2901 + }, + { + "epoch": 5.374074074074074, + "grad_norm": 2.100788991398898, + "learning_rate": 1.9434296633625614e-07, + "loss": 0.6123, + "step": 2902 + }, + { + "epoch": 5.375925925925926, + "grad_norm": 1.707196280231772, + "learning_rate": 1.943379935955625e-07, + "loss": 0.7042, + "step": 2903 + }, + { + "epoch": 5.377777777777778, + "grad_norm": 1.6607282413862121, + "learning_rate": 1.943330187338946e-07, + "loss": 0.6789, + "step": 2904 + }, + { + "epoch": 5.37962962962963, + "grad_norm": 1.7423264171455726, + "learning_rate": 1.9432804175136418e-07, + "loss": 0.6171, + "step": 2905 + }, + { + "epoch": 5.381481481481481, + "grad_norm": 1.460691236447556, + "learning_rate": 1.9432306264808322e-07, + "loss": 0.6685, + "step": 2906 + }, + { + "epoch": 5.383333333333334, + "grad_norm": 1.6251796806998862, + "learning_rate": 1.9431808142416362e-07, + "loss": 0.6527, + "step": 2907 + }, + { + "epoch": 5.385185185185185, + "grad_norm": 1.5083887460518863, + "learning_rate": 1.9431309807971738e-07, + "loss": 0.6948, + "step": 2908 + }, + { + "epoch": 5.387037037037037, + "grad_norm": 1.4483643763724703, + "learning_rate": 1.9430811261485656e-07, + "loss": 0.6867, + "step": 2909 + }, + { + "epoch": 5.388888888888889, + "grad_norm": 1.825427469411476, + "learning_rate": 1.9430312502969324e-07, + "loss": 0.6893, + "step": 2910 + }, + { + "epoch": 5.390740740740741, + "grad_norm": 1.4174216598914398, + "learning_rate": 1.9429813532433953e-07, + "loss": 0.6768, + "step": 2911 + }, + { + "epoch": 5.392592592592592, + "grad_norm": 1.6698229745333968, + "learning_rate": 1.9429314349890761e-07, + "loss": 0.67, + "step": 2912 + }, + { + "epoch": 5.394444444444445, + "grad_norm": 1.4357508555177596, + "learning_rate": 1.9428814955350976e-07, + "loss": 0.6722, + "step": 2913 + }, + { + "epoch": 5.396296296296296, + "grad_norm": 1.5923082468288956, + "learning_rate": 1.942831534882582e-07, + "loss": 0.6738, + "step": 2914 + }, + { + "epoch": 5.398148148148148, + "grad_norm": 1.6860674908288755, + "learning_rate": 1.942781553032653e-07, + "loss": 0.6478, + "step": 2915 + }, + { + "epoch": 5.4, + "grad_norm": 1.4974631615691536, + "learning_rate": 1.942731549986434e-07, + "loss": 0.7059, + "step": 2916 + }, + { + "epoch": 5.401851851851852, + "grad_norm": 2.2473767368622055, + "learning_rate": 1.9426815257450494e-07, + "loss": 0.6999, + "step": 2917 + }, + { + "epoch": 5.4037037037037035, + "grad_norm": 2.5227875253793886, + "learning_rate": 1.942631480309624e-07, + "loss": 0.6515, + "step": 2918 + }, + { + "epoch": 5.405555555555556, + "grad_norm": 1.6563997976461724, + "learning_rate": 1.9425814136812825e-07, + "loss": 0.6917, + "step": 2919 + }, + { + "epoch": 5.407407407407407, + "grad_norm": 1.597575290589142, + "learning_rate": 1.9425313258611508e-07, + "loss": 0.6379, + "step": 2920 + }, + { + "epoch": 5.409259259259259, + "grad_norm": 1.3576372341251943, + "learning_rate": 1.9424812168503554e-07, + "loss": 0.6315, + "step": 2921 + }, + { + "epoch": 5.411111111111111, + "grad_norm": 1.5506275926302855, + "learning_rate": 1.9424310866500223e-07, + "loss": 0.6518, + "step": 2922 + }, + { + "epoch": 5.412962962962963, + "grad_norm": 1.4749643270065134, + "learning_rate": 1.9423809352612786e-07, + "loss": 0.6838, + "step": 2923 + }, + { + "epoch": 5.4148148148148145, + "grad_norm": 2.6489837459389065, + "learning_rate": 1.942330762685252e-07, + "loss": 0.6824, + "step": 2924 + }, + { + "epoch": 5.416666666666667, + "grad_norm": 1.315890563185257, + "learning_rate": 1.9422805689230705e-07, + "loss": 0.7033, + "step": 2925 + }, + { + "epoch": 5.4185185185185185, + "grad_norm": 1.395458411256235, + "learning_rate": 1.9422303539758625e-07, + "loss": 0.6984, + "step": 2926 + }, + { + "epoch": 5.42037037037037, + "grad_norm": 1.5319872281239346, + "learning_rate": 1.9421801178447574e-07, + "loss": 0.6623, + "step": 2927 + }, + { + "epoch": 5.4222222222222225, + "grad_norm": 1.3846258145233896, + "learning_rate": 1.9421298605308845e-07, + "loss": 0.6561, + "step": 2928 + }, + { + "epoch": 5.424074074074074, + "grad_norm": 1.4989289725082253, + "learning_rate": 1.9420795820353733e-07, + "loss": 0.6482, + "step": 2929 + }, + { + "epoch": 5.425925925925926, + "grad_norm": 1.393535443972757, + "learning_rate": 1.9420292823593543e-07, + "loss": 0.718, + "step": 2930 + }, + { + "epoch": 5.427777777777778, + "grad_norm": 1.3144700050476068, + "learning_rate": 1.9419789615039588e-07, + "loss": 0.6396, + "step": 2931 + }, + { + "epoch": 5.42962962962963, + "grad_norm": 1.3998071811253625, + "learning_rate": 1.9419286194703177e-07, + "loss": 0.6484, + "step": 2932 + }, + { + "epoch": 5.431481481481481, + "grad_norm": 1.5983662613344372, + "learning_rate": 1.941878256259563e-07, + "loss": 0.7362, + "step": 2933 + }, + { + "epoch": 5.433333333333334, + "grad_norm": 1.4044020828608224, + "learning_rate": 1.9418278718728272e-07, + "loss": 0.671, + "step": 2934 + }, + { + "epoch": 5.435185185185185, + "grad_norm": 1.4399794269790391, + "learning_rate": 1.9417774663112426e-07, + "loss": 0.5834, + "step": 2935 + }, + { + "epoch": 5.437037037037037, + "grad_norm": 1.7459385014787752, + "learning_rate": 1.9417270395759426e-07, + "loss": 0.6903, + "step": 2936 + }, + { + "epoch": 5.438888888888889, + "grad_norm": 1.5243699475353711, + "learning_rate": 1.9416765916680615e-07, + "loss": 0.6505, + "step": 2937 + }, + { + "epoch": 5.440740740740741, + "grad_norm": 1.3793364691779872, + "learning_rate": 1.9416261225887326e-07, + "loss": 0.6662, + "step": 2938 + }, + { + "epoch": 5.442592592592592, + "grad_norm": 2.0953323288135444, + "learning_rate": 1.9415756323390914e-07, + "loss": 0.6106, + "step": 2939 + }, + { + "epoch": 5.444444444444445, + "grad_norm": 1.4290060655734755, + "learning_rate": 1.9415251209202728e-07, + "loss": 0.6317, + "step": 2940 + }, + { + "epoch": 5.446296296296296, + "grad_norm": 1.4455754787910795, + "learning_rate": 1.941474588333412e-07, + "loss": 0.6952, + "step": 2941 + }, + { + "epoch": 5.448148148148148, + "grad_norm": 1.4285820996856942, + "learning_rate": 1.941424034579646e-07, + "loss": 0.6396, + "step": 2942 + }, + { + "epoch": 5.45, + "grad_norm": 1.5889285596447713, + "learning_rate": 1.94137345966011e-07, + "loss": 0.6912, + "step": 2943 + }, + { + "epoch": 5.451851851851852, + "grad_norm": 1.4167829590991743, + "learning_rate": 1.9413228635759424e-07, + "loss": 0.6863, + "step": 2944 + }, + { + "epoch": 5.453703703703704, + "grad_norm": 1.3428993035324839, + "learning_rate": 1.94127224632828e-07, + "loss": 0.6498, + "step": 2945 + }, + { + "epoch": 5.455555555555556, + "grad_norm": 1.3497550044754816, + "learning_rate": 1.9412216079182614e-07, + "loss": 0.6878, + "step": 2946 + }, + { + "epoch": 5.457407407407407, + "grad_norm": 1.3612805883911834, + "learning_rate": 1.9411709483470244e-07, + "loss": 0.6614, + "step": 2947 + }, + { + "epoch": 5.459259259259259, + "grad_norm": 1.7452345996108638, + "learning_rate": 1.9411202676157087e-07, + "loss": 0.7028, + "step": 2948 + }, + { + "epoch": 5.461111111111111, + "grad_norm": 1.3947965069952588, + "learning_rate": 1.9410695657254528e-07, + "loss": 0.6493, + "step": 2949 + }, + { + "epoch": 5.462962962962963, + "grad_norm": 1.3409229335695583, + "learning_rate": 1.9410188426773976e-07, + "loss": 0.627, + "step": 2950 + }, + { + "epoch": 5.464814814814815, + "grad_norm": 1.3318809602071386, + "learning_rate": 1.9409680984726828e-07, + "loss": 0.6494, + "step": 2951 + }, + { + "epoch": 5.466666666666667, + "grad_norm": 1.8300137133764702, + "learning_rate": 1.9409173331124498e-07, + "loss": 0.6994, + "step": 2952 + }, + { + "epoch": 5.468518518518518, + "grad_norm": 1.500389161582396, + "learning_rate": 1.940866546597839e-07, + "loss": 0.7004, + "step": 2953 + }, + { + "epoch": 5.470370370370371, + "grad_norm": 3.6801809309329516, + "learning_rate": 1.9408157389299938e-07, + "loss": 0.6904, + "step": 2954 + }, + { + "epoch": 5.472222222222222, + "grad_norm": 1.5258044809238842, + "learning_rate": 1.940764910110055e-07, + "loss": 0.6866, + "step": 2955 + }, + { + "epoch": 5.474074074074074, + "grad_norm": 1.902625106246917, + "learning_rate": 1.9407140601391657e-07, + "loss": 0.7115, + "step": 2956 + }, + { + "epoch": 5.475925925925926, + "grad_norm": 1.5513732363924757, + "learning_rate": 1.94066318901847e-07, + "loss": 0.684, + "step": 2957 + }, + { + "epoch": 5.477777777777778, + "grad_norm": 2.095748644244585, + "learning_rate": 1.9406122967491108e-07, + "loss": 0.6706, + "step": 2958 + }, + { + "epoch": 5.479629629629629, + "grad_norm": 1.9262204957794142, + "learning_rate": 1.9405613833322324e-07, + "loss": 0.6594, + "step": 2959 + }, + { + "epoch": 5.481481481481482, + "grad_norm": 1.444565092012416, + "learning_rate": 1.9405104487689797e-07, + "loss": 0.6749, + "step": 2960 + }, + { + "epoch": 5.483333333333333, + "grad_norm": 1.5424753216598721, + "learning_rate": 1.9404594930604975e-07, + "loss": 0.6772, + "step": 2961 + }, + { + "epoch": 5.485185185185185, + "grad_norm": 1.5047592108047783, + "learning_rate": 1.9404085162079317e-07, + "loss": 0.6924, + "step": 2962 + }, + { + "epoch": 5.487037037037037, + "grad_norm": 1.4706940258969183, + "learning_rate": 1.9403575182124286e-07, + "loss": 0.68, + "step": 2963 + }, + { + "epoch": 5.488888888888889, + "grad_norm": 1.6688407574218445, + "learning_rate": 1.9403064990751343e-07, + "loss": 0.668, + "step": 2964 + }, + { + "epoch": 5.4907407407407405, + "grad_norm": 1.449326482274683, + "learning_rate": 1.9402554587971961e-07, + "loss": 0.6346, + "step": 2965 + }, + { + "epoch": 5.492592592592593, + "grad_norm": 1.4919651190182803, + "learning_rate": 1.9402043973797615e-07, + "loss": 0.64, + "step": 2966 + }, + { + "epoch": 5.4944444444444445, + "grad_norm": 1.5165074081158856, + "learning_rate": 1.9401533148239785e-07, + "loss": 0.6545, + "step": 2967 + }, + { + "epoch": 5.496296296296296, + "grad_norm": 1.7157365599221481, + "learning_rate": 1.9401022111309958e-07, + "loss": 0.6837, + "step": 2968 + }, + { + "epoch": 5.4981481481481485, + "grad_norm": 1.8374082642507168, + "learning_rate": 1.9400510863019616e-07, + "loss": 0.6841, + "step": 2969 + }, + { + "epoch": 5.5, + "grad_norm": 1.7782057848141277, + "learning_rate": 1.9399999403380262e-07, + "loss": 0.6537, + "step": 2970 + }, + { + "epoch": 5.5018518518518515, + "grad_norm": 1.6193677417046048, + "learning_rate": 1.939948773240339e-07, + "loss": 0.6722, + "step": 2971 + }, + { + "epoch": 5.503703703703704, + "grad_norm": 1.839276451566974, + "learning_rate": 1.9398975850100509e-07, + "loss": 0.6238, + "step": 2972 + }, + { + "epoch": 5.5055555555555555, + "grad_norm": 1.8725145797813199, + "learning_rate": 1.939846375648312e-07, + "loss": 0.7245, + "step": 2973 + }, + { + "epoch": 5.507407407407407, + "grad_norm": 1.3441020382176563, + "learning_rate": 1.9397951451562737e-07, + "loss": 0.6669, + "step": 2974 + }, + { + "epoch": 5.5092592592592595, + "grad_norm": 1.8091126481893145, + "learning_rate": 1.9397438935350885e-07, + "loss": 0.667, + "step": 2975 + }, + { + "epoch": 5.511111111111111, + "grad_norm": 1.910275571247403, + "learning_rate": 1.9396926207859085e-07, + "loss": 0.6747, + "step": 2976 + }, + { + "epoch": 5.512962962962963, + "grad_norm": 1.9830925759769786, + "learning_rate": 1.9396413269098857e-07, + "loss": 0.6754, + "step": 2977 + }, + { + "epoch": 5.514814814814815, + "grad_norm": 1.606592644759462, + "learning_rate": 1.939590011908174e-07, + "loss": 0.6865, + "step": 2978 + }, + { + "epoch": 5.516666666666667, + "grad_norm": 1.3742644681080447, + "learning_rate": 1.939538675781927e-07, + "loss": 0.6367, + "step": 2979 + }, + { + "epoch": 5.518518518518518, + "grad_norm": 1.8783109195767198, + "learning_rate": 1.9394873185322988e-07, + "loss": 0.6959, + "step": 2980 + }, + { + "epoch": 5.520370370370371, + "grad_norm": 1.5921200834309328, + "learning_rate": 1.939435940160444e-07, + "loss": 0.6891, + "step": 2981 + }, + { + "epoch": 5.522222222222222, + "grad_norm": 1.612228922987339, + "learning_rate": 1.939384540667518e-07, + "loss": 0.6371, + "step": 2982 + }, + { + "epoch": 5.524074074074074, + "grad_norm": 2.7696971499542316, + "learning_rate": 1.9393331200546764e-07, + "loss": 0.7107, + "step": 2983 + }, + { + "epoch": 5.525925925925926, + "grad_norm": 1.5503760663612347, + "learning_rate": 1.9392816783230745e-07, + "loss": 0.6602, + "step": 2984 + }, + { + "epoch": 5.527777777777778, + "grad_norm": 1.808780330076644, + "learning_rate": 1.9392302154738695e-07, + "loss": 0.7219, + "step": 2985 + }, + { + "epoch": 5.52962962962963, + "grad_norm": 2.316889557786525, + "learning_rate": 1.9391787315082188e-07, + "loss": 0.6877, + "step": 2986 + }, + { + "epoch": 5.531481481481482, + "grad_norm": 1.6347250733106344, + "learning_rate": 1.939127226427279e-07, + "loss": 0.6808, + "step": 2987 + }, + { + "epoch": 5.533333333333333, + "grad_norm": 1.4205303371170537, + "learning_rate": 1.939075700232209e-07, + "loss": 0.6629, + "step": 2988 + }, + { + "epoch": 5.535185185185185, + "grad_norm": 1.5062129418245878, + "learning_rate": 1.9390241529241664e-07, + "loss": 0.6631, + "step": 2989 + }, + { + "epoch": 5.537037037037037, + "grad_norm": 1.31335146159003, + "learning_rate": 1.9389725845043104e-07, + "loss": 0.6749, + "step": 2990 + }, + { + "epoch": 5.538888888888889, + "grad_norm": 1.8628955335828814, + "learning_rate": 1.9389209949738006e-07, + "loss": 0.6761, + "step": 2991 + }, + { + "epoch": 5.540740740740741, + "grad_norm": 1.669216218729797, + "learning_rate": 1.9388693843337968e-07, + "loss": 0.6742, + "step": 2992 + }, + { + "epoch": 5.542592592592593, + "grad_norm": 1.3340010398856033, + "learning_rate": 1.9388177525854595e-07, + "loss": 0.6662, + "step": 2993 + }, + { + "epoch": 5.544444444444444, + "grad_norm": 1.4617330410696125, + "learning_rate": 1.938766099729949e-07, + "loss": 0.6174, + "step": 2994 + }, + { + "epoch": 5.546296296296296, + "grad_norm": 1.313019414908476, + "learning_rate": 1.9387144257684272e-07, + "loss": 0.6588, + "step": 2995 + }, + { + "epoch": 5.548148148148148, + "grad_norm": 1.613916167300068, + "learning_rate": 1.9386627307020554e-07, + "loss": 0.645, + "step": 2996 + }, + { + "epoch": 5.55, + "grad_norm": 2.122689499817843, + "learning_rate": 1.938611014531996e-07, + "loss": 0.6673, + "step": 2997 + }, + { + "epoch": 5.551851851851852, + "grad_norm": 1.641597250070579, + "learning_rate": 1.9385592772594118e-07, + "loss": 0.6734, + "step": 2998 + }, + { + "epoch": 5.553703703703704, + "grad_norm": 1.465225389793985, + "learning_rate": 1.938507518885466e-07, + "loss": 0.7056, + "step": 2999 + }, + { + "epoch": 5.555555555555555, + "grad_norm": 1.3560676765894122, + "learning_rate": 1.9384557394113226e-07, + "loss": 0.6356, + "step": 3000 + }, + { + "epoch": 5.557407407407408, + "grad_norm": 1.5017648465486404, + "learning_rate": 1.938403938838145e-07, + "loss": 0.6855, + "step": 3001 + }, + { + "epoch": 5.559259259259259, + "grad_norm": 1.5082982066661013, + "learning_rate": 1.9383521171670982e-07, + "loss": 0.6598, + "step": 3002 + }, + { + "epoch": 5.561111111111111, + "grad_norm": 1.590000135989519, + "learning_rate": 1.9383002743993473e-07, + "loss": 0.6623, + "step": 3003 + }, + { + "epoch": 5.562962962962963, + "grad_norm": 1.477823356411927, + "learning_rate": 1.9382484105360576e-07, + "loss": 0.6555, + "step": 3004 + }, + { + "epoch": 5.564814814814815, + "grad_norm": 1.5761015836830574, + "learning_rate": 1.9381965255783957e-07, + "loss": 0.6403, + "step": 3005 + }, + { + "epoch": 5.566666666666666, + "grad_norm": 1.4465205071979121, + "learning_rate": 1.9381446195275277e-07, + "loss": 0.6442, + "step": 3006 + }, + { + "epoch": 5.568518518518519, + "grad_norm": 1.5656541107552768, + "learning_rate": 1.9380926923846207e-07, + "loss": 0.6626, + "step": 3007 + }, + { + "epoch": 5.57037037037037, + "grad_norm": 1.8849170633009384, + "learning_rate": 1.9380407441508424e-07, + "loss": 0.6529, + "step": 3008 + }, + { + "epoch": 5.572222222222222, + "grad_norm": 1.566007676480427, + "learning_rate": 1.9379887748273602e-07, + "loss": 0.6427, + "step": 3009 + }, + { + "epoch": 5.574074074074074, + "grad_norm": 1.5031049804944971, + "learning_rate": 1.9379367844153429e-07, + "loss": 0.744, + "step": 3010 + }, + { + "epoch": 5.575925925925926, + "grad_norm": 1.607387330475157, + "learning_rate": 1.9378847729159593e-07, + "loss": 0.6593, + "step": 3011 + }, + { + "epoch": 5.5777777777777775, + "grad_norm": 2.6693408806417747, + "learning_rate": 1.9378327403303788e-07, + "loss": 0.6268, + "step": 3012 + }, + { + "epoch": 5.57962962962963, + "grad_norm": 1.6469348887414648, + "learning_rate": 1.937780686659771e-07, + "loss": 0.669, + "step": 3013 + }, + { + "epoch": 5.5814814814814815, + "grad_norm": 1.8577305173064904, + "learning_rate": 1.9377286119053066e-07, + "loss": 0.6673, + "step": 3014 + }, + { + "epoch": 5.583333333333333, + "grad_norm": 2.3770772668479903, + "learning_rate": 1.937676516068156e-07, + "loss": 0.6682, + "step": 3015 + }, + { + "epoch": 5.5851851851851855, + "grad_norm": 1.386629585628963, + "learning_rate": 1.9376243991494907e-07, + "loss": 0.6466, + "step": 3016 + }, + { + "epoch": 5.587037037037037, + "grad_norm": 1.5880745414668402, + "learning_rate": 1.9375722611504823e-07, + "loss": 0.6361, + "step": 3017 + }, + { + "epoch": 5.588888888888889, + "grad_norm": 1.75952976542882, + "learning_rate": 1.937520102072303e-07, + "loss": 0.6849, + "step": 3018 + }, + { + "epoch": 5.590740740740741, + "grad_norm": 1.5641833439271742, + "learning_rate": 1.9374679219161257e-07, + "loss": 0.687, + "step": 3019 + }, + { + "epoch": 5.592592592592593, + "grad_norm": 2.111528469463874, + "learning_rate": 1.9374157206831233e-07, + "loss": 0.7033, + "step": 3020 + }, + { + "epoch": 5.594444444444444, + "grad_norm": 1.7181832468719178, + "learning_rate": 1.9373634983744695e-07, + "loss": 0.6586, + "step": 3021 + }, + { + "epoch": 5.5962962962962965, + "grad_norm": 2.360895119629398, + "learning_rate": 1.937311254991339e-07, + "loss": 0.6948, + "step": 3022 + }, + { + "epoch": 5.598148148148148, + "grad_norm": 1.5390042183476398, + "learning_rate": 1.937258990534905e-07, + "loss": 0.6548, + "step": 3023 + }, + { + "epoch": 5.6, + "grad_norm": 1.514027297406205, + "learning_rate": 1.9372067050063436e-07, + "loss": 0.6698, + "step": 3024 + }, + { + "epoch": 5.601851851851852, + "grad_norm": 1.402929899519556, + "learning_rate": 1.9371543984068302e-07, + "loss": 0.6732, + "step": 3025 + }, + { + "epoch": 5.603703703703704, + "grad_norm": 1.7746518325099097, + "learning_rate": 1.9371020707375403e-07, + "loss": 0.6702, + "step": 3026 + }, + { + "epoch": 5.605555555555555, + "grad_norm": 1.5342597824255881, + "learning_rate": 1.9370497219996508e-07, + "loss": 0.6569, + "step": 3027 + }, + { + "epoch": 5.607407407407408, + "grad_norm": 1.465997532523462, + "learning_rate": 1.9369973521943388e-07, + "loss": 0.6735, + "step": 3028 + }, + { + "epoch": 5.609259259259259, + "grad_norm": 2.206975163870859, + "learning_rate": 1.9369449613227813e-07, + "loss": 0.6644, + "step": 3029 + }, + { + "epoch": 5.611111111111111, + "grad_norm": 1.5014885917314045, + "learning_rate": 1.9368925493861566e-07, + "loss": 0.6706, + "step": 3030 + }, + { + "epoch": 5.612962962962963, + "grad_norm": 1.8209765441989565, + "learning_rate": 1.9368401163856426e-07, + "loss": 0.6654, + "step": 3031 + }, + { + "epoch": 5.614814814814815, + "grad_norm": 1.491323596112947, + "learning_rate": 1.9367876623224184e-07, + "loss": 0.6973, + "step": 3032 + }, + { + "epoch": 5.616666666666667, + "grad_norm": 1.9744024188488531, + "learning_rate": 1.9367351871976632e-07, + "loss": 0.7124, + "step": 3033 + }, + { + "epoch": 5.618518518518519, + "grad_norm": 1.6289982305355026, + "learning_rate": 1.936682691012557e-07, + "loss": 0.6658, + "step": 3034 + }, + { + "epoch": 5.62037037037037, + "grad_norm": 1.6225262331838626, + "learning_rate": 1.93663017376828e-07, + "loss": 0.6636, + "step": 3035 + }, + { + "epoch": 5.622222222222222, + "grad_norm": 1.8898997313628014, + "learning_rate": 1.9365776354660125e-07, + "loss": 0.6458, + "step": 3036 + }, + { + "epoch": 5.624074074074074, + "grad_norm": 1.9593879943134564, + "learning_rate": 1.9365250761069363e-07, + "loss": 0.6788, + "step": 3037 + }, + { + "epoch": 5.625925925925926, + "grad_norm": 5.840143770684436, + "learning_rate": 1.9364724956922328e-07, + "loss": 0.6522, + "step": 3038 + }, + { + "epoch": 5.627777777777778, + "grad_norm": 2.0273225284926957, + "learning_rate": 1.9364198942230844e-07, + "loss": 0.683, + "step": 3039 + }, + { + "epoch": 5.62962962962963, + "grad_norm": 2.083897234532407, + "learning_rate": 1.9363672717006732e-07, + "loss": 0.6732, + "step": 3040 + }, + { + "epoch": 5.631481481481481, + "grad_norm": 1.8030036642060996, + "learning_rate": 1.9363146281261827e-07, + "loss": 0.6749, + "step": 3041 + }, + { + "epoch": 5.633333333333333, + "grad_norm": 1.5549134621890215, + "learning_rate": 1.9362619635007963e-07, + "loss": 0.7169, + "step": 3042 + }, + { + "epoch": 5.635185185185185, + "grad_norm": 1.6499408971479643, + "learning_rate": 1.9362092778256983e-07, + "loss": 0.7181, + "step": 3043 + }, + { + "epoch": 5.637037037037037, + "grad_norm": 1.6541788016346053, + "learning_rate": 1.9361565711020728e-07, + "loss": 0.6681, + "step": 3044 + }, + { + "epoch": 5.638888888888889, + "grad_norm": 2.5772332485467793, + "learning_rate": 1.936103843331105e-07, + "loss": 0.6896, + "step": 3045 + }, + { + "epoch": 5.640740740740741, + "grad_norm": 1.862076521665792, + "learning_rate": 1.9360510945139806e-07, + "loss": 0.6752, + "step": 3046 + }, + { + "epoch": 5.642592592592592, + "grad_norm": 1.7795509222981707, + "learning_rate": 1.935998324651885e-07, + "loss": 0.7239, + "step": 3047 + }, + { + "epoch": 5.644444444444445, + "grad_norm": 1.6288779946017031, + "learning_rate": 1.935945533746005e-07, + "loss": 0.6827, + "step": 3048 + }, + { + "epoch": 5.646296296296296, + "grad_norm": 1.7353777879347545, + "learning_rate": 1.9358927217975276e-07, + "loss": 0.6292, + "step": 3049 + }, + { + "epoch": 5.648148148148148, + "grad_norm": 1.8564949579085719, + "learning_rate": 1.93583988880764e-07, + "loss": 0.6373, + "step": 3050 + }, + { + "epoch": 5.65, + "grad_norm": 1.939407507598791, + "learning_rate": 1.9357870347775297e-07, + "loss": 0.6486, + "step": 3051 + }, + { + "epoch": 5.651851851851852, + "grad_norm": 1.9194230505513739, + "learning_rate": 1.9357341597083857e-07, + "loss": 0.6855, + "step": 3052 + }, + { + "epoch": 5.6537037037037035, + "grad_norm": 1.94071649557587, + "learning_rate": 1.9356812636013963e-07, + "loss": 0.692, + "step": 3053 + }, + { + "epoch": 5.655555555555556, + "grad_norm": 1.6718659366554096, + "learning_rate": 1.9356283464577506e-07, + "loss": 0.6662, + "step": 3054 + }, + { + "epoch": 5.657407407407407, + "grad_norm": 2.0050321572741545, + "learning_rate": 1.9355754082786384e-07, + "loss": 0.6783, + "step": 3055 + }, + { + "epoch": 5.659259259259259, + "grad_norm": 1.5337280438773258, + "learning_rate": 1.9355224490652504e-07, + "loss": 0.6768, + "step": 3056 + }, + { + "epoch": 5.661111111111111, + "grad_norm": 1.5448094146964897, + "learning_rate": 1.9354694688187768e-07, + "loss": 0.6696, + "step": 3057 + }, + { + "epoch": 5.662962962962963, + "grad_norm": 1.860632010249602, + "learning_rate": 1.935416467540409e-07, + "loss": 0.6922, + "step": 3058 + }, + { + "epoch": 5.6648148148148145, + "grad_norm": 1.6359189036459534, + "learning_rate": 1.9353634452313382e-07, + "loss": 0.6665, + "step": 3059 + }, + { + "epoch": 5.666666666666667, + "grad_norm": 1.434756173356067, + "learning_rate": 1.9353104018927565e-07, + "loss": 0.6484, + "step": 3060 + }, + { + "epoch": 5.6685185185185185, + "grad_norm": 1.948186153162462, + "learning_rate": 1.9352573375258575e-07, + "loss": 0.67, + "step": 3061 + }, + { + "epoch": 5.67037037037037, + "grad_norm": 1.5982433520538097, + "learning_rate": 1.9352042521318327e-07, + "loss": 0.6913, + "step": 3062 + }, + { + "epoch": 5.6722222222222225, + "grad_norm": 1.380339450597154, + "learning_rate": 1.9351511457118765e-07, + "loss": 0.6742, + "step": 3063 + }, + { + "epoch": 5.674074074074074, + "grad_norm": 1.4197555959778585, + "learning_rate": 1.9350980182671827e-07, + "loss": 0.6668, + "step": 3064 + }, + { + "epoch": 5.675925925925926, + "grad_norm": 1.5029882107411736, + "learning_rate": 1.9350448697989462e-07, + "loss": 0.672, + "step": 3065 + }, + { + "epoch": 5.677777777777778, + "grad_norm": 1.6434691743536687, + "learning_rate": 1.9349917003083611e-07, + "loss": 0.6807, + "step": 3066 + }, + { + "epoch": 5.67962962962963, + "grad_norm": 1.5870592522694542, + "learning_rate": 1.9349385097966233e-07, + "loss": 0.6708, + "step": 3067 + }, + { + "epoch": 5.681481481481481, + "grad_norm": 1.6101027911119095, + "learning_rate": 1.9348852982649283e-07, + "loss": 0.67, + "step": 3068 + }, + { + "epoch": 5.683333333333334, + "grad_norm": 1.3936503593260219, + "learning_rate": 1.934832065714473e-07, + "loss": 0.6972, + "step": 3069 + }, + { + "epoch": 5.685185185185185, + "grad_norm": 1.7572130240760366, + "learning_rate": 1.9347788121464537e-07, + "loss": 0.6467, + "step": 3070 + }, + { + "epoch": 5.687037037037037, + "grad_norm": 1.547500074074858, + "learning_rate": 1.934725537562068e-07, + "loss": 0.7068, + "step": 3071 + }, + { + "epoch": 5.688888888888889, + "grad_norm": 1.4973576040083854, + "learning_rate": 1.9346722419625135e-07, + "loss": 0.6773, + "step": 3072 + }, + { + "epoch": 5.690740740740741, + "grad_norm": 1.3716341793583535, + "learning_rate": 1.9346189253489884e-07, + "loss": 0.682, + "step": 3073 + }, + { + "epoch": 5.692592592592592, + "grad_norm": 1.7507547486927795, + "learning_rate": 1.9345655877226913e-07, + "loss": 0.6852, + "step": 3074 + }, + { + "epoch": 5.694444444444445, + "grad_norm": 1.3766742209370904, + "learning_rate": 1.9345122290848218e-07, + "loss": 0.6749, + "step": 3075 + }, + { + "epoch": 5.696296296296296, + "grad_norm": 2.4314264555509375, + "learning_rate": 1.9344588494365794e-07, + "loss": 0.6577, + "step": 3076 + }, + { + "epoch": 5.698148148148148, + "grad_norm": 1.3732823957425164, + "learning_rate": 1.9344054487791638e-07, + "loss": 0.6797, + "step": 3077 + }, + { + "epoch": 5.7, + "grad_norm": 2.441507971640275, + "learning_rate": 1.934352027113776e-07, + "loss": 0.669, + "step": 3078 + }, + { + "epoch": 5.701851851851852, + "grad_norm": 1.7895974928068723, + "learning_rate": 1.9342985844416173e-07, + "loss": 0.6864, + "step": 3079 + }, + { + "epoch": 5.703703703703704, + "grad_norm": 1.515956115089075, + "learning_rate": 1.9342451207638888e-07, + "loss": 0.6937, + "step": 3080 + }, + { + "epoch": 5.705555555555556, + "grad_norm": 1.547323622781393, + "learning_rate": 1.9341916360817921e-07, + "loss": 0.7071, + "step": 3081 + }, + { + "epoch": 5.707407407407407, + "grad_norm": 1.655857180601228, + "learning_rate": 1.9341381303965308e-07, + "loss": 0.6191, + "step": 3082 + }, + { + "epoch": 5.709259259259259, + "grad_norm": 1.484552359747053, + "learning_rate": 1.934084603709307e-07, + "loss": 0.6599, + "step": 3083 + }, + { + "epoch": 5.711111111111111, + "grad_norm": 1.5035470451854136, + "learning_rate": 1.9340310560213244e-07, + "loss": 0.6687, + "step": 3084 + }, + { + "epoch": 5.712962962962963, + "grad_norm": 1.3915232327243305, + "learning_rate": 1.9339774873337868e-07, + "loss": 0.6757, + "step": 3085 + }, + { + "epoch": 5.714814814814815, + "grad_norm": 1.7470536229939362, + "learning_rate": 1.9339238976478986e-07, + "loss": 0.6644, + "step": 3086 + }, + { + "epoch": 5.716666666666667, + "grad_norm": 1.7876411304128315, + "learning_rate": 1.9338702869648648e-07, + "loss": 0.6545, + "step": 3087 + }, + { + "epoch": 5.718518518518518, + "grad_norm": 1.5541025370563417, + "learning_rate": 1.9338166552858906e-07, + "loss": 0.6564, + "step": 3088 + }, + { + "epoch": 5.72037037037037, + "grad_norm": 1.6010514811084968, + "learning_rate": 1.9337630026121818e-07, + "loss": 0.6891, + "step": 3089 + }, + { + "epoch": 5.722222222222222, + "grad_norm": 1.5908301722684488, + "learning_rate": 1.9337093289449446e-07, + "loss": 0.6584, + "step": 3090 + }, + { + "epoch": 5.724074074074074, + "grad_norm": 1.6367252369865792, + "learning_rate": 1.9336556342853857e-07, + "loss": 0.6718, + "step": 3091 + }, + { + "epoch": 5.725925925925926, + "grad_norm": 1.7485577159550505, + "learning_rate": 1.9336019186347125e-07, + "loss": 0.7073, + "step": 3092 + }, + { + "epoch": 5.727777777777778, + "grad_norm": 1.4950591970547817, + "learning_rate": 1.9335481819941326e-07, + "loss": 0.6573, + "step": 3093 + }, + { + "epoch": 5.729629629629629, + "grad_norm": 1.377075708431481, + "learning_rate": 1.9334944243648538e-07, + "loss": 0.6629, + "step": 3094 + }, + { + "epoch": 5.731481481481482, + "grad_norm": 1.307829341741579, + "learning_rate": 1.9334406457480853e-07, + "loss": 0.6168, + "step": 3095 + }, + { + "epoch": 5.733333333333333, + "grad_norm": 1.4463013563843563, + "learning_rate": 1.9333868461450358e-07, + "loss": 0.6969, + "step": 3096 + }, + { + "epoch": 5.735185185185185, + "grad_norm": 1.7536236222455248, + "learning_rate": 1.933333025556915e-07, + "loss": 0.6263, + "step": 3097 + }, + { + "epoch": 5.737037037037037, + "grad_norm": 1.4390663806014874, + "learning_rate": 1.9332791839849328e-07, + "loss": 0.6141, + "step": 3098 + }, + { + "epoch": 5.738888888888889, + "grad_norm": 1.3569344420250968, + "learning_rate": 1.9332253214302997e-07, + "loss": 0.6828, + "step": 3099 + }, + { + "epoch": 5.7407407407407405, + "grad_norm": 3.348758968827594, + "learning_rate": 1.933171437894227e-07, + "loss": 0.6576, + "step": 3100 + }, + { + "epoch": 5.742592592592593, + "grad_norm": 1.805984983846325, + "learning_rate": 1.9331175333779257e-07, + "loss": 0.7046, + "step": 3101 + }, + { + "epoch": 5.7444444444444445, + "grad_norm": 1.3755396359386274, + "learning_rate": 1.933063607882608e-07, + "loss": 0.6598, + "step": 3102 + }, + { + "epoch": 5.746296296296296, + "grad_norm": 1.5206046602809848, + "learning_rate": 1.9330096614094864e-07, + "loss": 0.6279, + "step": 3103 + }, + { + "epoch": 5.7481481481481485, + "grad_norm": 1.9727222083918303, + "learning_rate": 1.9329556939597733e-07, + "loss": 0.6818, + "step": 3104 + }, + { + "epoch": 5.75, + "grad_norm": 1.4488535226291053, + "learning_rate": 1.9329017055346828e-07, + "loss": 0.6259, + "step": 3105 + }, + { + "epoch": 5.7518518518518515, + "grad_norm": 1.8723026657408528, + "learning_rate": 1.9328476961354277e-07, + "loss": 0.6493, + "step": 3106 + }, + { + "epoch": 5.753703703703704, + "grad_norm": 1.6088414825401331, + "learning_rate": 1.932793665763223e-07, + "loss": 0.6763, + "step": 3107 + }, + { + "epoch": 5.7555555555555555, + "grad_norm": 1.7677963540157626, + "learning_rate": 1.9327396144192835e-07, + "loss": 0.6611, + "step": 3108 + }, + { + "epoch": 5.757407407407407, + "grad_norm": 1.844647273626407, + "learning_rate": 1.932685542104824e-07, + "loss": 0.7264, + "step": 3109 + }, + { + "epoch": 5.7592592592592595, + "grad_norm": 1.6325255691915002, + "learning_rate": 1.9326314488210603e-07, + "loss": 0.6503, + "step": 3110 + }, + { + "epoch": 5.761111111111111, + "grad_norm": 1.3434655203475394, + "learning_rate": 1.932577334569209e-07, + "loss": 0.6529, + "step": 3111 + }, + { + "epoch": 5.762962962962963, + "grad_norm": 1.6255110816283564, + "learning_rate": 1.932523199350486e-07, + "loss": 0.6508, + "step": 3112 + }, + { + "epoch": 5.764814814814815, + "grad_norm": 1.6089545124745757, + "learning_rate": 1.932469043166109e-07, + "loss": 0.6728, + "step": 3113 + }, + { + "epoch": 5.766666666666667, + "grad_norm": 1.7591660200264743, + "learning_rate": 1.9324148660172952e-07, + "loss": 0.6494, + "step": 3114 + }, + { + "epoch": 5.768518518518518, + "grad_norm": 1.479934145887186, + "learning_rate": 1.9323606679052627e-07, + "loss": 0.6833, + "step": 3115 + }, + { + "epoch": 5.770370370370371, + "grad_norm": 2.0597281586506795, + "learning_rate": 1.9323064488312303e-07, + "loss": 0.6371, + "step": 3116 + }, + { + "epoch": 5.772222222222222, + "grad_norm": 1.4814336960370498, + "learning_rate": 1.932252208796417e-07, + "loss": 0.7106, + "step": 3117 + }, + { + "epoch": 5.774074074074074, + "grad_norm": 1.5476183995742245, + "learning_rate": 1.932197947802042e-07, + "loss": 0.7032, + "step": 3118 + }, + { + "epoch": 5.775925925925926, + "grad_norm": 1.5465393185472458, + "learning_rate": 1.932143665849325e-07, + "loss": 0.6848, + "step": 3119 + }, + { + "epoch": 5.777777777777778, + "grad_norm": 1.5982362981341727, + "learning_rate": 1.9320893629394873e-07, + "loss": 0.6983, + "step": 3120 + }, + { + "epoch": 5.77962962962963, + "grad_norm": 1.5609548423717912, + "learning_rate": 1.9320350390737484e-07, + "loss": 0.6436, + "step": 3121 + }, + { + "epoch": 5.781481481481482, + "grad_norm": 1.482513581745419, + "learning_rate": 1.931980694253331e-07, + "loss": 0.6482, + "step": 3122 + }, + { + "epoch": 5.783333333333333, + "grad_norm": 1.5968450403624015, + "learning_rate": 1.9319263284794564e-07, + "loss": 0.6284, + "step": 3123 + }, + { + "epoch": 5.785185185185185, + "grad_norm": 1.4210885808309044, + "learning_rate": 1.931871941753347e-07, + "loss": 0.6655, + "step": 3124 + }, + { + "epoch": 5.787037037037037, + "grad_norm": 1.5514072248636095, + "learning_rate": 1.931817534076225e-07, + "loss": 0.7029, + "step": 3125 + }, + { + "epoch": 5.788888888888889, + "grad_norm": 1.6207300826295732, + "learning_rate": 1.9317631054493142e-07, + "loss": 0.6508, + "step": 3126 + }, + { + "epoch": 5.790740740740741, + "grad_norm": 1.61700088372504, + "learning_rate": 1.9317086558738383e-07, + "loss": 0.6496, + "step": 3127 + }, + { + "epoch": 5.792592592592593, + "grad_norm": 1.6775660047990084, + "learning_rate": 1.931654185351021e-07, + "loss": 0.676, + "step": 3128 + }, + { + "epoch": 5.794444444444444, + "grad_norm": 1.4122099768719387, + "learning_rate": 1.9315996938820876e-07, + "loss": 0.7009, + "step": 3129 + }, + { + "epoch": 5.796296296296296, + "grad_norm": 1.593078724967063, + "learning_rate": 1.931545181468263e-07, + "loss": 0.6658, + "step": 3130 + }, + { + "epoch": 5.798148148148148, + "grad_norm": 1.989623385071106, + "learning_rate": 1.9314906481107722e-07, + "loss": 0.6589, + "step": 3131 + }, + { + "epoch": 5.8, + "grad_norm": 1.8452054582470114, + "learning_rate": 1.9314360938108424e-07, + "loss": 0.6918, + "step": 3132 + }, + { + "epoch": 5.801851851851852, + "grad_norm": 1.6139587687033337, + "learning_rate": 1.931381518569699e-07, + "loss": 0.6327, + "step": 3133 + }, + { + "epoch": 5.803703703703704, + "grad_norm": 1.8197311604550148, + "learning_rate": 1.9313269223885692e-07, + "loss": 0.6453, + "step": 3134 + }, + { + "epoch": 5.805555555555555, + "grad_norm": 2.2888817665587418, + "learning_rate": 1.9312723052686812e-07, + "loss": 0.7086, + "step": 3135 + }, + { + "epoch": 5.807407407407408, + "grad_norm": 1.4006108452242423, + "learning_rate": 1.9312176672112626e-07, + "loss": 0.6337, + "step": 3136 + }, + { + "epoch": 5.809259259259259, + "grad_norm": 1.4705205932652263, + "learning_rate": 1.9311630082175415e-07, + "loss": 0.6773, + "step": 3137 + }, + { + "epoch": 5.811111111111111, + "grad_norm": 1.663415331183204, + "learning_rate": 1.931108328288747e-07, + "loss": 0.697, + "step": 3138 + }, + { + "epoch": 5.812962962962963, + "grad_norm": 1.9988591218228136, + "learning_rate": 1.9310536274261083e-07, + "loss": 0.6556, + "step": 3139 + }, + { + "epoch": 5.814814814814815, + "grad_norm": 1.6198445757075786, + "learning_rate": 1.9309989056308553e-07, + "loss": 0.6886, + "step": 3140 + }, + { + "epoch": 5.816666666666666, + "grad_norm": 1.7800971799868306, + "learning_rate": 1.9309441629042186e-07, + "loss": 0.7088, + "step": 3141 + }, + { + "epoch": 5.818518518518519, + "grad_norm": 1.5991795396771342, + "learning_rate": 1.9308893992474285e-07, + "loss": 0.6969, + "step": 3142 + }, + { + "epoch": 5.82037037037037, + "grad_norm": 1.530754210435707, + "learning_rate": 1.9308346146617164e-07, + "loss": 0.6736, + "step": 3143 + }, + { + "epoch": 5.822222222222222, + "grad_norm": 1.7214837627237491, + "learning_rate": 1.9307798091483143e-07, + "loss": 0.651, + "step": 3144 + }, + { + "epoch": 5.824074074074074, + "grad_norm": 1.4735048569022127, + "learning_rate": 1.930724982708454e-07, + "loss": 0.6376, + "step": 3145 + }, + { + "epoch": 5.825925925925926, + "grad_norm": 1.606932832107485, + "learning_rate": 1.9306701353433685e-07, + "loss": 0.6213, + "step": 3146 + }, + { + "epoch": 5.8277777777777775, + "grad_norm": 1.472813998472956, + "learning_rate": 1.9306152670542906e-07, + "loss": 0.7253, + "step": 3147 + }, + { + "epoch": 5.82962962962963, + "grad_norm": 1.8671947681157124, + "learning_rate": 1.9305603778424537e-07, + "loss": 0.6667, + "step": 3148 + }, + { + "epoch": 5.8314814814814815, + "grad_norm": 1.5151768405053028, + "learning_rate": 1.9305054677090924e-07, + "loss": 0.6783, + "step": 3149 + }, + { + "epoch": 5.833333333333333, + "grad_norm": 1.6116926068269934, + "learning_rate": 1.930450536655441e-07, + "loss": 0.6297, + "step": 3150 + }, + { + "epoch": 5.8351851851851855, + "grad_norm": 1.6594230998501391, + "learning_rate": 1.9303955846827343e-07, + "loss": 0.741, + "step": 3151 + }, + { + "epoch": 5.837037037037037, + "grad_norm": 1.6451829488166798, + "learning_rate": 1.930340611792208e-07, + "loss": 0.6389, + "step": 3152 + }, + { + "epoch": 5.838888888888889, + "grad_norm": 1.5208046282215946, + "learning_rate": 1.930285617985098e-07, + "loss": 0.6726, + "step": 3153 + }, + { + "epoch": 5.840740740740741, + "grad_norm": 1.399174385757347, + "learning_rate": 1.9302306032626404e-07, + "loss": 0.6238, + "step": 3154 + }, + { + "epoch": 5.842592592592593, + "grad_norm": 1.6934333327737037, + "learning_rate": 1.9301755676260729e-07, + "loss": 0.6697, + "step": 3155 + }, + { + "epoch": 5.844444444444444, + "grad_norm": 1.6870100432688353, + "learning_rate": 1.9301205110766322e-07, + "loss": 0.688, + "step": 3156 + }, + { + "epoch": 5.8462962962962965, + "grad_norm": 1.3879402122138877, + "learning_rate": 1.930065433615556e-07, + "loss": 0.6294, + "step": 3157 + }, + { + "epoch": 5.848148148148148, + "grad_norm": 1.8457281896677984, + "learning_rate": 1.930010335244083e-07, + "loss": 0.6578, + "step": 3158 + }, + { + "epoch": 5.85, + "grad_norm": 1.735811771558393, + "learning_rate": 1.9299552159634515e-07, + "loss": 0.6395, + "step": 3159 + }, + { + "epoch": 5.851851851851852, + "grad_norm": 1.4172907329530469, + "learning_rate": 1.9299000757749013e-07, + "loss": 0.6818, + "step": 3160 + }, + { + "epoch": 5.853703703703704, + "grad_norm": 1.3985172088259052, + "learning_rate": 1.929844914679672e-07, + "loss": 0.6377, + "step": 3161 + }, + { + "epoch": 5.855555555555555, + "grad_norm": 1.3511012971509335, + "learning_rate": 1.929789732679003e-07, + "loss": 0.6998, + "step": 3162 + }, + { + "epoch": 5.857407407407408, + "grad_norm": 1.4673256331251812, + "learning_rate": 1.9297345297741362e-07, + "loss": 0.6731, + "step": 3163 + }, + { + "epoch": 5.859259259259259, + "grad_norm": 1.5154356547355405, + "learning_rate": 1.9296793059663118e-07, + "loss": 0.6685, + "step": 3164 + }, + { + "epoch": 5.861111111111111, + "grad_norm": 1.471094740838117, + "learning_rate": 1.9296240612567716e-07, + "loss": 0.6708, + "step": 3165 + }, + { + "epoch": 5.862962962962963, + "grad_norm": 1.6823911525751123, + "learning_rate": 1.9295687956467577e-07, + "loss": 0.7071, + "step": 3166 + }, + { + "epoch": 5.864814814814815, + "grad_norm": 1.549637030960512, + "learning_rate": 1.9295135091375126e-07, + "loss": 0.7085, + "step": 3167 + }, + { + "epoch": 5.866666666666667, + "grad_norm": 1.4823879490741958, + "learning_rate": 1.9294582017302794e-07, + "loss": 0.7546, + "step": 3168 + }, + { + "epoch": 5.868518518518519, + "grad_norm": 1.6035790223586877, + "learning_rate": 1.9294028734263014e-07, + "loss": 0.6597, + "step": 3169 + }, + { + "epoch": 5.87037037037037, + "grad_norm": 1.3380787951200714, + "learning_rate": 1.9293475242268223e-07, + "loss": 0.655, + "step": 3170 + }, + { + "epoch": 5.872222222222222, + "grad_norm": 1.6254994848876927, + "learning_rate": 1.929292154133087e-07, + "loss": 0.6769, + "step": 3171 + }, + { + "epoch": 5.874074074074074, + "grad_norm": 1.6421715807954698, + "learning_rate": 1.9292367631463402e-07, + "loss": 0.6455, + "step": 3172 + }, + { + "epoch": 5.875925925925926, + "grad_norm": 1.6381023825796484, + "learning_rate": 1.9291813512678272e-07, + "loss": 0.6654, + "step": 3173 + }, + { + "epoch": 5.877777777777778, + "grad_norm": 1.8532628640486033, + "learning_rate": 1.9291259184987938e-07, + "loss": 0.6371, + "step": 3174 + }, + { + "epoch": 5.87962962962963, + "grad_norm": 1.6330188547559685, + "learning_rate": 1.9290704648404863e-07, + "loss": 0.6715, + "step": 3175 + }, + { + "epoch": 5.881481481481481, + "grad_norm": 1.504033728320947, + "learning_rate": 1.9290149902941515e-07, + "loss": 0.6551, + "step": 3176 + }, + { + "epoch": 5.883333333333333, + "grad_norm": 2.0499396534022436, + "learning_rate": 1.9289594948610364e-07, + "loss": 0.6942, + "step": 3177 + }, + { + "epoch": 5.885185185185185, + "grad_norm": 1.5782185557272341, + "learning_rate": 1.9289039785423888e-07, + "loss": 0.6325, + "step": 3178 + }, + { + "epoch": 5.887037037037037, + "grad_norm": 1.448502024093899, + "learning_rate": 1.9288484413394567e-07, + "loss": 0.6532, + "step": 3179 + }, + { + "epoch": 5.888888888888889, + "grad_norm": 1.3941981189264658, + "learning_rate": 1.9287928832534893e-07, + "loss": 0.6546, + "step": 3180 + }, + { + "epoch": 5.890740740740741, + "grad_norm": 2.0057391865140755, + "learning_rate": 1.9287373042857356e-07, + "loss": 0.6574, + "step": 3181 + }, + { + "epoch": 5.892592592592592, + "grad_norm": 2.0301992019694723, + "learning_rate": 1.9286817044374443e-07, + "loss": 0.6781, + "step": 3182 + }, + { + "epoch": 5.894444444444445, + "grad_norm": 1.6096883955639443, + "learning_rate": 1.9286260837098662e-07, + "loss": 0.7255, + "step": 3183 + }, + { + "epoch": 5.896296296296296, + "grad_norm": 1.4246677287421317, + "learning_rate": 1.9285704421042516e-07, + "loss": 0.6526, + "step": 3184 + }, + { + "epoch": 5.898148148148148, + "grad_norm": 1.7999254664651674, + "learning_rate": 1.9285147796218516e-07, + "loss": 0.6591, + "step": 3185 + }, + { + "epoch": 5.9, + "grad_norm": 1.7472846916172196, + "learning_rate": 1.9284590962639176e-07, + "loss": 0.6475, + "step": 3186 + }, + { + "epoch": 5.901851851851852, + "grad_norm": 1.5182371952362772, + "learning_rate": 1.9284033920317013e-07, + "loss": 0.6531, + "step": 3187 + }, + { + "epoch": 5.9037037037037035, + "grad_norm": 1.399734562440533, + "learning_rate": 1.9283476669264554e-07, + "loss": 0.6824, + "step": 3188 + }, + { + "epoch": 5.905555555555556, + "grad_norm": 1.4505831566307255, + "learning_rate": 1.9282919209494323e-07, + "loss": 0.6564, + "step": 3189 + }, + { + "epoch": 5.907407407407407, + "grad_norm": 1.6983977132524728, + "learning_rate": 1.9282361541018858e-07, + "loss": 0.6363, + "step": 3190 + }, + { + "epoch": 5.909259259259259, + "grad_norm": 1.462861434244709, + "learning_rate": 1.9281803663850699e-07, + "loss": 0.6915, + "step": 3191 + }, + { + "epoch": 5.911111111111111, + "grad_norm": 1.6590387438118874, + "learning_rate": 1.9281245578002378e-07, + "loss": 0.6384, + "step": 3192 + }, + { + "epoch": 5.912962962962963, + "grad_norm": 1.5221404311931686, + "learning_rate": 1.9280687283486455e-07, + "loss": 0.7088, + "step": 3193 + }, + { + "epoch": 5.9148148148148145, + "grad_norm": 1.6158752464994992, + "learning_rate": 1.9280128780315472e-07, + "loss": 0.6702, + "step": 3194 + }, + { + "epoch": 5.916666666666667, + "grad_norm": 1.4166402620577003, + "learning_rate": 1.927957006850199e-07, + "loss": 0.6814, + "step": 3195 + }, + { + "epoch": 5.9185185185185185, + "grad_norm": 1.4815897112790115, + "learning_rate": 1.9279011148058572e-07, + "loss": 0.6381, + "step": 3196 + }, + { + "epoch": 5.92037037037037, + "grad_norm": 1.4596040691377503, + "learning_rate": 1.9278452018997779e-07, + "loss": 0.68, + "step": 3197 + }, + { + "epoch": 5.9222222222222225, + "grad_norm": 1.3901922117362056, + "learning_rate": 1.9277892681332183e-07, + "loss": 0.6459, + "step": 3198 + }, + { + "epoch": 5.924074074074074, + "grad_norm": 1.9296858392826894, + "learning_rate": 1.9277333135074368e-07, + "loss": 0.6415, + "step": 3199 + }, + { + "epoch": 5.925925925925926, + "grad_norm": 1.5138859044023336, + "learning_rate": 1.92767733802369e-07, + "loss": 0.6552, + "step": 3200 + }, + { + "epoch": 5.927777777777778, + "grad_norm": 1.7278625766545153, + "learning_rate": 1.9276213416832374e-07, + "loss": 0.6726, + "step": 3201 + }, + { + "epoch": 5.92962962962963, + "grad_norm": 1.4943273887276074, + "learning_rate": 1.927565324487338e-07, + "loss": 0.6791, + "step": 3202 + }, + { + "epoch": 5.931481481481481, + "grad_norm": 1.6859281712062053, + "learning_rate": 1.9275092864372504e-07, + "loss": 0.6731, + "step": 3203 + }, + { + "epoch": 5.933333333333334, + "grad_norm": 1.5249263227319036, + "learning_rate": 1.9274532275342352e-07, + "loss": 0.6411, + "step": 3204 + }, + { + "epoch": 5.935185185185185, + "grad_norm": 1.3001713793707412, + "learning_rate": 1.9273971477795525e-07, + "loss": 0.6836, + "step": 3205 + }, + { + "epoch": 5.937037037037037, + "grad_norm": 1.8766222587097663, + "learning_rate": 1.9273410471744632e-07, + "loss": 0.6625, + "step": 3206 + }, + { + "epoch": 5.938888888888889, + "grad_norm": 1.567684611784952, + "learning_rate": 1.9272849257202283e-07, + "loss": 0.673, + "step": 3207 + }, + { + "epoch": 5.940740740740741, + "grad_norm": 2.072647953912945, + "learning_rate": 1.92722878341811e-07, + "loss": 0.6637, + "step": 3208 + }, + { + "epoch": 5.942592592592592, + "grad_norm": 1.3276483101642031, + "learning_rate": 1.92717262026937e-07, + "loss": 0.67, + "step": 3209 + }, + { + "epoch": 5.944444444444445, + "grad_norm": 1.6291842182860714, + "learning_rate": 1.9271164362752718e-07, + "loss": 0.6295, + "step": 3210 + }, + { + "epoch": 5.946296296296296, + "grad_norm": 1.6531296674625306, + "learning_rate": 1.9270602314370778e-07, + "loss": 0.7137, + "step": 3211 + }, + { + "epoch": 5.948148148148148, + "grad_norm": 1.5011625741864751, + "learning_rate": 1.927004005756052e-07, + "loss": 0.6711, + "step": 3212 + }, + { + "epoch": 5.95, + "grad_norm": 2.2772641566950367, + "learning_rate": 1.9269477592334587e-07, + "loss": 0.6786, + "step": 3213 + }, + { + "epoch": 5.951851851851852, + "grad_norm": 2.3301676799243016, + "learning_rate": 1.926891491870562e-07, + "loss": 0.6043, + "step": 3214 + }, + { + "epoch": 5.953703703703704, + "grad_norm": 2.1565166126831206, + "learning_rate": 1.9268352036686274e-07, + "loss": 0.6381, + "step": 3215 + }, + { + "epoch": 5.955555555555556, + "grad_norm": 1.3965667144729528, + "learning_rate": 1.9267788946289199e-07, + "loss": 0.6936, + "step": 3216 + }, + { + "epoch": 5.957407407407407, + "grad_norm": 1.668933927186015, + "learning_rate": 1.9267225647527058e-07, + "loss": 0.6989, + "step": 3217 + }, + { + "epoch": 5.959259259259259, + "grad_norm": 1.6814189012613956, + "learning_rate": 1.9266662140412515e-07, + "loss": 0.6764, + "step": 3218 + }, + { + "epoch": 5.961111111111111, + "grad_norm": 1.6573762483619858, + "learning_rate": 1.926609842495824e-07, + "loss": 0.6587, + "step": 3219 + }, + { + "epoch": 5.962962962962963, + "grad_norm": 1.5925934986670407, + "learning_rate": 1.9265534501176905e-07, + "loss": 0.6646, + "step": 3220 + }, + { + "epoch": 5.964814814814815, + "grad_norm": 1.7386825765795433, + "learning_rate": 1.926497036908119e-07, + "loss": 0.7053, + "step": 3221 + }, + { + "epoch": 5.966666666666667, + "grad_norm": 1.5513124747883649, + "learning_rate": 1.9264406028683778e-07, + "loss": 0.6114, + "step": 3222 + }, + { + "epoch": 5.968518518518518, + "grad_norm": 1.741564320479593, + "learning_rate": 1.9263841479997357e-07, + "loss": 0.6316, + "step": 3223 + }, + { + "epoch": 5.97037037037037, + "grad_norm": 1.5200633163767672, + "learning_rate": 1.926327672303462e-07, + "loss": 0.6797, + "step": 3224 + }, + { + "epoch": 5.972222222222222, + "grad_norm": 1.577544443816808, + "learning_rate": 1.9262711757808262e-07, + "loss": 0.6952, + "step": 3225 + }, + { + "epoch": 5.974074074074074, + "grad_norm": 1.5480104319689156, + "learning_rate": 1.9262146584330987e-07, + "loss": 0.6544, + "step": 3226 + }, + { + "epoch": 5.975925925925926, + "grad_norm": 1.5372275894781748, + "learning_rate": 1.92615812026155e-07, + "loss": 0.681, + "step": 3227 + }, + { + "epoch": 5.977777777777778, + "grad_norm": 1.700402767559671, + "learning_rate": 1.9261015612674515e-07, + "loss": 0.687, + "step": 3228 + }, + { + "epoch": 5.979629629629629, + "grad_norm": 1.5279800214160841, + "learning_rate": 1.9260449814520747e-07, + "loss": 0.6141, + "step": 3229 + }, + { + "epoch": 5.981481481481482, + "grad_norm": 1.6164714346226579, + "learning_rate": 1.9259883808166918e-07, + "loss": 0.6804, + "step": 3230 + }, + { + "epoch": 5.983333333333333, + "grad_norm": 1.5886694542372342, + "learning_rate": 1.9259317593625747e-07, + "loss": 0.6719, + "step": 3231 + }, + { + "epoch": 5.985185185185185, + "grad_norm": 1.49270136076904, + "learning_rate": 1.925875117090997e-07, + "loss": 0.6736, + "step": 3232 + }, + { + "epoch": 5.987037037037037, + "grad_norm": 1.5241185598741427, + "learning_rate": 1.9258184540032322e-07, + "loss": 0.6624, + "step": 3233 + }, + { + "epoch": 5.988888888888889, + "grad_norm": 1.4837138337792564, + "learning_rate": 1.925761770100554e-07, + "loss": 0.6876, + "step": 3234 + }, + { + "epoch": 5.9907407407407405, + "grad_norm": 1.6546796169015763, + "learning_rate": 1.9257050653842369e-07, + "loss": 0.6527, + "step": 3235 + }, + { + "epoch": 5.992592592592593, + "grad_norm": 1.5188511838123737, + "learning_rate": 1.9256483398555553e-07, + "loss": 0.6526, + "step": 3236 + }, + { + "epoch": 5.9944444444444445, + "grad_norm": 1.5601513374579616, + "learning_rate": 1.9255915935157857e-07, + "loss": 0.6818, + "step": 3237 + }, + { + "epoch": 5.996296296296296, + "grad_norm": 1.5844502052633924, + "learning_rate": 1.9255348263662027e-07, + "loss": 0.6483, + "step": 3238 + }, + { + "epoch": 5.9981481481481485, + "grad_norm": 1.695389321959413, + "learning_rate": 1.9254780384080833e-07, + "loss": 0.6888, + "step": 3239 + }, + { + "epoch": 6.0, + "grad_norm": 1.506239263979736, + "learning_rate": 1.925421229642704e-07, + "loss": 0.6766, + "step": 3240 + }, + { + "epoch": 6.0018518518518515, + "grad_norm": 1.4448384705576789, + "learning_rate": 1.9253644000713422e-07, + "loss": 0.658, + "step": 3241 + }, + { + "epoch": 6.003703703703704, + "grad_norm": 1.5068671226051193, + "learning_rate": 1.9253075496952754e-07, + "loss": 0.663, + "step": 3242 + }, + { + "epoch": 6.0055555555555555, + "grad_norm": 1.9336947608505894, + "learning_rate": 1.9252506785157817e-07, + "loss": 0.667, + "step": 3243 + }, + { + "epoch": 6.007407407407407, + "grad_norm": 1.4939981006308398, + "learning_rate": 1.9251937865341397e-07, + "loss": 0.6319, + "step": 3244 + }, + { + "epoch": 6.0092592592592595, + "grad_norm": 1.5241668308773684, + "learning_rate": 1.9251368737516286e-07, + "loss": 0.6379, + "step": 3245 + }, + { + "epoch": 6.011111111111111, + "grad_norm": 1.9521554429013426, + "learning_rate": 1.925079940169528e-07, + "loss": 0.6845, + "step": 3246 + }, + { + "epoch": 6.012962962962963, + "grad_norm": 1.5794745158179309, + "learning_rate": 1.925022985789118e-07, + "loss": 0.6585, + "step": 3247 + }, + { + "epoch": 6.014814814814815, + "grad_norm": 1.9881908666505577, + "learning_rate": 1.924966010611679e-07, + "loss": 0.661, + "step": 3248 + }, + { + "epoch": 6.016666666666667, + "grad_norm": 1.402943567940391, + "learning_rate": 1.9249090146384918e-07, + "loss": 0.6384, + "step": 3249 + }, + { + "epoch": 6.018518518518518, + "grad_norm": 1.4462048670835508, + "learning_rate": 1.9248519978708377e-07, + "loss": 0.6747, + "step": 3250 + }, + { + "epoch": 6.020370370370371, + "grad_norm": 1.6993090392276184, + "learning_rate": 1.9247949603099988e-07, + "loss": 0.6143, + "step": 3251 + }, + { + "epoch": 6.022222222222222, + "grad_norm": 1.5586666410459007, + "learning_rate": 1.924737901957258e-07, + "loss": 0.6645, + "step": 3252 + }, + { + "epoch": 6.024074074074074, + "grad_norm": 1.5249833849458387, + "learning_rate": 1.9246808228138972e-07, + "loss": 0.6642, + "step": 3253 + }, + { + "epoch": 6.025925925925926, + "grad_norm": 1.6049132156048016, + "learning_rate": 1.9246237228812002e-07, + "loss": 0.6373, + "step": 3254 + }, + { + "epoch": 6.027777777777778, + "grad_norm": 1.665420901538134, + "learning_rate": 1.9245666021604507e-07, + "loss": 0.6542, + "step": 3255 + }, + { + "epoch": 6.029629629629629, + "grad_norm": 1.5923442851880225, + "learning_rate": 1.9245094606529332e-07, + "loss": 0.6684, + "step": 3256 + }, + { + "epoch": 6.031481481481482, + "grad_norm": 1.547366313597984, + "learning_rate": 1.9244522983599315e-07, + "loss": 0.6337, + "step": 3257 + }, + { + "epoch": 6.033333333333333, + "grad_norm": 2.153715134123784, + "learning_rate": 1.924395115282732e-07, + "loss": 0.6686, + "step": 3258 + }, + { + "epoch": 6.035185185185185, + "grad_norm": 1.4563291619516572, + "learning_rate": 1.924337911422619e-07, + "loss": 0.6763, + "step": 3259 + }, + { + "epoch": 6.037037037037037, + "grad_norm": 1.6643606281204921, + "learning_rate": 1.9242806867808795e-07, + "loss": 0.6638, + "step": 3260 + }, + { + "epoch": 6.038888888888889, + "grad_norm": 1.548870741599961, + "learning_rate": 1.9242234413588002e-07, + "loss": 0.6497, + "step": 3261 + }, + { + "epoch": 6.040740740740741, + "grad_norm": 2.9327633456583677, + "learning_rate": 1.9241661751576676e-07, + "loss": 0.6674, + "step": 3262 + }, + { + "epoch": 6.042592592592593, + "grad_norm": 1.4901888064230668, + "learning_rate": 1.9241088881787694e-07, + "loss": 0.6462, + "step": 3263 + }, + { + "epoch": 6.044444444444444, + "grad_norm": 1.4065076667917518, + "learning_rate": 1.9240515804233936e-07, + "loss": 0.6399, + "step": 3264 + }, + { + "epoch": 6.046296296296297, + "grad_norm": 1.350434313806594, + "learning_rate": 1.9239942518928286e-07, + "loss": 0.6731, + "step": 3265 + }, + { + "epoch": 6.048148148148148, + "grad_norm": 1.4158879874539907, + "learning_rate": 1.9239369025883628e-07, + "loss": 0.6648, + "step": 3266 + }, + { + "epoch": 6.05, + "grad_norm": 1.9114391748446615, + "learning_rate": 1.9238795325112868e-07, + "loss": 0.67, + "step": 3267 + }, + { + "epoch": 6.051851851851852, + "grad_norm": 1.6138123657063617, + "learning_rate": 1.9238221416628892e-07, + "loss": 0.6532, + "step": 3268 + }, + { + "epoch": 6.053703703703704, + "grad_norm": 1.4063975574167729, + "learning_rate": 1.923764730044461e-07, + "loss": 0.6801, + "step": 3269 + }, + { + "epoch": 6.055555555555555, + "grad_norm": 1.466317451244446, + "learning_rate": 1.923707297657293e-07, + "loss": 0.681, + "step": 3270 + }, + { + "epoch": 6.057407407407408, + "grad_norm": 2.0148717854543308, + "learning_rate": 1.9236498445026753e-07, + "loss": 0.7087, + "step": 3271 + }, + { + "epoch": 6.059259259259259, + "grad_norm": 1.6243257950685415, + "learning_rate": 1.9235923705819013e-07, + "loss": 0.6456, + "step": 3272 + }, + { + "epoch": 6.061111111111111, + "grad_norm": 1.6272678880330924, + "learning_rate": 1.9235348758962622e-07, + "loss": 0.6736, + "step": 3273 + }, + { + "epoch": 6.062962962962963, + "grad_norm": 2.1243603306086842, + "learning_rate": 1.9234773604470507e-07, + "loss": 0.6277, + "step": 3274 + }, + { + "epoch": 6.064814814814815, + "grad_norm": 1.4682004796994141, + "learning_rate": 1.9234198242355602e-07, + "loss": 0.7142, + "step": 3275 + }, + { + "epoch": 6.066666666666666, + "grad_norm": 1.4514682179896041, + "learning_rate": 1.923362267263084e-07, + "loss": 0.6753, + "step": 3276 + }, + { + "epoch": 6.068518518518519, + "grad_norm": 1.4264938055029956, + "learning_rate": 1.923304689530916e-07, + "loss": 0.6361, + "step": 3277 + }, + { + "epoch": 6.07037037037037, + "grad_norm": 1.3774428955600742, + "learning_rate": 1.923247091040351e-07, + "loss": 0.6927, + "step": 3278 + }, + { + "epoch": 6.072222222222222, + "grad_norm": 1.8117583015479766, + "learning_rate": 1.9231894717926842e-07, + "loss": 0.671, + "step": 3279 + }, + { + "epoch": 6.074074074074074, + "grad_norm": 1.3815959151207655, + "learning_rate": 1.9231318317892106e-07, + "loss": 0.6762, + "step": 3280 + }, + { + "epoch": 6.075925925925926, + "grad_norm": 1.519901756847425, + "learning_rate": 1.923074171031226e-07, + "loss": 0.7097, + "step": 3281 + }, + { + "epoch": 6.0777777777777775, + "grad_norm": 1.703388260101047, + "learning_rate": 1.9230164895200271e-07, + "loss": 0.6981, + "step": 3282 + }, + { + "epoch": 6.07962962962963, + "grad_norm": 1.3739243694967267, + "learning_rate": 1.9229587872569105e-07, + "loss": 0.6897, + "step": 3283 + }, + { + "epoch": 6.0814814814814815, + "grad_norm": 2.1783638495690867, + "learning_rate": 1.9229010642431742e-07, + "loss": 0.6794, + "step": 3284 + }, + { + "epoch": 6.083333333333333, + "grad_norm": 1.6497742497369867, + "learning_rate": 1.9228433204801148e-07, + "loss": 0.6836, + "step": 3285 + }, + { + "epoch": 6.0851851851851855, + "grad_norm": 2.0938996057325054, + "learning_rate": 1.9227855559690314e-07, + "loss": 0.6669, + "step": 3286 + }, + { + "epoch": 6.087037037037037, + "grad_norm": 1.4443119554837531, + "learning_rate": 1.9227277707112226e-07, + "loss": 0.6875, + "step": 3287 + }, + { + "epoch": 6.088888888888889, + "grad_norm": 3.909007861073493, + "learning_rate": 1.9226699647079872e-07, + "loss": 0.6662, + "step": 3288 + }, + { + "epoch": 6.090740740740741, + "grad_norm": 1.5774446082880966, + "learning_rate": 1.9226121379606254e-07, + "loss": 0.6785, + "step": 3289 + }, + { + "epoch": 6.092592592592593, + "grad_norm": 1.5570381030724414, + "learning_rate": 1.9225542904704365e-07, + "loss": 0.6661, + "step": 3290 + }, + { + "epoch": 6.094444444444444, + "grad_norm": 1.7722882775673867, + "learning_rate": 1.922496422238722e-07, + "loss": 0.6532, + "step": 3291 + }, + { + "epoch": 6.0962962962962965, + "grad_norm": 1.4153832125684849, + "learning_rate": 1.922438533266782e-07, + "loss": 0.7035, + "step": 3292 + }, + { + "epoch": 6.098148148148148, + "grad_norm": 2.0858320184451853, + "learning_rate": 1.9223806235559189e-07, + "loss": 0.647, + "step": 3293 + }, + { + "epoch": 6.1, + "grad_norm": 1.6764944847519188, + "learning_rate": 1.922322693107434e-07, + "loss": 0.6528, + "step": 3294 + }, + { + "epoch": 6.101851851851852, + "grad_norm": 1.7162558996793766, + "learning_rate": 1.9222647419226297e-07, + "loss": 0.6384, + "step": 3295 + }, + { + "epoch": 6.103703703703704, + "grad_norm": 1.6493952329606891, + "learning_rate": 1.9222067700028095e-07, + "loss": 0.6685, + "step": 3296 + }, + { + "epoch": 6.105555555555555, + "grad_norm": 1.481788648070295, + "learning_rate": 1.9221487773492764e-07, + "loss": 0.6973, + "step": 3297 + }, + { + "epoch": 6.107407407407408, + "grad_norm": 1.4495162455776365, + "learning_rate": 1.9220907639633342e-07, + "loss": 0.6299, + "step": 3298 + }, + { + "epoch": 6.109259259259259, + "grad_norm": 1.555265097779018, + "learning_rate": 1.9220327298462874e-07, + "loss": 0.7063, + "step": 3299 + }, + { + "epoch": 6.111111111111111, + "grad_norm": 1.6035271386899408, + "learning_rate": 1.9219746749994403e-07, + "loss": 0.6269, + "step": 3300 + }, + { + "epoch": 6.112962962962963, + "grad_norm": 1.3245827542119324, + "learning_rate": 1.9219165994240987e-07, + "loss": 0.695, + "step": 3301 + }, + { + "epoch": 6.114814814814815, + "grad_norm": 1.5082492443585795, + "learning_rate": 1.9218585031215682e-07, + "loss": 0.6433, + "step": 3302 + }, + { + "epoch": 6.116666666666666, + "grad_norm": 2.3193651821465053, + "learning_rate": 1.9218003860931548e-07, + "loss": 0.6656, + "step": 3303 + }, + { + "epoch": 6.118518518518519, + "grad_norm": 1.517198966669395, + "learning_rate": 1.921742248340165e-07, + "loss": 0.6111, + "step": 3304 + }, + { + "epoch": 6.12037037037037, + "grad_norm": 1.638136947163177, + "learning_rate": 1.921684089863906e-07, + "loss": 0.6501, + "step": 3305 + }, + { + "epoch": 6.122222222222222, + "grad_norm": 1.565969798221956, + "learning_rate": 1.9216259106656855e-07, + "loss": 0.6786, + "step": 3306 + }, + { + "epoch": 6.124074074074074, + "grad_norm": 1.7694339421775493, + "learning_rate": 1.9215677107468113e-07, + "loss": 0.6658, + "step": 3307 + }, + { + "epoch": 6.125925925925926, + "grad_norm": 1.6803687136826442, + "learning_rate": 1.921509490108592e-07, + "loss": 0.6685, + "step": 3308 + }, + { + "epoch": 6.127777777777778, + "grad_norm": 1.4359361649591924, + "learning_rate": 1.921451248752337e-07, + "loss": 0.6363, + "step": 3309 + }, + { + "epoch": 6.12962962962963, + "grad_norm": 1.4338033274768724, + "learning_rate": 1.921392986679355e-07, + "loss": 0.7184, + "step": 3310 + }, + { + "epoch": 6.131481481481481, + "grad_norm": 1.5341599626288873, + "learning_rate": 1.921334703890956e-07, + "loss": 0.6685, + "step": 3311 + }, + { + "epoch": 6.133333333333334, + "grad_norm": 1.7760408779864978, + "learning_rate": 1.921276400388451e-07, + "loss": 0.6626, + "step": 3312 + }, + { + "epoch": 6.135185185185185, + "grad_norm": 1.701511965565045, + "learning_rate": 1.92121807617315e-07, + "loss": 0.7086, + "step": 3313 + }, + { + "epoch": 6.137037037037037, + "grad_norm": 1.5488584755189525, + "learning_rate": 1.9211597312463648e-07, + "loss": 0.6388, + "step": 3314 + }, + { + "epoch": 6.138888888888889, + "grad_norm": 1.6314173807773145, + "learning_rate": 1.921101365609407e-07, + "loss": 0.6746, + "step": 3315 + }, + { + "epoch": 6.140740740740741, + "grad_norm": 1.8598083034126227, + "learning_rate": 1.9210429792635888e-07, + "loss": 0.684, + "step": 3316 + }, + { + "epoch": 6.142592592592592, + "grad_norm": 1.5424155430017175, + "learning_rate": 1.9209845722102228e-07, + "loss": 0.6425, + "step": 3317 + }, + { + "epoch": 6.144444444444445, + "grad_norm": 1.6110149182668472, + "learning_rate": 1.9209261444506227e-07, + "loss": 0.6346, + "step": 3318 + }, + { + "epoch": 6.146296296296296, + "grad_norm": 1.444019835423443, + "learning_rate": 1.9208676959861012e-07, + "loss": 0.6801, + "step": 3319 + }, + { + "epoch": 6.148148148148148, + "grad_norm": 1.6515702559174343, + "learning_rate": 1.920809226817973e-07, + "loss": 0.672, + "step": 3320 + }, + { + "epoch": 6.15, + "grad_norm": 1.7143028084078356, + "learning_rate": 1.9207507369475527e-07, + "loss": 0.6689, + "step": 3321 + }, + { + "epoch": 6.151851851851852, + "grad_norm": 1.51699561976835, + "learning_rate": 1.920692226376155e-07, + "loss": 0.6571, + "step": 3322 + }, + { + "epoch": 6.1537037037037035, + "grad_norm": 1.3188794834709554, + "learning_rate": 1.9206336951050952e-07, + "loss": 0.6946, + "step": 3323 + }, + { + "epoch": 6.155555555555556, + "grad_norm": 1.7836721402771154, + "learning_rate": 1.9205751431356897e-07, + "loss": 0.6682, + "step": 3324 + }, + { + "epoch": 6.157407407407407, + "grad_norm": 1.695294532941119, + "learning_rate": 1.920516570469255e-07, + "loss": 0.6956, + "step": 3325 + }, + { + "epoch": 6.159259259259259, + "grad_norm": 1.7260113051081005, + "learning_rate": 1.9204579771071075e-07, + "loss": 0.6461, + "step": 3326 + }, + { + "epoch": 6.161111111111111, + "grad_norm": 1.4013776990874338, + "learning_rate": 1.9203993630505648e-07, + "loss": 0.6735, + "step": 3327 + }, + { + "epoch": 6.162962962962963, + "grad_norm": 1.606015622454816, + "learning_rate": 1.9203407283009444e-07, + "loss": 0.67, + "step": 3328 + }, + { + "epoch": 6.1648148148148145, + "grad_norm": 3.0184006279633704, + "learning_rate": 1.920282072859565e-07, + "loss": 0.6759, + "step": 3329 + }, + { + "epoch": 6.166666666666667, + "grad_norm": 1.6341492547872276, + "learning_rate": 1.9202233967277454e-07, + "loss": 0.6684, + "step": 3330 + }, + { + "epoch": 6.1685185185185185, + "grad_norm": 1.5650332345731641, + "learning_rate": 1.9201646999068042e-07, + "loss": 0.6342, + "step": 3331 + }, + { + "epoch": 6.17037037037037, + "grad_norm": 1.5776098419792035, + "learning_rate": 1.9201059823980615e-07, + "loss": 0.6507, + "step": 3332 + }, + { + "epoch": 6.1722222222222225, + "grad_norm": 1.3912193707984237, + "learning_rate": 1.9200472442028375e-07, + "loss": 0.6561, + "step": 3333 + }, + { + "epoch": 6.174074074074074, + "grad_norm": 1.4712625538489306, + "learning_rate": 1.9199884853224525e-07, + "loss": 0.6567, + "step": 3334 + }, + { + "epoch": 6.175925925925926, + "grad_norm": 1.7647268734713886, + "learning_rate": 1.9199297057582278e-07, + "loss": 0.6435, + "step": 3335 + }, + { + "epoch": 6.177777777777778, + "grad_norm": 2.1072799249397223, + "learning_rate": 1.9198709055114848e-07, + "loss": 0.6861, + "step": 3336 + }, + { + "epoch": 6.17962962962963, + "grad_norm": 1.4598892021068544, + "learning_rate": 1.9198120845835453e-07, + "loss": 0.6537, + "step": 3337 + }, + { + "epoch": 6.181481481481481, + "grad_norm": 1.9817834624125323, + "learning_rate": 1.9197532429757324e-07, + "loss": 0.6801, + "step": 3338 + }, + { + "epoch": 6.183333333333334, + "grad_norm": 1.5218399752139922, + "learning_rate": 1.9196943806893683e-07, + "loss": 0.6722, + "step": 3339 + }, + { + "epoch": 6.185185185185185, + "grad_norm": 1.5618825449371072, + "learning_rate": 1.9196354977257766e-07, + "loss": 0.6953, + "step": 3340 + }, + { + "epoch": 6.187037037037037, + "grad_norm": 1.4268630152187416, + "learning_rate": 1.919576594086281e-07, + "loss": 0.6756, + "step": 3341 + }, + { + "epoch": 6.188888888888889, + "grad_norm": 1.5081163502280814, + "learning_rate": 1.9195176697722066e-07, + "loss": 0.6499, + "step": 3342 + }, + { + "epoch": 6.190740740740741, + "grad_norm": 1.4403414514101327, + "learning_rate": 1.9194587247848775e-07, + "loss": 0.6882, + "step": 3343 + }, + { + "epoch": 6.192592592592592, + "grad_norm": 1.4113805168701803, + "learning_rate": 1.9193997591256188e-07, + "loss": 0.702, + "step": 3344 + }, + { + "epoch": 6.194444444444445, + "grad_norm": 1.537333510024604, + "learning_rate": 1.9193407727957566e-07, + "loss": 0.6628, + "step": 3345 + }, + { + "epoch": 6.196296296296296, + "grad_norm": 1.8040425820769224, + "learning_rate": 1.919281765796617e-07, + "loss": 0.632, + "step": 3346 + }, + { + "epoch": 6.198148148148148, + "grad_norm": 1.442920474064865, + "learning_rate": 1.9192227381295263e-07, + "loss": 0.6866, + "step": 3347 + }, + { + "epoch": 6.2, + "grad_norm": 1.633535299852807, + "learning_rate": 1.919163689795812e-07, + "loss": 0.6948, + "step": 3348 + }, + { + "epoch": 6.201851851851852, + "grad_norm": 1.5353752964391654, + "learning_rate": 1.919104620796802e-07, + "loss": 0.6664, + "step": 3349 + }, + { + "epoch": 6.203703703703703, + "grad_norm": 1.306342887143042, + "learning_rate": 1.919045531133823e-07, + "loss": 0.6205, + "step": 3350 + }, + { + "epoch": 6.205555555555556, + "grad_norm": 1.642092118730288, + "learning_rate": 1.9189864208082048e-07, + "loss": 0.6901, + "step": 3351 + }, + { + "epoch": 6.207407407407407, + "grad_norm": 1.5140846820955263, + "learning_rate": 1.9189272898212764e-07, + "loss": 0.6551, + "step": 3352 + }, + { + "epoch": 6.209259259259259, + "grad_norm": 1.8087419917613925, + "learning_rate": 1.9188681381743662e-07, + "loss": 0.671, + "step": 3353 + }, + { + "epoch": 6.211111111111111, + "grad_norm": 1.5323281068740797, + "learning_rate": 1.9188089658688048e-07, + "loss": 0.6753, + "step": 3354 + }, + { + "epoch": 6.212962962962963, + "grad_norm": 1.4914873329476706, + "learning_rate": 1.9187497729059224e-07, + "loss": 0.6522, + "step": 3355 + }, + { + "epoch": 6.214814814814815, + "grad_norm": 1.5152457738013503, + "learning_rate": 1.9186905592870497e-07, + "loss": 0.6804, + "step": 3356 + }, + { + "epoch": 6.216666666666667, + "grad_norm": 1.6805050649653726, + "learning_rate": 1.9186313250135182e-07, + "loss": 0.7242, + "step": 3357 + }, + { + "epoch": 6.218518518518518, + "grad_norm": 2.016589350831726, + "learning_rate": 1.9185720700866597e-07, + "loss": 0.6773, + "step": 3358 + }, + { + "epoch": 6.220370370370371, + "grad_norm": 1.5867126132779024, + "learning_rate": 1.9185127945078058e-07, + "loss": 0.629, + "step": 3359 + }, + { + "epoch": 6.222222222222222, + "grad_norm": 2.555724963147106, + "learning_rate": 1.91845349827829e-07, + "loss": 0.6583, + "step": 3360 + }, + { + "epoch": 6.224074074074074, + "grad_norm": 1.4505794963276681, + "learning_rate": 1.9183941813994454e-07, + "loss": 0.6393, + "step": 3361 + }, + { + "epoch": 6.225925925925926, + "grad_norm": 5.767031031874181, + "learning_rate": 1.9183348438726048e-07, + "loss": 0.6557, + "step": 3362 + }, + { + "epoch": 6.227777777777778, + "grad_norm": 1.61949811131035, + "learning_rate": 1.9182754856991028e-07, + "loss": 0.6476, + "step": 3363 + }, + { + "epoch": 6.229629629629629, + "grad_norm": 1.410102440854385, + "learning_rate": 1.918216106880274e-07, + "loss": 0.6442, + "step": 3364 + }, + { + "epoch": 6.231481481481482, + "grad_norm": 1.5828107819617578, + "learning_rate": 1.9181567074174532e-07, + "loss": 0.6822, + "step": 3365 + }, + { + "epoch": 6.233333333333333, + "grad_norm": 1.4482426048154045, + "learning_rate": 1.9180972873119758e-07, + "loss": 0.6839, + "step": 3366 + }, + { + "epoch": 6.235185185185185, + "grad_norm": 1.5254398064451258, + "learning_rate": 1.9180378465651782e-07, + "loss": 0.6432, + "step": 3367 + }, + { + "epoch": 6.237037037037037, + "grad_norm": 1.5712406047304206, + "learning_rate": 1.9179783851783963e-07, + "loss": 0.6447, + "step": 3368 + }, + { + "epoch": 6.238888888888889, + "grad_norm": 1.6101021332698338, + "learning_rate": 1.917918903152967e-07, + "loss": 0.6204, + "step": 3369 + }, + { + "epoch": 6.2407407407407405, + "grad_norm": 2.074215443583755, + "learning_rate": 1.917859400490228e-07, + "loss": 0.6996, + "step": 3370 + }, + { + "epoch": 6.242592592592593, + "grad_norm": 1.5712661098611045, + "learning_rate": 1.9177998771915164e-07, + "loss": 0.6581, + "step": 3371 + }, + { + "epoch": 6.2444444444444445, + "grad_norm": 1.6320109272099848, + "learning_rate": 1.9177403332581712e-07, + "loss": 0.64, + "step": 3372 + }, + { + "epoch": 6.246296296296296, + "grad_norm": 1.4119372007155595, + "learning_rate": 1.9176807686915307e-07, + "loss": 0.7032, + "step": 3373 + }, + { + "epoch": 6.2481481481481485, + "grad_norm": 1.5996084972469322, + "learning_rate": 1.917621183492934e-07, + "loss": 0.6708, + "step": 3374 + }, + { + "epoch": 6.25, + "grad_norm": 1.5482380105102012, + "learning_rate": 1.917561577663721e-07, + "loss": 0.6754, + "step": 3375 + }, + { + "epoch": 6.2518518518518515, + "grad_norm": 1.6278065338317644, + "learning_rate": 1.9175019512052315e-07, + "loss": 0.6495, + "step": 3376 + }, + { + "epoch": 6.253703703703704, + "grad_norm": 1.427443309089482, + "learning_rate": 1.9174423041188064e-07, + "loss": 0.7212, + "step": 3377 + }, + { + "epoch": 6.2555555555555555, + "grad_norm": 1.5541245674329187, + "learning_rate": 1.9173826364057867e-07, + "loss": 0.6534, + "step": 3378 + }, + { + "epoch": 6.257407407407407, + "grad_norm": 1.4798706511108735, + "learning_rate": 1.9173229480675136e-07, + "loss": 0.6872, + "step": 3379 + }, + { + "epoch": 6.2592592592592595, + "grad_norm": 1.5741355778396606, + "learning_rate": 1.9172632391053293e-07, + "loss": 0.6242, + "step": 3380 + }, + { + "epoch": 6.261111111111111, + "grad_norm": 2.1736452190263127, + "learning_rate": 1.9172035095205762e-07, + "loss": 0.6251, + "step": 3381 + }, + { + "epoch": 6.262962962962963, + "grad_norm": 1.5680883386519138, + "learning_rate": 1.9171437593145966e-07, + "loss": 0.6979, + "step": 3382 + }, + { + "epoch": 6.264814814814815, + "grad_norm": 1.7182523419892173, + "learning_rate": 1.917083988488735e-07, + "loss": 0.672, + "step": 3383 + }, + { + "epoch": 6.266666666666667, + "grad_norm": 2.085816581645039, + "learning_rate": 1.917024197044334e-07, + "loss": 0.6975, + "step": 3384 + }, + { + "epoch": 6.268518518518518, + "grad_norm": 1.3934471906208208, + "learning_rate": 1.916964384982739e-07, + "loss": 0.7055, + "step": 3385 + }, + { + "epoch": 6.270370370370371, + "grad_norm": 1.4409719661635803, + "learning_rate": 1.9169045523052941e-07, + "loss": 0.6744, + "step": 3386 + }, + { + "epoch": 6.272222222222222, + "grad_norm": 1.9071123625518116, + "learning_rate": 1.9168446990133442e-07, + "loss": 0.6614, + "step": 3387 + }, + { + "epoch": 6.274074074074074, + "grad_norm": 2.5389267641304727, + "learning_rate": 1.9167848251082358e-07, + "loss": 0.6764, + "step": 3388 + }, + { + "epoch": 6.275925925925926, + "grad_norm": 1.3045398433591127, + "learning_rate": 1.9167249305913144e-07, + "loss": 0.6419, + "step": 3389 + }, + { + "epoch": 6.277777777777778, + "grad_norm": 1.4265062862635696, + "learning_rate": 1.9166650154639267e-07, + "loss": 0.655, + "step": 3390 + }, + { + "epoch": 6.279629629629629, + "grad_norm": 1.4740999658806446, + "learning_rate": 1.9166050797274198e-07, + "loss": 0.7062, + "step": 3391 + }, + { + "epoch": 6.281481481481482, + "grad_norm": 2.99327903327846, + "learning_rate": 1.9165451233831413e-07, + "loss": 0.6196, + "step": 3392 + }, + { + "epoch": 6.283333333333333, + "grad_norm": 1.489313161249133, + "learning_rate": 1.9164851464324392e-07, + "loss": 0.6732, + "step": 3393 + }, + { + "epoch": 6.285185185185185, + "grad_norm": 1.8268014928067233, + "learning_rate": 1.9164251488766618e-07, + "loss": 0.6937, + "step": 3394 + }, + { + "epoch": 6.287037037037037, + "grad_norm": 1.4010652760287288, + "learning_rate": 1.916365130717158e-07, + "loss": 0.6728, + "step": 3395 + }, + { + "epoch": 6.288888888888889, + "grad_norm": 1.4955383789510575, + "learning_rate": 1.9163050919552776e-07, + "loss": 0.7062, + "step": 3396 + }, + { + "epoch": 6.290740740740741, + "grad_norm": 1.8298734868077546, + "learning_rate": 1.9162450325923697e-07, + "loss": 0.6779, + "step": 3397 + }, + { + "epoch": 6.292592592592593, + "grad_norm": 1.3846873769370864, + "learning_rate": 1.9161849526297852e-07, + "loss": 0.709, + "step": 3398 + }, + { + "epoch": 6.294444444444444, + "grad_norm": 1.659133186209603, + "learning_rate": 1.9161248520688745e-07, + "loss": 0.6388, + "step": 3399 + }, + { + "epoch": 6.296296296296296, + "grad_norm": 1.5794134707910168, + "learning_rate": 1.916064730910989e-07, + "loss": 0.6755, + "step": 3400 + }, + { + "epoch": 6.298148148148148, + "grad_norm": 1.417430610217626, + "learning_rate": 1.9160045891574806e-07, + "loss": 0.655, + "step": 3401 + }, + { + "epoch": 6.3, + "grad_norm": 1.669536168966088, + "learning_rate": 1.915944426809701e-07, + "loss": 0.6259, + "step": 3402 + }, + { + "epoch": 6.301851851851852, + "grad_norm": 1.6540145764154959, + "learning_rate": 1.9158842438690028e-07, + "loss": 0.7047, + "step": 3403 + }, + { + "epoch": 6.303703703703704, + "grad_norm": 1.581507051187245, + "learning_rate": 1.9158240403367396e-07, + "loss": 0.6847, + "step": 3404 + }, + { + "epoch": 6.305555555555555, + "grad_norm": 1.8494292893526751, + "learning_rate": 1.9157638162142646e-07, + "loss": 0.6586, + "step": 3405 + }, + { + "epoch": 6.307407407407408, + "grad_norm": 1.6001401114536582, + "learning_rate": 1.9157035715029317e-07, + "loss": 0.6206, + "step": 3406 + }, + { + "epoch": 6.309259259259259, + "grad_norm": 1.9812505126575766, + "learning_rate": 1.9156433062040956e-07, + "loss": 0.6366, + "step": 3407 + }, + { + "epoch": 6.311111111111111, + "grad_norm": 2.277901288785482, + "learning_rate": 1.915583020319111e-07, + "loss": 0.6157, + "step": 3408 + }, + { + "epoch": 6.312962962962963, + "grad_norm": 1.707848761298398, + "learning_rate": 1.9155227138493335e-07, + "loss": 0.6422, + "step": 3409 + }, + { + "epoch": 6.314814814814815, + "grad_norm": 1.6746797176861448, + "learning_rate": 1.915462386796119e-07, + "loss": 0.6668, + "step": 3410 + }, + { + "epoch": 6.316666666666666, + "grad_norm": 1.6696580726154555, + "learning_rate": 1.9154020391608236e-07, + "loss": 0.672, + "step": 3411 + }, + { + "epoch": 6.318518518518519, + "grad_norm": 1.8529384731080532, + "learning_rate": 1.9153416709448043e-07, + "loss": 0.6763, + "step": 3412 + }, + { + "epoch": 6.32037037037037, + "grad_norm": 1.6689760975037833, + "learning_rate": 1.915281282149418e-07, + "loss": 0.6466, + "step": 3413 + }, + { + "epoch": 6.322222222222222, + "grad_norm": 1.426932617951893, + "learning_rate": 1.9152208727760227e-07, + "loss": 0.6564, + "step": 3414 + }, + { + "epoch": 6.324074074074074, + "grad_norm": 2.129197528045345, + "learning_rate": 1.9151604428259765e-07, + "loss": 0.6852, + "step": 3415 + }, + { + "epoch": 6.325925925925926, + "grad_norm": 1.785447640339465, + "learning_rate": 1.9150999923006378e-07, + "loss": 0.6749, + "step": 3416 + }, + { + "epoch": 6.3277777777777775, + "grad_norm": 1.4903525606466723, + "learning_rate": 1.9150395212013662e-07, + "loss": 0.6073, + "step": 3417 + }, + { + "epoch": 6.32962962962963, + "grad_norm": 1.4565758031313258, + "learning_rate": 1.9149790295295206e-07, + "loss": 0.6496, + "step": 3418 + }, + { + "epoch": 6.3314814814814815, + "grad_norm": 1.6642606339460744, + "learning_rate": 1.9149185172864616e-07, + "loss": 0.6908, + "step": 3419 + }, + { + "epoch": 6.333333333333333, + "grad_norm": 1.5488521842397405, + "learning_rate": 1.9148579844735495e-07, + "loss": 0.6404, + "step": 3420 + }, + { + "epoch": 6.3351851851851855, + "grad_norm": 1.497859446088985, + "learning_rate": 1.914797431092145e-07, + "loss": 0.6602, + "step": 3421 + }, + { + "epoch": 6.337037037037037, + "grad_norm": 1.6853275632443632, + "learning_rate": 1.9147368571436097e-07, + "loss": 0.6597, + "step": 3422 + }, + { + "epoch": 6.338888888888889, + "grad_norm": 1.698456252837249, + "learning_rate": 1.914676262629306e-07, + "loss": 0.5796, + "step": 3423 + }, + { + "epoch": 6.340740740740741, + "grad_norm": 1.56287755830112, + "learning_rate": 1.9146156475505952e-07, + "loss": 0.684, + "step": 3424 + }, + { + "epoch": 6.342592592592593, + "grad_norm": 1.5510770876196125, + "learning_rate": 1.9145550119088404e-07, + "loss": 0.6434, + "step": 3425 + }, + { + "epoch": 6.344444444444444, + "grad_norm": 1.3610545902526667, + "learning_rate": 1.9144943557054053e-07, + "loss": 0.5937, + "step": 3426 + }, + { + "epoch": 6.3462962962962965, + "grad_norm": 1.4976729337028059, + "learning_rate": 1.9144336789416534e-07, + "loss": 0.6803, + "step": 3427 + }, + { + "epoch": 6.348148148148148, + "grad_norm": 1.5556266803670558, + "learning_rate": 1.9143729816189488e-07, + "loss": 0.6959, + "step": 3428 + }, + { + "epoch": 6.35, + "grad_norm": 1.833288957501599, + "learning_rate": 1.9143122637386564e-07, + "loss": 0.6675, + "step": 3429 + }, + { + "epoch": 6.351851851851852, + "grad_norm": 1.7097929956779876, + "learning_rate": 1.9142515253021404e-07, + "loss": 0.6574, + "step": 3430 + }, + { + "epoch": 6.353703703703704, + "grad_norm": 1.4826663772887276, + "learning_rate": 1.9141907663107677e-07, + "loss": 0.6822, + "step": 3431 + }, + { + "epoch": 6.355555555555555, + "grad_norm": 1.4284192455158202, + "learning_rate": 1.9141299867659034e-07, + "loss": 0.6872, + "step": 3432 + }, + { + "epoch": 6.357407407407408, + "grad_norm": 1.6063410989501834, + "learning_rate": 1.9140691866689144e-07, + "loss": 0.6284, + "step": 3433 + }, + { + "epoch": 6.359259259259259, + "grad_norm": 1.3710009448688145, + "learning_rate": 1.9140083660211675e-07, + "loss": 0.6845, + "step": 3434 + }, + { + "epoch": 6.361111111111111, + "grad_norm": 1.37125158829942, + "learning_rate": 1.91394752482403e-07, + "loss": 0.6778, + "step": 3435 + }, + { + "epoch": 6.362962962962963, + "grad_norm": 1.7920836689663495, + "learning_rate": 1.91388666307887e-07, + "loss": 0.6614, + "step": 3436 + }, + { + "epoch": 6.364814814814815, + "grad_norm": 1.6169473332441728, + "learning_rate": 1.9138257807870555e-07, + "loss": 0.6898, + "step": 3437 + }, + { + "epoch": 6.366666666666666, + "grad_norm": 1.4511184099836392, + "learning_rate": 1.9137648779499558e-07, + "loss": 0.6488, + "step": 3438 + }, + { + "epoch": 6.368518518518519, + "grad_norm": 3.2756727132891124, + "learning_rate": 1.9137039545689397e-07, + "loss": 0.6264, + "step": 3439 + }, + { + "epoch": 6.37037037037037, + "grad_norm": 1.6320513703683337, + "learning_rate": 1.9136430106453774e-07, + "loss": 0.6493, + "step": 3440 + }, + { + "epoch": 6.372222222222222, + "grad_norm": 1.372712142012325, + "learning_rate": 1.9135820461806386e-07, + "loss": 0.6551, + "step": 3441 + }, + { + "epoch": 6.374074074074074, + "grad_norm": 1.4233008427160678, + "learning_rate": 1.913521061176094e-07, + "loss": 0.6752, + "step": 3442 + }, + { + "epoch": 6.375925925925926, + "grad_norm": 1.60618672177728, + "learning_rate": 1.913460055633115e-07, + "loss": 0.6168, + "step": 3443 + }, + { + "epoch": 6.377777777777778, + "grad_norm": 1.763724534295357, + "learning_rate": 1.9133990295530732e-07, + "loss": 0.6238, + "step": 3444 + }, + { + "epoch": 6.37962962962963, + "grad_norm": 1.4017229732786478, + "learning_rate": 1.9133379829373405e-07, + "loss": 0.6375, + "step": 3445 + }, + { + "epoch": 6.381481481481481, + "grad_norm": 2.3358142122918957, + "learning_rate": 1.913276915787289e-07, + "loss": 0.6765, + "step": 3446 + }, + { + "epoch": 6.383333333333334, + "grad_norm": 1.5248172607926531, + "learning_rate": 1.9132158281042923e-07, + "loss": 0.6469, + "step": 3447 + }, + { + "epoch": 6.385185185185185, + "grad_norm": 1.5012034867503168, + "learning_rate": 1.9131547198897233e-07, + "loss": 0.6709, + "step": 3448 + }, + { + "epoch": 6.387037037037037, + "grad_norm": 2.0285030171956313, + "learning_rate": 1.9130935911449563e-07, + "loss": 0.6705, + "step": 3449 + }, + { + "epoch": 6.388888888888889, + "grad_norm": 1.4929696543172324, + "learning_rate": 1.9130324418713654e-07, + "loss": 0.5815, + "step": 3450 + }, + { + "epoch": 6.390740740740741, + "grad_norm": 1.748207755802049, + "learning_rate": 1.9129712720703254e-07, + "loss": 0.6531, + "step": 3451 + }, + { + "epoch": 6.392592592592592, + "grad_norm": 1.6102653597159946, + "learning_rate": 1.9129100817432115e-07, + "loss": 0.6631, + "step": 3452 + }, + { + "epoch": 6.394444444444445, + "grad_norm": 1.4354879816857333, + "learning_rate": 1.9128488708913997e-07, + "loss": 0.6617, + "step": 3453 + }, + { + "epoch": 6.396296296296296, + "grad_norm": 1.5323256088910597, + "learning_rate": 1.912787639516266e-07, + "loss": 0.6822, + "step": 3454 + }, + { + "epoch": 6.398148148148148, + "grad_norm": 1.975147709669045, + "learning_rate": 1.9127263876191871e-07, + "loss": 0.6353, + "step": 3455 + }, + { + "epoch": 6.4, + "grad_norm": 1.7335395939795983, + "learning_rate": 1.91266511520154e-07, + "loss": 0.6685, + "step": 3456 + }, + { + "epoch": 6.401851851851852, + "grad_norm": 1.470428367473061, + "learning_rate": 1.9126038222647024e-07, + "loss": 0.7029, + "step": 3457 + }, + { + "epoch": 6.4037037037037035, + "grad_norm": 1.6491674131842289, + "learning_rate": 1.9125425088100525e-07, + "loss": 0.6468, + "step": 3458 + }, + { + "epoch": 6.405555555555556, + "grad_norm": 1.7837052191056182, + "learning_rate": 1.9124811748389681e-07, + "loss": 0.6621, + "step": 3459 + }, + { + "epoch": 6.407407407407407, + "grad_norm": 1.4773470602095442, + "learning_rate": 1.912419820352829e-07, + "loss": 0.6694, + "step": 3460 + }, + { + "epoch": 6.409259259259259, + "grad_norm": 2.41615950812344, + "learning_rate": 1.9123584453530142e-07, + "loss": 0.6541, + "step": 3461 + }, + { + "epoch": 6.411111111111111, + "grad_norm": 1.6551929261096554, + "learning_rate": 1.9122970498409035e-07, + "loss": 0.7091, + "step": 3462 + }, + { + "epoch": 6.412962962962963, + "grad_norm": 1.6496518259984632, + "learning_rate": 1.9122356338178773e-07, + "loss": 0.6238, + "step": 3463 + }, + { + "epoch": 6.4148148148148145, + "grad_norm": 2.4245246179191136, + "learning_rate": 1.9121741972853168e-07, + "loss": 0.64, + "step": 3464 + }, + { + "epoch": 6.416666666666667, + "grad_norm": 1.4971063033915384, + "learning_rate": 1.9121127402446024e-07, + "loss": 0.6356, + "step": 3465 + }, + { + "epoch": 6.4185185185185185, + "grad_norm": 1.6169379581882568, + "learning_rate": 1.9120512626971168e-07, + "loss": 0.6235, + "step": 3466 + }, + { + "epoch": 6.42037037037037, + "grad_norm": 1.4174126049171334, + "learning_rate": 1.9119897646442412e-07, + "loss": 0.6566, + "step": 3467 + }, + { + "epoch": 6.4222222222222225, + "grad_norm": 1.6776269916317015, + "learning_rate": 1.9119282460873593e-07, + "loss": 0.6309, + "step": 3468 + }, + { + "epoch": 6.424074074074074, + "grad_norm": 1.3895209628854723, + "learning_rate": 1.9118667070278532e-07, + "loss": 0.6894, + "step": 3469 + }, + { + "epoch": 6.425925925925926, + "grad_norm": 1.69209431763122, + "learning_rate": 1.9118051474671074e-07, + "loss": 0.6196, + "step": 3470 + }, + { + "epoch": 6.427777777777778, + "grad_norm": 1.8103124746368358, + "learning_rate": 1.9117435674065052e-07, + "loss": 0.6652, + "step": 3471 + }, + { + "epoch": 6.42962962962963, + "grad_norm": 1.7201860399694917, + "learning_rate": 1.9116819668474316e-07, + "loss": 0.6523, + "step": 3472 + }, + { + "epoch": 6.431481481481481, + "grad_norm": 1.5894130993481135, + "learning_rate": 1.9116203457912709e-07, + "loss": 0.6768, + "step": 3473 + }, + { + "epoch": 6.433333333333334, + "grad_norm": 1.43929195989301, + "learning_rate": 1.9115587042394092e-07, + "loss": 0.6382, + "step": 3474 + }, + { + "epoch": 6.435185185185185, + "grad_norm": 1.8346544278283612, + "learning_rate": 1.9114970421932322e-07, + "loss": 0.6961, + "step": 3475 + }, + { + "epoch": 6.437037037037037, + "grad_norm": 1.7034550503509098, + "learning_rate": 1.9114353596541254e-07, + "loss": 0.6615, + "step": 3476 + }, + { + "epoch": 6.438888888888889, + "grad_norm": 1.3651589031920115, + "learning_rate": 1.911373656623477e-07, + "loss": 0.6859, + "step": 3477 + }, + { + "epoch": 6.440740740740741, + "grad_norm": 1.6360328518097047, + "learning_rate": 1.9113119331026733e-07, + "loss": 0.6176, + "step": 3478 + }, + { + "epoch": 6.442592592592592, + "grad_norm": 1.5022692946560237, + "learning_rate": 1.9112501890931022e-07, + "loss": 0.6492, + "step": 3479 + }, + { + "epoch": 6.444444444444445, + "grad_norm": 1.6584397649180205, + "learning_rate": 1.911188424596152e-07, + "loss": 0.7087, + "step": 3480 + }, + { + "epoch": 6.446296296296296, + "grad_norm": 3.9558014621885182, + "learning_rate": 1.9111266396132116e-07, + "loss": 0.6575, + "step": 3481 + }, + { + "epoch": 6.448148148148148, + "grad_norm": 1.491705065355962, + "learning_rate": 1.9110648341456693e-07, + "loss": 0.6141, + "step": 3482 + }, + { + "epoch": 6.45, + "grad_norm": 1.8074815983797556, + "learning_rate": 1.9110030081949153e-07, + "loss": 0.6846, + "step": 3483 + }, + { + "epoch": 6.451851851851852, + "grad_norm": 1.568925684188308, + "learning_rate": 1.9109411617623391e-07, + "loss": 0.6168, + "step": 3484 + }, + { + "epoch": 6.453703703703704, + "grad_norm": 1.3328503187688294, + "learning_rate": 1.910879294849332e-07, + "loss": 0.6754, + "step": 3485 + }, + { + "epoch": 6.455555555555556, + "grad_norm": 1.339476004343173, + "learning_rate": 1.9108174074572842e-07, + "loss": 0.6873, + "step": 3486 + }, + { + "epoch": 6.457407407407407, + "grad_norm": 1.6843423899888483, + "learning_rate": 1.9107554995875875e-07, + "loss": 0.6704, + "step": 3487 + }, + { + "epoch": 6.459259259259259, + "grad_norm": 1.4159829128260302, + "learning_rate": 1.9106935712416333e-07, + "loss": 0.6825, + "step": 3488 + }, + { + "epoch": 6.461111111111111, + "grad_norm": 1.4616694142510485, + "learning_rate": 1.9106316224208143e-07, + "loss": 0.5907, + "step": 3489 + }, + { + "epoch": 6.462962962962963, + "grad_norm": 1.4261532937922716, + "learning_rate": 1.9105696531265235e-07, + "loss": 0.647, + "step": 3490 + }, + { + "epoch": 6.464814814814815, + "grad_norm": 1.4342219018111817, + "learning_rate": 1.9105076633601533e-07, + "loss": 0.6576, + "step": 3491 + }, + { + "epoch": 6.466666666666667, + "grad_norm": 1.3907042454793173, + "learning_rate": 1.9104456531230985e-07, + "loss": 0.6505, + "step": 3492 + }, + { + "epoch": 6.468518518518518, + "grad_norm": 1.4112376101651003, + "learning_rate": 1.910383622416752e-07, + "loss": 0.6942, + "step": 3493 + }, + { + "epoch": 6.470370370370371, + "grad_norm": 1.5109287401691613, + "learning_rate": 1.9103215712425097e-07, + "loss": 0.6151, + "step": 3494 + }, + { + "epoch": 6.472222222222222, + "grad_norm": 1.552618521467297, + "learning_rate": 1.910259499601766e-07, + "loss": 0.617, + "step": 3495 + }, + { + "epoch": 6.474074074074074, + "grad_norm": 1.8608266173216272, + "learning_rate": 1.9101974074959164e-07, + "loss": 0.6969, + "step": 3496 + }, + { + "epoch": 6.475925925925926, + "grad_norm": 1.37219659410563, + "learning_rate": 1.910135294926357e-07, + "loss": 0.6549, + "step": 3497 + }, + { + "epoch": 6.477777777777778, + "grad_norm": 1.4335034329481964, + "learning_rate": 1.9100731618944844e-07, + "loss": 0.6518, + "step": 3498 + }, + { + "epoch": 6.479629629629629, + "grad_norm": 1.4625744892098476, + "learning_rate": 1.9100110084016952e-07, + "loss": 0.6817, + "step": 3499 + }, + { + "epoch": 6.481481481481482, + "grad_norm": 1.4956408438940878, + "learning_rate": 1.909948834449387e-07, + "loss": 0.6449, + "step": 3500 + }, + { + "epoch": 6.483333333333333, + "grad_norm": 1.496066186014099, + "learning_rate": 1.9098866400389578e-07, + "loss": 0.7182, + "step": 3501 + }, + { + "epoch": 6.485185185185185, + "grad_norm": 1.9098552202998413, + "learning_rate": 1.9098244251718055e-07, + "loss": 0.6873, + "step": 3502 + }, + { + "epoch": 6.487037037037037, + "grad_norm": 1.6937796441403588, + "learning_rate": 1.9097621898493293e-07, + "loss": 0.6474, + "step": 3503 + }, + { + "epoch": 6.488888888888889, + "grad_norm": 1.5185315334522649, + "learning_rate": 1.909699934072928e-07, + "loss": 0.6832, + "step": 3504 + }, + { + "epoch": 6.4907407407407405, + "grad_norm": 1.5695052033692563, + "learning_rate": 1.9096376578440016e-07, + "loss": 0.6771, + "step": 3505 + }, + { + "epoch": 6.492592592592593, + "grad_norm": 1.3800690951375854, + "learning_rate": 1.90957536116395e-07, + "loss": 0.6784, + "step": 3506 + }, + { + "epoch": 6.4944444444444445, + "grad_norm": 1.544958976020382, + "learning_rate": 1.909513044034174e-07, + "loss": 0.6474, + "step": 3507 + }, + { + "epoch": 6.496296296296296, + "grad_norm": 1.5256162306385321, + "learning_rate": 1.9094507064560746e-07, + "loss": 0.645, + "step": 3508 + }, + { + "epoch": 6.4981481481481485, + "grad_norm": 1.3915880481965817, + "learning_rate": 1.9093883484310533e-07, + "loss": 0.6614, + "step": 3509 + }, + { + "epoch": 6.5, + "grad_norm": 1.378096839744085, + "learning_rate": 1.909325969960512e-07, + "loss": 0.6216, + "step": 3510 + }, + { + "epoch": 6.5018518518518515, + "grad_norm": 1.6025282410958077, + "learning_rate": 1.9092635710458533e-07, + "loss": 0.6386, + "step": 3511 + }, + { + "epoch": 6.503703703703704, + "grad_norm": 1.5917477300156002, + "learning_rate": 1.9092011516884798e-07, + "loss": 0.6416, + "step": 3512 + }, + { + "epoch": 6.5055555555555555, + "grad_norm": 1.6282878708885695, + "learning_rate": 1.9091387118897953e-07, + "loss": 0.6758, + "step": 3513 + }, + { + "epoch": 6.507407407407407, + "grad_norm": 1.311585953143118, + "learning_rate": 1.909076251651203e-07, + "loss": 0.6643, + "step": 3514 + }, + { + "epoch": 6.5092592592592595, + "grad_norm": 1.5683270155089606, + "learning_rate": 1.9090137709741081e-07, + "loss": 0.634, + "step": 3515 + }, + { + "epoch": 6.511111111111111, + "grad_norm": 1.7821499159645384, + "learning_rate": 1.9089512698599145e-07, + "loss": 0.6604, + "step": 3516 + }, + { + "epoch": 6.512962962962963, + "grad_norm": 1.3659166617388496, + "learning_rate": 1.9088887483100277e-07, + "loss": 0.646, + "step": 3517 + }, + { + "epoch": 6.514814814814815, + "grad_norm": 1.52707141954496, + "learning_rate": 1.908826206325853e-07, + "loss": 0.621, + "step": 3518 + }, + { + "epoch": 6.516666666666667, + "grad_norm": 1.5315816528514754, + "learning_rate": 1.9087636439087973e-07, + "loss": 0.6574, + "step": 3519 + }, + { + "epoch": 6.518518518518518, + "grad_norm": 1.552779362275153, + "learning_rate": 1.9087010610602666e-07, + "loss": 0.654, + "step": 3520 + }, + { + "epoch": 6.520370370370371, + "grad_norm": 1.537784393880499, + "learning_rate": 1.908638457781668e-07, + "loss": 0.6445, + "step": 3521 + }, + { + "epoch": 6.522222222222222, + "grad_norm": 1.5310748957344984, + "learning_rate": 1.9085758340744092e-07, + "loss": 0.6498, + "step": 3522 + }, + { + "epoch": 6.524074074074074, + "grad_norm": 1.4601432378948946, + "learning_rate": 1.9085131899398976e-07, + "loss": 0.6578, + "step": 3523 + }, + { + "epoch": 6.525925925925926, + "grad_norm": 1.4413110373869804, + "learning_rate": 1.9084505253795424e-07, + "loss": 0.6807, + "step": 3524 + }, + { + "epoch": 6.527777777777778, + "grad_norm": 1.6464147492343302, + "learning_rate": 1.908387840394752e-07, + "loss": 0.6839, + "step": 3525 + }, + { + "epoch": 6.52962962962963, + "grad_norm": 1.6304665083380474, + "learning_rate": 1.908325134986936e-07, + "loss": 0.6613, + "step": 3526 + }, + { + "epoch": 6.531481481481482, + "grad_norm": 1.3851430322764795, + "learning_rate": 1.9082624091575034e-07, + "loss": 0.6326, + "step": 3527 + }, + { + "epoch": 6.533333333333333, + "grad_norm": 1.404006762516901, + "learning_rate": 1.9081996629078654e-07, + "loss": 0.6728, + "step": 3528 + }, + { + "epoch": 6.535185185185185, + "grad_norm": 1.4498564116386556, + "learning_rate": 1.9081368962394322e-07, + "loss": 0.6824, + "step": 3529 + }, + { + "epoch": 6.537037037037037, + "grad_norm": 1.3628743587318175, + "learning_rate": 1.9080741091536153e-07, + "loss": 0.6304, + "step": 3530 + }, + { + "epoch": 6.538888888888889, + "grad_norm": 1.3221319601036166, + "learning_rate": 1.9080113016518263e-07, + "loss": 0.6433, + "step": 3531 + }, + { + "epoch": 6.540740740740741, + "grad_norm": 1.5930103250396523, + "learning_rate": 1.9079484737354768e-07, + "loss": 0.6647, + "step": 3532 + }, + { + "epoch": 6.542592592592593, + "grad_norm": 1.4803340619749408, + "learning_rate": 1.9078856254059797e-07, + "loss": 0.7207, + "step": 3533 + }, + { + "epoch": 6.544444444444444, + "grad_norm": 1.583338190977288, + "learning_rate": 1.907822756664748e-07, + "loss": 0.6541, + "step": 3534 + }, + { + "epoch": 6.546296296296296, + "grad_norm": 1.3597349589840217, + "learning_rate": 1.9077598675131948e-07, + "loss": 0.6915, + "step": 3535 + }, + { + "epoch": 6.548148148148148, + "grad_norm": 1.659837047491889, + "learning_rate": 1.9076969579527348e-07, + "loss": 0.6493, + "step": 3536 + }, + { + "epoch": 6.55, + "grad_norm": 2.021272565884714, + "learning_rate": 1.907634027984782e-07, + "loss": 0.6426, + "step": 3537 + }, + { + "epoch": 6.551851851851852, + "grad_norm": 1.7785422519535794, + "learning_rate": 1.907571077610751e-07, + "loss": 0.6715, + "step": 3538 + }, + { + "epoch": 6.553703703703704, + "grad_norm": 1.342462330809328, + "learning_rate": 1.9075081068320572e-07, + "loss": 0.6402, + "step": 3539 + }, + { + "epoch": 6.555555555555555, + "grad_norm": 1.4327846649043527, + "learning_rate": 1.907445115650116e-07, + "loss": 0.6529, + "step": 3540 + }, + { + "epoch": 6.557407407407408, + "grad_norm": 1.4304568368638708, + "learning_rate": 1.9073821040663446e-07, + "loss": 0.6411, + "step": 3541 + }, + { + "epoch": 6.559259259259259, + "grad_norm": 1.603628663431344, + "learning_rate": 1.9073190720821592e-07, + "loss": 0.6416, + "step": 3542 + }, + { + "epoch": 6.561111111111111, + "grad_norm": 1.7769364542069177, + "learning_rate": 1.9072560196989764e-07, + "loss": 0.6461, + "step": 3543 + }, + { + "epoch": 6.562962962962963, + "grad_norm": 1.9872393557421308, + "learning_rate": 1.9071929469182142e-07, + "loss": 0.7213, + "step": 3544 + }, + { + "epoch": 6.564814814814815, + "grad_norm": 1.640454704186781, + "learning_rate": 1.9071298537412907e-07, + "loss": 0.639, + "step": 3545 + }, + { + "epoch": 6.566666666666666, + "grad_norm": 1.4313170831766509, + "learning_rate": 1.9070667401696247e-07, + "loss": 0.6702, + "step": 3546 + }, + { + "epoch": 6.568518518518519, + "grad_norm": 1.4968096501907018, + "learning_rate": 1.9070036062046343e-07, + "loss": 0.6095, + "step": 3547 + }, + { + "epoch": 6.57037037037037, + "grad_norm": 1.5418600560473843, + "learning_rate": 1.9069404518477397e-07, + "loss": 0.6482, + "step": 3548 + }, + { + "epoch": 6.572222222222222, + "grad_norm": 1.557955934344611, + "learning_rate": 1.9068772771003604e-07, + "loss": 0.6443, + "step": 3549 + }, + { + "epoch": 6.574074074074074, + "grad_norm": 2.5980352575194696, + "learning_rate": 1.9068140819639172e-07, + "loss": 0.6083, + "step": 3550 + }, + { + "epoch": 6.575925925925926, + "grad_norm": 1.6872405238305705, + "learning_rate": 1.9067508664398303e-07, + "loss": 0.6354, + "step": 3551 + }, + { + "epoch": 6.5777777777777775, + "grad_norm": 1.5996374547292114, + "learning_rate": 1.9066876305295214e-07, + "loss": 0.6386, + "step": 3552 + }, + { + "epoch": 6.57962962962963, + "grad_norm": 1.9227447752325662, + "learning_rate": 1.906624374234412e-07, + "loss": 0.6105, + "step": 3553 + }, + { + "epoch": 6.5814814814814815, + "grad_norm": 1.381125193874707, + "learning_rate": 1.9065610975559241e-07, + "loss": 0.6662, + "step": 3554 + }, + { + "epoch": 6.583333333333333, + "grad_norm": 1.6763112740820068, + "learning_rate": 1.906497800495481e-07, + "loss": 0.6725, + "step": 3555 + }, + { + "epoch": 6.5851851851851855, + "grad_norm": 1.5304111243140848, + "learning_rate": 1.9064344830545046e-07, + "loss": 0.6392, + "step": 3556 + }, + { + "epoch": 6.587037037037037, + "grad_norm": 1.9180125056119757, + "learning_rate": 1.90637114523442e-07, + "loss": 0.6653, + "step": 3557 + }, + { + "epoch": 6.588888888888889, + "grad_norm": 3.1404883221884057, + "learning_rate": 1.90630778703665e-07, + "loss": 0.6594, + "step": 3558 + }, + { + "epoch": 6.590740740740741, + "grad_norm": 1.5320784304267354, + "learning_rate": 1.9062444084626192e-07, + "loss": 0.6627, + "step": 3559 + }, + { + "epoch": 6.592592592592593, + "grad_norm": 1.698347706238461, + "learning_rate": 1.9061810095137532e-07, + "loss": 0.6518, + "step": 3560 + }, + { + "epoch": 6.594444444444444, + "grad_norm": 1.8503443559882269, + "learning_rate": 1.9061175901914767e-07, + "loss": 0.6625, + "step": 3561 + }, + { + "epoch": 6.5962962962962965, + "grad_norm": 1.502712207905687, + "learning_rate": 1.906054150497216e-07, + "loss": 0.679, + "step": 3562 + }, + { + "epoch": 6.598148148148148, + "grad_norm": 2.2420178811252027, + "learning_rate": 1.905990690432397e-07, + "loss": 0.6714, + "step": 3563 + }, + { + "epoch": 6.6, + "grad_norm": 1.426132603147637, + "learning_rate": 1.9059272099984466e-07, + "loss": 0.6355, + "step": 3564 + }, + { + "epoch": 6.601851851851852, + "grad_norm": 3.2311387759751464, + "learning_rate": 1.9058637091967923e-07, + "loss": 0.6385, + "step": 3565 + }, + { + "epoch": 6.603703703703704, + "grad_norm": 1.464351256965262, + "learning_rate": 1.9058001880288614e-07, + "loss": 0.6815, + "step": 3566 + }, + { + "epoch": 6.605555555555555, + "grad_norm": 1.59727281260324, + "learning_rate": 1.9057366464960822e-07, + "loss": 0.6669, + "step": 3567 + }, + { + "epoch": 6.607407407407408, + "grad_norm": 1.555545966934019, + "learning_rate": 1.9056730845998833e-07, + "loss": 0.6608, + "step": 3568 + }, + { + "epoch": 6.609259259259259, + "grad_norm": 1.3354981240292743, + "learning_rate": 1.9056095023416936e-07, + "loss": 0.5987, + "step": 3569 + }, + { + "epoch": 6.611111111111111, + "grad_norm": 1.3640399136781258, + "learning_rate": 1.9055458997229426e-07, + "loss": 0.666, + "step": 3570 + }, + { + "epoch": 6.612962962962963, + "grad_norm": 1.304527838523876, + "learning_rate": 1.9054822767450602e-07, + "loss": 0.6297, + "step": 3571 + }, + { + "epoch": 6.614814814814815, + "grad_norm": 2.028089013989317, + "learning_rate": 1.9054186334094772e-07, + "loss": 0.6668, + "step": 3572 + }, + { + "epoch": 6.616666666666667, + "grad_norm": 1.412820961063143, + "learning_rate": 1.9053549697176244e-07, + "loss": 0.6453, + "step": 3573 + }, + { + "epoch": 6.618518518518519, + "grad_norm": 1.519384259862545, + "learning_rate": 1.9052912856709328e-07, + "loss": 0.6152, + "step": 3574 + }, + { + "epoch": 6.62037037037037, + "grad_norm": 1.5701999244434224, + "learning_rate": 1.9052275812708344e-07, + "loss": 0.6794, + "step": 3575 + }, + { + "epoch": 6.622222222222222, + "grad_norm": 1.5601762010325246, + "learning_rate": 1.9051638565187616e-07, + "loss": 0.6816, + "step": 3576 + }, + { + "epoch": 6.624074074074074, + "grad_norm": 1.549440481439686, + "learning_rate": 1.9051001114161463e-07, + "loss": 0.6436, + "step": 3577 + }, + { + "epoch": 6.625925925925926, + "grad_norm": 1.434176175710681, + "learning_rate": 1.905036345964423e-07, + "loss": 0.6433, + "step": 3578 + }, + { + "epoch": 6.627777777777778, + "grad_norm": 1.6055836311587846, + "learning_rate": 1.9049725601650244e-07, + "loss": 0.645, + "step": 3579 + }, + { + "epoch": 6.62962962962963, + "grad_norm": 1.5525979025070047, + "learning_rate": 1.9049087540193845e-07, + "loss": 0.6428, + "step": 3580 + }, + { + "epoch": 6.631481481481481, + "grad_norm": 3.4383964362447075, + "learning_rate": 1.9048449275289382e-07, + "loss": 0.6159, + "step": 3581 + }, + { + "epoch": 6.633333333333333, + "grad_norm": 1.593866565104126, + "learning_rate": 1.9047810806951206e-07, + "loss": 0.687, + "step": 3582 + }, + { + "epoch": 6.635185185185185, + "grad_norm": 1.490724425194913, + "learning_rate": 1.9047172135193666e-07, + "loss": 0.6403, + "step": 3583 + }, + { + "epoch": 6.637037037037037, + "grad_norm": 1.5031928217828185, + "learning_rate": 1.904653326003113e-07, + "loss": 0.6476, + "step": 3584 + }, + { + "epoch": 6.638888888888889, + "grad_norm": 1.5238868620861905, + "learning_rate": 1.9045894181477952e-07, + "loss": 0.6851, + "step": 3585 + }, + { + "epoch": 6.640740740740741, + "grad_norm": 1.4788607025563554, + "learning_rate": 1.9045254899548507e-07, + "loss": 0.6539, + "step": 3586 + }, + { + "epoch": 6.642592592592592, + "grad_norm": 1.3643331152600338, + "learning_rate": 1.9044615414257165e-07, + "loss": 0.6786, + "step": 3587 + }, + { + "epoch": 6.644444444444445, + "grad_norm": 1.5213643119843239, + "learning_rate": 1.9043975725618302e-07, + "loss": 0.6691, + "step": 3588 + }, + { + "epoch": 6.646296296296296, + "grad_norm": 1.8462854866239424, + "learning_rate": 1.90433358336463e-07, + "loss": 0.6895, + "step": 3589 + }, + { + "epoch": 6.648148148148148, + "grad_norm": 1.9198249255725623, + "learning_rate": 1.9042695738355552e-07, + "loss": 0.7524, + "step": 3590 + }, + { + "epoch": 6.65, + "grad_norm": 1.6441331792564244, + "learning_rate": 1.9042055439760442e-07, + "loss": 0.687, + "step": 3591 + }, + { + "epoch": 6.651851851851852, + "grad_norm": 2.226794634160272, + "learning_rate": 1.9041414937875368e-07, + "loss": 0.6534, + "step": 3592 + }, + { + "epoch": 6.6537037037037035, + "grad_norm": 1.5938911055605058, + "learning_rate": 1.904077423271473e-07, + "loss": 0.6493, + "step": 3593 + }, + { + "epoch": 6.655555555555556, + "grad_norm": 1.5257997175416245, + "learning_rate": 1.9040133324292932e-07, + "loss": 0.6844, + "step": 3594 + }, + { + "epoch": 6.657407407407407, + "grad_norm": 1.4899952780270258, + "learning_rate": 1.9039492212624387e-07, + "loss": 0.6552, + "step": 3595 + }, + { + "epoch": 6.659259259259259, + "grad_norm": 1.65106697519256, + "learning_rate": 1.9038850897723503e-07, + "loss": 0.6443, + "step": 3596 + }, + { + "epoch": 6.661111111111111, + "grad_norm": 1.460551136135121, + "learning_rate": 1.9038209379604703e-07, + "loss": 0.6533, + "step": 3597 + }, + { + "epoch": 6.662962962962963, + "grad_norm": 1.404160694039314, + "learning_rate": 1.9037567658282407e-07, + "loss": 0.682, + "step": 3598 + }, + { + "epoch": 6.6648148148148145, + "grad_norm": 1.5239984959272146, + "learning_rate": 1.9036925733771048e-07, + "loss": 0.6504, + "step": 3599 + }, + { + "epoch": 6.666666666666667, + "grad_norm": 1.5171195025470456, + "learning_rate": 1.9036283606085054e-07, + "loss": 0.6393, + "step": 3600 + }, + { + "epoch": 6.6685185185185185, + "grad_norm": 1.590037615031641, + "learning_rate": 1.9035641275238858e-07, + "loss": 0.6761, + "step": 3601 + }, + { + "epoch": 6.67037037037037, + "grad_norm": 1.4477573761909848, + "learning_rate": 1.903499874124691e-07, + "loss": 0.6504, + "step": 3602 + }, + { + "epoch": 6.6722222222222225, + "grad_norm": 1.3808780730625259, + "learning_rate": 1.903435600412365e-07, + "loss": 0.6613, + "step": 3603 + }, + { + "epoch": 6.674074074074074, + "grad_norm": 1.7024238755083634, + "learning_rate": 1.903371306388353e-07, + "loss": 0.6665, + "step": 3604 + }, + { + "epoch": 6.675925925925926, + "grad_norm": 1.6534460938459639, + "learning_rate": 1.9033069920541007e-07, + "loss": 0.6325, + "step": 3605 + }, + { + "epoch": 6.677777777777778, + "grad_norm": 1.7680368366360395, + "learning_rate": 1.9032426574110535e-07, + "loss": 0.6625, + "step": 3606 + }, + { + "epoch": 6.67962962962963, + "grad_norm": 1.6637677084100126, + "learning_rate": 1.9031783024606587e-07, + "loss": 0.6501, + "step": 3607 + }, + { + "epoch": 6.681481481481481, + "grad_norm": 1.6660782812856678, + "learning_rate": 1.9031139272043623e-07, + "loss": 0.68, + "step": 3608 + }, + { + "epoch": 6.683333333333334, + "grad_norm": 1.67232389224004, + "learning_rate": 1.9030495316436118e-07, + "loss": 0.6674, + "step": 3609 + }, + { + "epoch": 6.685185185185185, + "grad_norm": 1.708669316071796, + "learning_rate": 1.9029851157798556e-07, + "loss": 0.64, + "step": 3610 + }, + { + "epoch": 6.687037037037037, + "grad_norm": 1.5381193120897307, + "learning_rate": 1.9029206796145413e-07, + "loss": 0.6742, + "step": 3611 + }, + { + "epoch": 6.688888888888889, + "grad_norm": 1.557536196178727, + "learning_rate": 1.902856223149118e-07, + "loss": 0.6127, + "step": 3612 + }, + { + "epoch": 6.690740740740741, + "grad_norm": 1.5339706282702676, + "learning_rate": 1.9027917463850347e-07, + "loss": 0.6371, + "step": 3613 + }, + { + "epoch": 6.692592592592592, + "grad_norm": 1.5591672007068569, + "learning_rate": 1.902727249323741e-07, + "loss": 0.63, + "step": 3614 + }, + { + "epoch": 6.694444444444445, + "grad_norm": 1.451981717697895, + "learning_rate": 1.902662731966687e-07, + "loss": 0.6193, + "step": 3615 + }, + { + "epoch": 6.696296296296296, + "grad_norm": 1.5182456827043982, + "learning_rate": 1.9025981943153228e-07, + "loss": 0.6551, + "step": 3616 + }, + { + "epoch": 6.698148148148148, + "grad_norm": 1.8914097147039861, + "learning_rate": 1.9025336363711002e-07, + "loss": 0.6821, + "step": 3617 + }, + { + "epoch": 6.7, + "grad_norm": 1.8119385081592834, + "learning_rate": 1.9024690581354698e-07, + "loss": 0.642, + "step": 3618 + }, + { + "epoch": 6.701851851851852, + "grad_norm": 1.564284701728419, + "learning_rate": 1.902404459609884e-07, + "loss": 0.6594, + "step": 3619 + }, + { + "epoch": 6.703703703703704, + "grad_norm": 1.6367265966760483, + "learning_rate": 1.9023398407957957e-07, + "loss": 0.6896, + "step": 3620 + }, + { + "epoch": 6.705555555555556, + "grad_norm": 1.9223921672916777, + "learning_rate": 1.9022752016946565e-07, + "loss": 0.6774, + "step": 3621 + }, + { + "epoch": 6.707407407407407, + "grad_norm": 1.6486392685417746, + "learning_rate": 1.90221054230792e-07, + "loss": 0.6648, + "step": 3622 + }, + { + "epoch": 6.709259259259259, + "grad_norm": 1.659731780565725, + "learning_rate": 1.9021458626370403e-07, + "loss": 0.6911, + "step": 3623 + }, + { + "epoch": 6.711111111111111, + "grad_norm": 1.4729867084364658, + "learning_rate": 1.9020811626834719e-07, + "loss": 0.7183, + "step": 3624 + }, + { + "epoch": 6.712962962962963, + "grad_norm": 1.6601496240955846, + "learning_rate": 1.9020164424486685e-07, + "loss": 0.6424, + "step": 3625 + }, + { + "epoch": 6.714814814814815, + "grad_norm": 1.565639553086471, + "learning_rate": 1.9019517019340856e-07, + "loss": 0.6663, + "step": 3626 + }, + { + "epoch": 6.716666666666667, + "grad_norm": 1.620155777521443, + "learning_rate": 1.9018869411411788e-07, + "loss": 0.6577, + "step": 3627 + }, + { + "epoch": 6.718518518518518, + "grad_norm": 1.5013395508326475, + "learning_rate": 1.9018221600714043e-07, + "loss": 0.6677, + "step": 3628 + }, + { + "epoch": 6.72037037037037, + "grad_norm": 1.4475065971540078, + "learning_rate": 1.901757358726218e-07, + "loss": 0.6266, + "step": 3629 + }, + { + "epoch": 6.722222222222222, + "grad_norm": 1.5601766751913895, + "learning_rate": 1.9016925371070776e-07, + "loss": 0.6363, + "step": 3630 + }, + { + "epoch": 6.724074074074074, + "grad_norm": 1.5927068616657944, + "learning_rate": 1.9016276952154398e-07, + "loss": 0.6368, + "step": 3631 + }, + { + "epoch": 6.725925925925926, + "grad_norm": 1.5776497160193574, + "learning_rate": 1.9015628330527625e-07, + "loss": 0.6527, + "step": 3632 + }, + { + "epoch": 6.727777777777778, + "grad_norm": 1.6473345836128068, + "learning_rate": 1.901497950620504e-07, + "loss": 0.6347, + "step": 3633 + }, + { + "epoch": 6.729629629629629, + "grad_norm": 1.570245606739739, + "learning_rate": 1.9014330479201234e-07, + "loss": 0.6591, + "step": 3634 + }, + { + "epoch": 6.731481481481482, + "grad_norm": 1.45709132223943, + "learning_rate": 1.9013681249530795e-07, + "loss": 0.6541, + "step": 3635 + }, + { + "epoch": 6.733333333333333, + "grad_norm": 1.4581327408399674, + "learning_rate": 1.9013031817208322e-07, + "loss": 0.6507, + "step": 3636 + }, + { + "epoch": 6.735185185185185, + "grad_norm": 1.8865447386482699, + "learning_rate": 1.9012382182248413e-07, + "loss": 0.61, + "step": 3637 + }, + { + "epoch": 6.737037037037037, + "grad_norm": 1.5314080752434895, + "learning_rate": 1.901173234466568e-07, + "loss": 0.6661, + "step": 3638 + }, + { + "epoch": 6.738888888888889, + "grad_norm": 1.696917313911211, + "learning_rate": 1.901108230447472e-07, + "loss": 0.6447, + "step": 3639 + }, + { + "epoch": 6.7407407407407405, + "grad_norm": 1.559217656282439, + "learning_rate": 1.9010432061690163e-07, + "loss": 0.6322, + "step": 3640 + }, + { + "epoch": 6.742592592592593, + "grad_norm": 1.631279597257817, + "learning_rate": 1.9009781616326616e-07, + "loss": 0.6121, + "step": 3641 + }, + { + "epoch": 6.7444444444444445, + "grad_norm": 1.6531603742333165, + "learning_rate": 1.900913096839871e-07, + "loss": 0.6272, + "step": 3642 + }, + { + "epoch": 6.746296296296296, + "grad_norm": 1.5893929247462644, + "learning_rate": 1.9008480117921075e-07, + "loss": 0.6904, + "step": 3643 + }, + { + "epoch": 6.7481481481481485, + "grad_norm": 1.6028986652736323, + "learning_rate": 1.900782906490834e-07, + "loss": 0.6486, + "step": 3644 + }, + { + "epoch": 6.75, + "grad_norm": 1.5784819772103595, + "learning_rate": 1.9007177809375138e-07, + "loss": 0.6534, + "step": 3645 + }, + { + "epoch": 6.7518518518518515, + "grad_norm": 1.4843783997718978, + "learning_rate": 1.9006526351336118e-07, + "loss": 0.6357, + "step": 3646 + }, + { + "epoch": 6.753703703703704, + "grad_norm": 1.4347707211685412, + "learning_rate": 1.9005874690805926e-07, + "loss": 0.6379, + "step": 3647 + }, + { + "epoch": 6.7555555555555555, + "grad_norm": 1.8082011737370267, + "learning_rate": 1.900522282779921e-07, + "loss": 0.6775, + "step": 3648 + }, + { + "epoch": 6.757407407407407, + "grad_norm": 1.4323001375174702, + "learning_rate": 1.9004570762330627e-07, + "loss": 0.6421, + "step": 3649 + }, + { + "epoch": 6.7592592592592595, + "grad_norm": 1.4598483569102236, + "learning_rate": 1.9003918494414838e-07, + "loss": 0.6975, + "step": 3650 + }, + { + "epoch": 6.761111111111111, + "grad_norm": 1.5324125968973363, + "learning_rate": 1.9003266024066504e-07, + "loss": 0.7096, + "step": 3651 + }, + { + "epoch": 6.762962962962963, + "grad_norm": 1.4122644275958263, + "learning_rate": 1.9002613351300298e-07, + "loss": 0.7087, + "step": 3652 + }, + { + "epoch": 6.764814814814815, + "grad_norm": 1.6553660384444624, + "learning_rate": 1.9001960476130895e-07, + "loss": 0.6454, + "step": 3653 + }, + { + "epoch": 6.766666666666667, + "grad_norm": 1.6761040730676293, + "learning_rate": 1.9001307398572972e-07, + "loss": 0.6815, + "step": 3654 + }, + { + "epoch": 6.768518518518518, + "grad_norm": 1.5579402446549757, + "learning_rate": 1.9000654118641212e-07, + "loss": 0.6891, + "step": 3655 + }, + { + "epoch": 6.770370370370371, + "grad_norm": 1.5285465413775523, + "learning_rate": 1.90000006363503e-07, + "loss": 0.6922, + "step": 3656 + }, + { + "epoch": 6.772222222222222, + "grad_norm": 1.558157121276329, + "learning_rate": 1.899934695171493e-07, + "loss": 0.6595, + "step": 3657 + }, + { + "epoch": 6.774074074074074, + "grad_norm": 1.468871471995222, + "learning_rate": 1.8998693064749798e-07, + "loss": 0.6886, + "step": 3658 + }, + { + "epoch": 6.775925925925926, + "grad_norm": 1.5910103800078768, + "learning_rate": 1.8998038975469608e-07, + "loss": 0.7032, + "step": 3659 + }, + { + "epoch": 6.777777777777778, + "grad_norm": 1.7030079963166107, + "learning_rate": 1.8997384683889063e-07, + "loss": 0.6726, + "step": 3660 + }, + { + "epoch": 6.77962962962963, + "grad_norm": 1.7253881426431534, + "learning_rate": 1.899673019002287e-07, + "loss": 0.6056, + "step": 3661 + }, + { + "epoch": 6.781481481481482, + "grad_norm": 1.5759875449971485, + "learning_rate": 1.8996075493885754e-07, + "loss": 0.663, + "step": 3662 + }, + { + "epoch": 6.783333333333333, + "grad_norm": 1.4675711051542797, + "learning_rate": 1.8995420595492426e-07, + "loss": 0.6151, + "step": 3663 + }, + { + "epoch": 6.785185185185185, + "grad_norm": 1.7946600064410578, + "learning_rate": 1.899476549485761e-07, + "loss": 0.6387, + "step": 3664 + }, + { + "epoch": 6.787037037037037, + "grad_norm": 1.5098823524452794, + "learning_rate": 1.8994110191996038e-07, + "loss": 0.6712, + "step": 3665 + }, + { + "epoch": 6.788888888888889, + "grad_norm": 1.7301205099205819, + "learning_rate": 1.899345468692244e-07, + "loss": 0.6464, + "step": 3666 + }, + { + "epoch": 6.790740740740741, + "grad_norm": 1.8317356681476744, + "learning_rate": 1.8992798979651555e-07, + "loss": 0.6296, + "step": 3667 + }, + { + "epoch": 6.792592592592593, + "grad_norm": 1.6819194696226583, + "learning_rate": 1.8992143070198122e-07, + "loss": 0.6691, + "step": 3668 + }, + { + "epoch": 6.794444444444444, + "grad_norm": 2.4692521962781586, + "learning_rate": 1.8991486958576893e-07, + "loss": 0.6427, + "step": 3669 + }, + { + "epoch": 6.796296296296296, + "grad_norm": 1.576866072527621, + "learning_rate": 1.899083064480262e-07, + "loss": 0.6642, + "step": 3670 + }, + { + "epoch": 6.798148148148148, + "grad_norm": 1.3835151645033728, + "learning_rate": 1.8990174128890053e-07, + "loss": 0.6744, + "step": 3671 + }, + { + "epoch": 6.8, + "grad_norm": 1.7439390373733747, + "learning_rate": 1.8989517410853953e-07, + "loss": 0.6354, + "step": 3672 + }, + { + "epoch": 6.801851851851852, + "grad_norm": 1.6657889949287623, + "learning_rate": 1.8988860490709087e-07, + "loss": 0.678, + "step": 3673 + }, + { + "epoch": 6.803703703703704, + "grad_norm": 1.689354563387892, + "learning_rate": 1.8988203368470223e-07, + "loss": 0.6813, + "step": 3674 + }, + { + "epoch": 6.805555555555555, + "grad_norm": 1.4532169248299824, + "learning_rate": 1.8987546044152138e-07, + "loss": 0.6091, + "step": 3675 + }, + { + "epoch": 6.807407407407408, + "grad_norm": 1.4390991828113617, + "learning_rate": 1.8986888517769607e-07, + "loss": 0.6739, + "step": 3676 + }, + { + "epoch": 6.809259259259259, + "grad_norm": 3.563430021277078, + "learning_rate": 1.8986230789337413e-07, + "loss": 0.6605, + "step": 3677 + }, + { + "epoch": 6.811111111111111, + "grad_norm": 1.877626494712757, + "learning_rate": 1.8985572858870346e-07, + "loss": 0.6101, + "step": 3678 + }, + { + "epoch": 6.812962962962963, + "grad_norm": 2.4495109425371644, + "learning_rate": 1.89849147263832e-07, + "loss": 0.6777, + "step": 3679 + }, + { + "epoch": 6.814814814814815, + "grad_norm": 1.6373458280822453, + "learning_rate": 1.8984256391890763e-07, + "loss": 0.6637, + "step": 3680 + }, + { + "epoch": 6.816666666666666, + "grad_norm": 2.0339961095190215, + "learning_rate": 1.8983597855407843e-07, + "loss": 0.693, + "step": 3681 + }, + { + "epoch": 6.818518518518519, + "grad_norm": 1.3967124979792362, + "learning_rate": 1.8982939116949247e-07, + "loss": 0.6882, + "step": 3682 + }, + { + "epoch": 6.82037037037037, + "grad_norm": 1.4535407344970497, + "learning_rate": 1.8982280176529777e-07, + "loss": 0.7006, + "step": 3683 + }, + { + "epoch": 6.822222222222222, + "grad_norm": 1.7322387700173034, + "learning_rate": 1.8981621034164256e-07, + "loss": 0.6533, + "step": 3684 + }, + { + "epoch": 6.824074074074074, + "grad_norm": 1.8912976923510763, + "learning_rate": 1.89809616898675e-07, + "loss": 0.6459, + "step": 3685 + }, + { + "epoch": 6.825925925925926, + "grad_norm": 1.5542267896306647, + "learning_rate": 1.8980302143654333e-07, + "loss": 0.6561, + "step": 3686 + }, + { + "epoch": 6.8277777777777775, + "grad_norm": 1.6631186467969499, + "learning_rate": 1.8979642395539584e-07, + "loss": 0.6461, + "step": 3687 + }, + { + "epoch": 6.82962962962963, + "grad_norm": 1.4249597361328639, + "learning_rate": 1.8978982445538088e-07, + "loss": 0.671, + "step": 3688 + }, + { + "epoch": 6.8314814814814815, + "grad_norm": 1.4921100682037167, + "learning_rate": 1.8978322293664677e-07, + "loss": 0.674, + "step": 3689 + }, + { + "epoch": 6.833333333333333, + "grad_norm": 1.5023788231474966, + "learning_rate": 1.8977661939934198e-07, + "loss": 0.6674, + "step": 3690 + }, + { + "epoch": 6.8351851851851855, + "grad_norm": 1.9942026792163534, + "learning_rate": 1.8977001384361496e-07, + "loss": 0.6784, + "step": 3691 + }, + { + "epoch": 6.837037037037037, + "grad_norm": 1.497253795214737, + "learning_rate": 1.8976340626961418e-07, + "loss": 0.6552, + "step": 3692 + }, + { + "epoch": 6.838888888888889, + "grad_norm": 1.6422850609691322, + "learning_rate": 1.8975679667748826e-07, + "loss": 0.6449, + "step": 3693 + }, + { + "epoch": 6.840740740740741, + "grad_norm": 1.3947561721057105, + "learning_rate": 1.8975018506738576e-07, + "loss": 0.6595, + "step": 3694 + }, + { + "epoch": 6.842592592592593, + "grad_norm": 2.75657147715647, + "learning_rate": 1.8974357143945537e-07, + "loss": 0.696, + "step": 3695 + }, + { + "epoch": 6.844444444444444, + "grad_norm": 1.857394533609168, + "learning_rate": 1.8973695579384573e-07, + "loss": 0.6461, + "step": 3696 + }, + { + "epoch": 6.8462962962962965, + "grad_norm": 1.997427286217202, + "learning_rate": 1.897303381307056e-07, + "loss": 0.6719, + "step": 3697 + }, + { + "epoch": 6.848148148148148, + "grad_norm": 1.416733749823453, + "learning_rate": 1.8972371845018377e-07, + "loss": 0.6154, + "step": 3698 + }, + { + "epoch": 6.85, + "grad_norm": 1.5585819483769643, + "learning_rate": 1.897170967524291e-07, + "loss": 0.679, + "step": 3699 + }, + { + "epoch": 6.851851851851852, + "grad_norm": 1.5393001317437378, + "learning_rate": 1.897104730375904e-07, + "loss": 0.6614, + "step": 3700 + }, + { + "epoch": 6.853703703703704, + "grad_norm": 1.661770609922812, + "learning_rate": 1.897038473058166e-07, + "loss": 0.6709, + "step": 3701 + }, + { + "epoch": 6.855555555555555, + "grad_norm": 1.5975764750358483, + "learning_rate": 1.8969721955725668e-07, + "loss": 0.6373, + "step": 3702 + }, + { + "epoch": 6.857407407407408, + "grad_norm": 1.3259836817862736, + "learning_rate": 1.8969058979205968e-07, + "loss": 0.6434, + "step": 3703 + }, + { + "epoch": 6.859259259259259, + "grad_norm": 1.5721941674254143, + "learning_rate": 1.896839580103746e-07, + "loss": 0.7058, + "step": 3704 + }, + { + "epoch": 6.861111111111111, + "grad_norm": 1.6728291920538325, + "learning_rate": 1.8967732421235057e-07, + "loss": 0.6341, + "step": 3705 + }, + { + "epoch": 6.862962962962963, + "grad_norm": 1.5774156773087555, + "learning_rate": 1.8967068839813673e-07, + "loss": 0.6804, + "step": 3706 + }, + { + "epoch": 6.864814814814815, + "grad_norm": 1.965298544747606, + "learning_rate": 1.8966405056788227e-07, + "loss": 0.6441, + "step": 3707 + }, + { + "epoch": 6.866666666666667, + "grad_norm": 1.7513797049797886, + "learning_rate": 1.8965741072173644e-07, + "loss": 0.6133, + "step": 3708 + }, + { + "epoch": 6.868518518518519, + "grad_norm": 1.5579099413716568, + "learning_rate": 1.8965076885984854e-07, + "loss": 0.6609, + "step": 3709 + }, + { + "epoch": 6.87037037037037, + "grad_norm": 1.6450827434769868, + "learning_rate": 1.8964412498236785e-07, + "loss": 0.6378, + "step": 3710 + }, + { + "epoch": 6.872222222222222, + "grad_norm": 1.3831758194677863, + "learning_rate": 1.8963747908944372e-07, + "loss": 0.6889, + "step": 3711 + }, + { + "epoch": 6.874074074074074, + "grad_norm": 1.5677892936721385, + "learning_rate": 1.8963083118122562e-07, + "loss": 0.6804, + "step": 3712 + }, + { + "epoch": 6.875925925925926, + "grad_norm": 1.6451382518344124, + "learning_rate": 1.8962418125786303e-07, + "loss": 0.6897, + "step": 3713 + }, + { + "epoch": 6.877777777777778, + "grad_norm": 1.465058575361201, + "learning_rate": 1.8961752931950544e-07, + "loss": 0.6288, + "step": 3714 + }, + { + "epoch": 6.87962962962963, + "grad_norm": 1.8808036509802426, + "learning_rate": 1.8961087536630235e-07, + "loss": 0.6385, + "step": 3715 + }, + { + "epoch": 6.881481481481481, + "grad_norm": 1.6016158267785385, + "learning_rate": 1.8960421939840344e-07, + "loss": 0.6166, + "step": 3716 + }, + { + "epoch": 6.883333333333333, + "grad_norm": 1.6260360623218741, + "learning_rate": 1.895975614159583e-07, + "loss": 0.6185, + "step": 3717 + }, + { + "epoch": 6.885185185185185, + "grad_norm": 1.5297775695269318, + "learning_rate": 1.8959090141911665e-07, + "loss": 0.6401, + "step": 3718 + }, + { + "epoch": 6.887037037037037, + "grad_norm": 1.7343884149624085, + "learning_rate": 1.8958423940802816e-07, + "loss": 0.6745, + "step": 3719 + }, + { + "epoch": 6.888888888888889, + "grad_norm": 1.3219886578391538, + "learning_rate": 1.8957757538284272e-07, + "loss": 0.6355, + "step": 3720 + }, + { + "epoch": 6.890740740740741, + "grad_norm": 1.4618830623508936, + "learning_rate": 1.8957090934371007e-07, + "loss": 0.6661, + "step": 3721 + }, + { + "epoch": 6.892592592592592, + "grad_norm": 1.8065799642056688, + "learning_rate": 1.8956424129078013e-07, + "loss": 0.6567, + "step": 3722 + }, + { + "epoch": 6.894444444444445, + "grad_norm": 1.4928704055693034, + "learning_rate": 1.8955757122420275e-07, + "loss": 0.6622, + "step": 3723 + }, + { + "epoch": 6.896296296296296, + "grad_norm": 1.8252254396519585, + "learning_rate": 1.8955089914412796e-07, + "loss": 0.6348, + "step": 3724 + }, + { + "epoch": 6.898148148148148, + "grad_norm": 1.7925279281791, + "learning_rate": 1.8954422505070574e-07, + "loss": 0.6472, + "step": 3725 + }, + { + "epoch": 6.9, + "grad_norm": 1.6770141151811553, + "learning_rate": 1.8953754894408616e-07, + "loss": 0.6658, + "step": 3726 + }, + { + "epoch": 6.901851851851852, + "grad_norm": 1.5584695713974657, + "learning_rate": 1.8953087082441924e-07, + "loss": 0.673, + "step": 3727 + }, + { + "epoch": 6.9037037037037035, + "grad_norm": 1.3328363643814065, + "learning_rate": 1.8952419069185527e-07, + "loss": 0.6337, + "step": 3728 + }, + { + "epoch": 6.905555555555556, + "grad_norm": 1.7023790785545256, + "learning_rate": 1.8951750854654428e-07, + "loss": 0.657, + "step": 3729 + }, + { + "epoch": 6.907407407407407, + "grad_norm": 1.57593606065029, + "learning_rate": 1.895108243886366e-07, + "loss": 0.6454, + "step": 3730 + }, + { + "epoch": 6.909259259259259, + "grad_norm": 1.5176957610124138, + "learning_rate": 1.8950413821828248e-07, + "loss": 0.661, + "step": 3731 + }, + { + "epoch": 6.911111111111111, + "grad_norm": 1.6553322127203016, + "learning_rate": 1.8949745003563222e-07, + "loss": 0.623, + "step": 3732 + }, + { + "epoch": 6.912962962962963, + "grad_norm": 1.448231726102951, + "learning_rate": 1.8949075984083622e-07, + "loss": 0.6115, + "step": 3733 + }, + { + "epoch": 6.9148148148148145, + "grad_norm": 1.649715706166609, + "learning_rate": 1.894840676340449e-07, + "loss": 0.6406, + "step": 3734 + }, + { + "epoch": 6.916666666666667, + "grad_norm": 1.3708481330181648, + "learning_rate": 1.894773734154087e-07, + "loss": 0.6363, + "step": 3735 + }, + { + "epoch": 6.9185185185185185, + "grad_norm": 1.8490920386925511, + "learning_rate": 1.8947067718507814e-07, + "loss": 0.6662, + "step": 3736 + }, + { + "epoch": 6.92037037037037, + "grad_norm": 1.4857734145358497, + "learning_rate": 1.8946397894320375e-07, + "loss": 0.7231, + "step": 3737 + }, + { + "epoch": 6.9222222222222225, + "grad_norm": 1.7840043358150104, + "learning_rate": 1.8945727868993613e-07, + "loss": 0.6666, + "step": 3738 + }, + { + "epoch": 6.924074074074074, + "grad_norm": 1.867143935152978, + "learning_rate": 1.8945057642542591e-07, + "loss": 0.6939, + "step": 3739 + }, + { + "epoch": 6.925925925925926, + "grad_norm": 2.375579610071121, + "learning_rate": 1.894438721498238e-07, + "loss": 0.6594, + "step": 3740 + }, + { + "epoch": 6.927777777777778, + "grad_norm": 1.6727899561474089, + "learning_rate": 1.8943716586328052e-07, + "loss": 0.6364, + "step": 3741 + }, + { + "epoch": 6.92962962962963, + "grad_norm": 2.1328839934262414, + "learning_rate": 1.8943045756594685e-07, + "loss": 0.6822, + "step": 3742 + }, + { + "epoch": 6.931481481481481, + "grad_norm": 1.9963577592510184, + "learning_rate": 1.894237472579736e-07, + "loss": 0.6465, + "step": 3743 + }, + { + "epoch": 6.933333333333334, + "grad_norm": 1.7912782459950478, + "learning_rate": 1.8941703493951163e-07, + "loss": 0.6431, + "step": 3744 + }, + { + "epoch": 6.935185185185185, + "grad_norm": 1.633204654362644, + "learning_rate": 1.8941032061071186e-07, + "loss": 0.6788, + "step": 3745 + }, + { + "epoch": 6.937037037037037, + "grad_norm": 1.5130681240939485, + "learning_rate": 1.8940360427172527e-07, + "loss": 0.67, + "step": 3746 + }, + { + "epoch": 6.938888888888889, + "grad_norm": 1.4701924131258375, + "learning_rate": 1.893968859227028e-07, + "loss": 0.6994, + "step": 3747 + }, + { + "epoch": 6.940740740740741, + "grad_norm": 1.5780033257574644, + "learning_rate": 1.8939016556379558e-07, + "loss": 0.633, + "step": 3748 + }, + { + "epoch": 6.942592592592592, + "grad_norm": 1.513845315573104, + "learning_rate": 1.8938344319515466e-07, + "loss": 0.6485, + "step": 3749 + }, + { + "epoch": 6.944444444444445, + "grad_norm": 1.5162303552476346, + "learning_rate": 1.8937671881693117e-07, + "loss": 0.663, + "step": 3750 + }, + { + "epoch": 6.946296296296296, + "grad_norm": 1.4872176066181257, + "learning_rate": 1.8936999242927625e-07, + "loss": 0.6435, + "step": 3751 + }, + { + "epoch": 6.948148148148148, + "grad_norm": 1.4513972699548876, + "learning_rate": 1.8936326403234123e-07, + "loss": 0.6382, + "step": 3752 + }, + { + "epoch": 6.95, + "grad_norm": 1.371152779136157, + "learning_rate": 1.8935653362627728e-07, + "loss": 0.6425, + "step": 3753 + }, + { + "epoch": 6.951851851851852, + "grad_norm": 1.5893033714815785, + "learning_rate": 1.893498012112358e-07, + "loss": 0.6855, + "step": 3754 + }, + { + "epoch": 6.953703703703704, + "grad_norm": 1.6018091762083748, + "learning_rate": 1.893430667873681e-07, + "loss": 0.6902, + "step": 3755 + }, + { + "epoch": 6.955555555555556, + "grad_norm": 1.5234220101690332, + "learning_rate": 1.8933633035482562e-07, + "loss": 0.6753, + "step": 3756 + }, + { + "epoch": 6.957407407407407, + "grad_norm": 1.5662858067571013, + "learning_rate": 1.893295919137598e-07, + "loss": 0.6766, + "step": 3757 + }, + { + "epoch": 6.959259259259259, + "grad_norm": 1.635046606984755, + "learning_rate": 1.8932285146432215e-07, + "loss": 0.6125, + "step": 3758 + }, + { + "epoch": 6.961111111111111, + "grad_norm": 1.526099270404238, + "learning_rate": 1.8931610900666416e-07, + "loss": 0.6756, + "step": 3759 + }, + { + "epoch": 6.962962962962963, + "grad_norm": 1.7483574053613806, + "learning_rate": 1.8930936454093752e-07, + "loss": 0.6631, + "step": 3760 + }, + { + "epoch": 6.964814814814815, + "grad_norm": 1.821187802000183, + "learning_rate": 1.8930261806729375e-07, + "loss": 0.6678, + "step": 3761 + }, + { + "epoch": 6.966666666666667, + "grad_norm": 1.8067898156884483, + "learning_rate": 1.8929586958588462e-07, + "loss": 0.6698, + "step": 3762 + }, + { + "epoch": 6.968518518518518, + "grad_norm": 1.4730568325665743, + "learning_rate": 1.892891190968618e-07, + "loss": 0.6941, + "step": 3763 + }, + { + "epoch": 6.97037037037037, + "grad_norm": 1.4322413224394548, + "learning_rate": 1.8928236660037708e-07, + "loss": 0.6619, + "step": 3764 + }, + { + "epoch": 6.972222222222222, + "grad_norm": 1.4437285302509855, + "learning_rate": 1.8927561209658227e-07, + "loss": 0.6906, + "step": 3765 + }, + { + "epoch": 6.974074074074074, + "grad_norm": 2.0368963930578716, + "learning_rate": 1.892688555856292e-07, + "loss": 0.6649, + "step": 3766 + }, + { + "epoch": 6.975925925925926, + "grad_norm": 1.913007026516596, + "learning_rate": 1.8926209706766986e-07, + "loss": 0.6778, + "step": 3767 + }, + { + "epoch": 6.977777777777778, + "grad_norm": 2.000599159827014, + "learning_rate": 1.8925533654285615e-07, + "loss": 0.6854, + "step": 3768 + }, + { + "epoch": 6.979629629629629, + "grad_norm": 1.5554951368903644, + "learning_rate": 1.8924857401134e-07, + "loss": 0.6495, + "step": 3769 + }, + { + "epoch": 6.981481481481482, + "grad_norm": 1.5708077309923056, + "learning_rate": 1.8924180947327356e-07, + "loss": 0.6618, + "step": 3770 + }, + { + "epoch": 6.983333333333333, + "grad_norm": 1.6402954939170293, + "learning_rate": 1.8923504292880885e-07, + "loss": 0.6049, + "step": 3771 + }, + { + "epoch": 6.985185185185185, + "grad_norm": 1.4131337947493847, + "learning_rate": 1.89228274378098e-07, + "loss": 0.6744, + "step": 3772 + }, + { + "epoch": 6.987037037037037, + "grad_norm": 1.5548263555384274, + "learning_rate": 1.8922150382129325e-07, + "loss": 0.6536, + "step": 3773 + }, + { + "epoch": 6.988888888888889, + "grad_norm": 1.4625756497529754, + "learning_rate": 1.8921473125854672e-07, + "loss": 0.6455, + "step": 3774 + }, + { + "epoch": 6.9907407407407405, + "grad_norm": 1.4586315382373547, + "learning_rate": 1.8920795669001076e-07, + "loss": 0.6196, + "step": 3775 + }, + { + "epoch": 6.992592592592593, + "grad_norm": 1.4401381843446186, + "learning_rate": 1.8920118011583765e-07, + "loss": 0.6488, + "step": 3776 + }, + { + "epoch": 6.9944444444444445, + "grad_norm": 1.976729657997227, + "learning_rate": 1.891944015361797e-07, + "loss": 0.6743, + "step": 3777 + }, + { + "epoch": 6.996296296296296, + "grad_norm": 1.9011699693512465, + "learning_rate": 1.891876209511894e-07, + "loss": 0.6468, + "step": 3778 + }, + { + "epoch": 6.9981481481481485, + "grad_norm": 1.5686409897996878, + "learning_rate": 1.8918083836101918e-07, + "loss": 0.6572, + "step": 3779 + }, + { + "epoch": 7.0, + "grad_norm": 1.584022508893578, + "learning_rate": 1.8917405376582142e-07, + "loss": 0.6883, + "step": 3780 + }, + { + "epoch": 7.0018518518518515, + "grad_norm": 1.5148635158449415, + "learning_rate": 1.891672671657488e-07, + "loss": 0.6761, + "step": 3781 + }, + { + "epoch": 7.003703703703704, + "grad_norm": 1.6201043954948564, + "learning_rate": 1.8916047856095384e-07, + "loss": 0.6561, + "step": 3782 + }, + { + "epoch": 7.0055555555555555, + "grad_norm": 1.319132924367095, + "learning_rate": 1.891536879515891e-07, + "loss": 0.6457, + "step": 3783 + }, + { + "epoch": 7.007407407407407, + "grad_norm": 1.9014528315996195, + "learning_rate": 1.891468953378074e-07, + "loss": 0.6261, + "step": 3784 + }, + { + "epoch": 7.0092592592592595, + "grad_norm": 1.904225659480633, + "learning_rate": 1.8914010071976133e-07, + "loss": 0.641, + "step": 3785 + }, + { + "epoch": 7.011111111111111, + "grad_norm": 1.4401106113315392, + "learning_rate": 1.8913330409760368e-07, + "loss": 0.6585, + "step": 3786 + }, + { + "epoch": 7.012962962962963, + "grad_norm": 1.4646495180750831, + "learning_rate": 1.8912650547148729e-07, + "loss": 0.6707, + "step": 3787 + }, + { + "epoch": 7.014814814814815, + "grad_norm": 2.010739386577859, + "learning_rate": 1.89119704841565e-07, + "loss": 0.6587, + "step": 3788 + }, + { + "epoch": 7.016666666666667, + "grad_norm": 1.6411988059033766, + "learning_rate": 1.891129022079897e-07, + "loss": 0.6316, + "step": 3789 + }, + { + "epoch": 7.018518518518518, + "grad_norm": 1.6861889459666546, + "learning_rate": 1.8910609757091432e-07, + "loss": 0.6351, + "step": 3790 + }, + { + "epoch": 7.020370370370371, + "grad_norm": 1.5455541701430804, + "learning_rate": 1.8909929093049186e-07, + "loss": 0.6631, + "step": 3791 + }, + { + "epoch": 7.022222222222222, + "grad_norm": 2.4503574803448123, + "learning_rate": 1.8909248228687535e-07, + "loss": 0.6903, + "step": 3792 + }, + { + "epoch": 7.024074074074074, + "grad_norm": 1.7845738642128253, + "learning_rate": 1.8908567164021786e-07, + "loss": 0.6531, + "step": 3793 + }, + { + "epoch": 7.025925925925926, + "grad_norm": 1.6139964760581305, + "learning_rate": 1.8907885899067254e-07, + "loss": 0.6728, + "step": 3794 + }, + { + "epoch": 7.027777777777778, + "grad_norm": 2.177508652230196, + "learning_rate": 1.8907204433839254e-07, + "loss": 0.6686, + "step": 3795 + }, + { + "epoch": 7.029629629629629, + "grad_norm": 1.6185494915398189, + "learning_rate": 1.8906522768353103e-07, + "loss": 0.6415, + "step": 3796 + }, + { + "epoch": 7.031481481481482, + "grad_norm": 1.2830642330663806, + "learning_rate": 1.8905840902624134e-07, + "loss": 0.6777, + "step": 3797 + }, + { + "epoch": 7.033333333333333, + "grad_norm": 1.5491670956500094, + "learning_rate": 1.8905158836667673e-07, + "loss": 0.6218, + "step": 3798 + }, + { + "epoch": 7.035185185185185, + "grad_norm": 1.5181712339371882, + "learning_rate": 1.8904476570499056e-07, + "loss": 0.6292, + "step": 3799 + }, + { + "epoch": 7.037037037037037, + "grad_norm": 1.7479168469024826, + "learning_rate": 1.8903794104133616e-07, + "loss": 0.6496, + "step": 3800 + }, + { + "epoch": 7.038888888888889, + "grad_norm": 2.322662597891969, + "learning_rate": 1.8903111437586707e-07, + "loss": 0.6502, + "step": 3801 + }, + { + "epoch": 7.040740740740741, + "grad_norm": 1.4617329741188945, + "learning_rate": 1.8902428570873675e-07, + "loss": 0.6729, + "step": 3802 + }, + { + "epoch": 7.042592592592593, + "grad_norm": 1.5448420207988391, + "learning_rate": 1.8901745504009866e-07, + "loss": 0.6436, + "step": 3803 + }, + { + "epoch": 7.044444444444444, + "grad_norm": 1.424090275074607, + "learning_rate": 1.8901062237010642e-07, + "loss": 0.6442, + "step": 3804 + }, + { + "epoch": 7.046296296296297, + "grad_norm": 1.4587107661069902, + "learning_rate": 1.8900378769891363e-07, + "loss": 0.6727, + "step": 3805 + }, + { + "epoch": 7.048148148148148, + "grad_norm": 1.4006843536790925, + "learning_rate": 1.8899695102667396e-07, + "loss": 0.6243, + "step": 3806 + }, + { + "epoch": 7.05, + "grad_norm": 1.6143833895325665, + "learning_rate": 1.8899011235354113e-07, + "loss": 0.6358, + "step": 3807 + }, + { + "epoch": 7.051851851851852, + "grad_norm": 1.3807486570238263, + "learning_rate": 1.8898327167966887e-07, + "loss": 0.6564, + "step": 3808 + }, + { + "epoch": 7.053703703703704, + "grad_norm": 1.3984914690669097, + "learning_rate": 1.8897642900521103e-07, + "loss": 0.6382, + "step": 3809 + }, + { + "epoch": 7.055555555555555, + "grad_norm": 1.4233921684234079, + "learning_rate": 1.8896958433032139e-07, + "loss": 0.6165, + "step": 3810 + }, + { + "epoch": 7.057407407407408, + "grad_norm": 1.5496563254416587, + "learning_rate": 1.8896273765515383e-07, + "loss": 0.6366, + "step": 3811 + }, + { + "epoch": 7.059259259259259, + "grad_norm": 2.7940927374900686, + "learning_rate": 1.8895588897986233e-07, + "loss": 0.693, + "step": 3812 + }, + { + "epoch": 7.061111111111111, + "grad_norm": 1.5990704089133272, + "learning_rate": 1.8894903830460082e-07, + "loss": 0.7033, + "step": 3813 + }, + { + "epoch": 7.062962962962963, + "grad_norm": 1.4485419736912386, + "learning_rate": 1.8894218562952336e-07, + "loss": 0.6842, + "step": 3814 + }, + { + "epoch": 7.064814814814815, + "grad_norm": 1.3575841490905451, + "learning_rate": 1.8893533095478403e-07, + "loss": 0.682, + "step": 3815 + }, + { + "epoch": 7.066666666666666, + "grad_norm": 1.4247753726405616, + "learning_rate": 1.889284742805369e-07, + "loss": 0.6481, + "step": 3816 + }, + { + "epoch": 7.068518518518519, + "grad_norm": 1.5431771573911641, + "learning_rate": 1.8892161560693612e-07, + "loss": 0.6602, + "step": 3817 + }, + { + "epoch": 7.07037037037037, + "grad_norm": 1.4913059094015684, + "learning_rate": 1.8891475493413594e-07, + "loss": 0.6524, + "step": 3818 + }, + { + "epoch": 7.072222222222222, + "grad_norm": 1.4411822425317622, + "learning_rate": 1.8890789226229055e-07, + "loss": 0.6269, + "step": 3819 + }, + { + "epoch": 7.074074074074074, + "grad_norm": 1.6310596446757368, + "learning_rate": 1.889010275915543e-07, + "loss": 0.6624, + "step": 3820 + }, + { + "epoch": 7.075925925925926, + "grad_norm": 1.4582472932053814, + "learning_rate": 1.888941609220815e-07, + "loss": 0.6399, + "step": 3821 + }, + { + "epoch": 7.0777777777777775, + "grad_norm": 1.6701647403780597, + "learning_rate": 1.8888729225402652e-07, + "loss": 0.6757, + "step": 3822 + }, + { + "epoch": 7.07962962962963, + "grad_norm": 1.4741038380070752, + "learning_rate": 1.888804215875438e-07, + "loss": 0.6725, + "step": 3823 + }, + { + "epoch": 7.0814814814814815, + "grad_norm": 1.354451323753022, + "learning_rate": 1.8887354892278782e-07, + "loss": 0.6461, + "step": 3824 + }, + { + "epoch": 7.083333333333333, + "grad_norm": 1.6106308005765915, + "learning_rate": 1.8886667425991307e-07, + "loss": 0.6099, + "step": 3825 + }, + { + "epoch": 7.0851851851851855, + "grad_norm": 1.404362210362807, + "learning_rate": 1.888597975990741e-07, + "loss": 0.6516, + "step": 3826 + }, + { + "epoch": 7.087037037037037, + "grad_norm": 1.9192295923238643, + "learning_rate": 1.8885291894042558e-07, + "loss": 0.653, + "step": 3827 + }, + { + "epoch": 7.088888888888889, + "grad_norm": 2.357469502616754, + "learning_rate": 1.8884603828412212e-07, + "loss": 0.6408, + "step": 3828 + }, + { + "epoch": 7.090740740740741, + "grad_norm": 1.6600772019635135, + "learning_rate": 1.8883915563031837e-07, + "loss": 0.6639, + "step": 3829 + }, + { + "epoch": 7.092592592592593, + "grad_norm": 1.65294751230379, + "learning_rate": 1.8883227097916917e-07, + "loss": 0.6489, + "step": 3830 + }, + { + "epoch": 7.094444444444444, + "grad_norm": 1.4917645208882038, + "learning_rate": 1.8882538433082922e-07, + "loss": 0.7064, + "step": 3831 + }, + { + "epoch": 7.0962962962962965, + "grad_norm": 1.483161226487801, + "learning_rate": 1.888184956854534e-07, + "loss": 0.6752, + "step": 3832 + }, + { + "epoch": 7.098148148148148, + "grad_norm": 1.6625645182672377, + "learning_rate": 1.8881160504319657e-07, + "loss": 0.6225, + "step": 3833 + }, + { + "epoch": 7.1, + "grad_norm": 1.6628820010557193, + "learning_rate": 1.8880471240421364e-07, + "loss": 0.6582, + "step": 3834 + }, + { + "epoch": 7.101851851851852, + "grad_norm": 1.3936728465431745, + "learning_rate": 1.887978177686596e-07, + "loss": 0.6603, + "step": 3835 + }, + { + "epoch": 7.103703703703704, + "grad_norm": 3.1120663049006723, + "learning_rate": 1.8879092113668943e-07, + "loss": 0.6477, + "step": 3836 + }, + { + "epoch": 7.105555555555555, + "grad_norm": 1.3528474015908034, + "learning_rate": 1.8878402250845821e-07, + "loss": 0.6973, + "step": 3837 + }, + { + "epoch": 7.107407407407408, + "grad_norm": 1.6064333266934185, + "learning_rate": 1.8877712188412104e-07, + "loss": 0.6668, + "step": 3838 + }, + { + "epoch": 7.109259259259259, + "grad_norm": 1.4184827569603229, + "learning_rate": 1.8877021926383302e-07, + "loss": 0.6415, + "step": 3839 + }, + { + "epoch": 7.111111111111111, + "grad_norm": 2.2987675229885323, + "learning_rate": 1.8876331464774942e-07, + "loss": 0.6212, + "step": 3840 + }, + { + "epoch": 7.112962962962963, + "grad_norm": 1.6548802174221124, + "learning_rate": 1.887564080360254e-07, + "loss": 0.6581, + "step": 3841 + }, + { + "epoch": 7.114814814814815, + "grad_norm": 1.979919832157401, + "learning_rate": 1.8874949942881625e-07, + "loss": 0.6536, + "step": 3842 + }, + { + "epoch": 7.116666666666666, + "grad_norm": 2.0173517985854117, + "learning_rate": 1.8874258882627733e-07, + "loss": 0.6358, + "step": 3843 + }, + { + "epoch": 7.118518518518519, + "grad_norm": 1.5985849065778481, + "learning_rate": 1.8873567622856402e-07, + "loss": 0.684, + "step": 3844 + }, + { + "epoch": 7.12037037037037, + "grad_norm": 1.6315226586683655, + "learning_rate": 1.887287616358317e-07, + "loss": 0.6514, + "step": 3845 + }, + { + "epoch": 7.122222222222222, + "grad_norm": 1.6036906764410606, + "learning_rate": 1.887218450482358e-07, + "loss": 0.6366, + "step": 3846 + }, + { + "epoch": 7.124074074074074, + "grad_norm": 1.8699340968748386, + "learning_rate": 1.887149264659319e-07, + "loss": 0.6668, + "step": 3847 + }, + { + "epoch": 7.125925925925926, + "grad_norm": 1.4819708051686205, + "learning_rate": 1.8870800588907547e-07, + "loss": 0.6476, + "step": 3848 + }, + { + "epoch": 7.127777777777778, + "grad_norm": 1.855927403652474, + "learning_rate": 1.8870108331782214e-07, + "loss": 0.6722, + "step": 3849 + }, + { + "epoch": 7.12962962962963, + "grad_norm": 1.6502540815895659, + "learning_rate": 1.886941587523276e-07, + "loss": 0.6174, + "step": 3850 + }, + { + "epoch": 7.131481481481481, + "grad_norm": 6.27316503381482, + "learning_rate": 1.8868723219274742e-07, + "loss": 0.6554, + "step": 3851 + }, + { + "epoch": 7.133333333333334, + "grad_norm": 1.7578849986145926, + "learning_rate": 1.8868030363923743e-07, + "loss": 0.6433, + "step": 3852 + }, + { + "epoch": 7.135185185185185, + "grad_norm": 1.6964535594880703, + "learning_rate": 1.8867337309195336e-07, + "loss": 0.6472, + "step": 3853 + }, + { + "epoch": 7.137037037037037, + "grad_norm": 1.5635892556909845, + "learning_rate": 1.88666440551051e-07, + "loss": 0.6939, + "step": 3854 + }, + { + "epoch": 7.138888888888889, + "grad_norm": 1.6794114761553867, + "learning_rate": 1.886595060166863e-07, + "loss": 0.6199, + "step": 3855 + }, + { + "epoch": 7.140740740740741, + "grad_norm": 1.4944608806373647, + "learning_rate": 1.8865256948901506e-07, + "loss": 0.6907, + "step": 3856 + }, + { + "epoch": 7.142592592592592, + "grad_norm": 1.5058651462008914, + "learning_rate": 1.8864563096819327e-07, + "loss": 0.676, + "step": 3857 + }, + { + "epoch": 7.144444444444445, + "grad_norm": 1.7225456820470928, + "learning_rate": 1.8863869045437694e-07, + "loss": 0.6401, + "step": 3858 + }, + { + "epoch": 7.146296296296296, + "grad_norm": 1.8313228140413405, + "learning_rate": 1.8863174794772214e-07, + "loss": 0.6427, + "step": 3859 + }, + { + "epoch": 7.148148148148148, + "grad_norm": 1.4074709176373037, + "learning_rate": 1.8862480344838492e-07, + "loss": 0.6611, + "step": 3860 + }, + { + "epoch": 7.15, + "grad_norm": 1.4773359487564475, + "learning_rate": 1.8861785695652141e-07, + "loss": 0.6293, + "step": 3861 + }, + { + "epoch": 7.151851851851852, + "grad_norm": 1.6334113374766588, + "learning_rate": 1.886109084722878e-07, + "loss": 0.6487, + "step": 3862 + }, + { + "epoch": 7.1537037037037035, + "grad_norm": 1.5871776141561962, + "learning_rate": 1.8860395799584027e-07, + "loss": 0.6684, + "step": 3863 + }, + { + "epoch": 7.155555555555556, + "grad_norm": 1.7675143086737106, + "learning_rate": 1.8859700552733515e-07, + "loss": 0.6612, + "step": 3864 + }, + { + "epoch": 7.157407407407407, + "grad_norm": 1.5625901916181437, + "learning_rate": 1.8859005106692873e-07, + "loss": 0.6959, + "step": 3865 + }, + { + "epoch": 7.159259259259259, + "grad_norm": 1.5362406437161675, + "learning_rate": 1.8858309461477735e-07, + "loss": 0.6408, + "step": 3866 + }, + { + "epoch": 7.161111111111111, + "grad_norm": 1.4598016385756507, + "learning_rate": 1.885761361710374e-07, + "loss": 0.6798, + "step": 3867 + }, + { + "epoch": 7.162962962962963, + "grad_norm": 1.5998399467113293, + "learning_rate": 1.8856917573586535e-07, + "loss": 0.6379, + "step": 3868 + }, + { + "epoch": 7.1648148148148145, + "grad_norm": 1.4686353698298469, + "learning_rate": 1.885622133094177e-07, + "loss": 0.631, + "step": 3869 + }, + { + "epoch": 7.166666666666667, + "grad_norm": 1.536428113332608, + "learning_rate": 1.8855524889185093e-07, + "loss": 0.6599, + "step": 3870 + }, + { + "epoch": 7.1685185185185185, + "grad_norm": 1.5964128859774331, + "learning_rate": 1.8854828248332165e-07, + "loss": 0.6348, + "step": 3871 + }, + { + "epoch": 7.17037037037037, + "grad_norm": 1.5687555028743951, + "learning_rate": 1.8854131408398651e-07, + "loss": 0.6582, + "step": 3872 + }, + { + "epoch": 7.1722222222222225, + "grad_norm": 1.5339340899242777, + "learning_rate": 1.885343436940021e-07, + "loss": 0.6743, + "step": 3873 + }, + { + "epoch": 7.174074074074074, + "grad_norm": 1.413969126390241, + "learning_rate": 1.8852737131352528e-07, + "loss": 0.6057, + "step": 3874 + }, + { + "epoch": 7.175925925925926, + "grad_norm": 1.389578942011871, + "learning_rate": 1.8852039694271265e-07, + "loss": 0.6284, + "step": 3875 + }, + { + "epoch": 7.177777777777778, + "grad_norm": 1.7142042908403565, + "learning_rate": 1.885134205817211e-07, + "loss": 0.6913, + "step": 3876 + }, + { + "epoch": 7.17962962962963, + "grad_norm": 2.6299219853837736, + "learning_rate": 1.8850644223070745e-07, + "loss": 0.6222, + "step": 3877 + }, + { + "epoch": 7.181481481481481, + "grad_norm": 2.3473674730398035, + "learning_rate": 1.884994618898286e-07, + "loss": 0.6575, + "step": 3878 + }, + { + "epoch": 7.183333333333334, + "grad_norm": 1.630867989586954, + "learning_rate": 1.8849247955924147e-07, + "loss": 0.6414, + "step": 3879 + }, + { + "epoch": 7.185185185185185, + "grad_norm": 1.8145038461911651, + "learning_rate": 1.884854952391031e-07, + "loss": 0.6507, + "step": 3880 + }, + { + "epoch": 7.187037037037037, + "grad_norm": 1.5987641775808559, + "learning_rate": 1.8847850892957045e-07, + "loss": 0.6711, + "step": 3881 + }, + { + "epoch": 7.188888888888889, + "grad_norm": 2.696866273756317, + "learning_rate": 1.884715206308006e-07, + "loss": 0.6517, + "step": 3882 + }, + { + "epoch": 7.190740740740741, + "grad_norm": 8.973496167083082, + "learning_rate": 1.884645303429507e-07, + "loss": 0.6508, + "step": 3883 + }, + { + "epoch": 7.192592592592592, + "grad_norm": 1.6659351260990651, + "learning_rate": 1.8845753806617791e-07, + "loss": 0.6288, + "step": 3884 + }, + { + "epoch": 7.194444444444445, + "grad_norm": 1.4719304183891508, + "learning_rate": 1.8845054380063937e-07, + "loss": 0.6818, + "step": 3885 + }, + { + "epoch": 7.196296296296296, + "grad_norm": 1.4892986386424063, + "learning_rate": 1.884435475464924e-07, + "loss": 0.643, + "step": 3886 + }, + { + "epoch": 7.198148148148148, + "grad_norm": 1.5833156254768301, + "learning_rate": 1.8843654930389429e-07, + "loss": 0.5923, + "step": 3887 + }, + { + "epoch": 7.2, + "grad_norm": 1.5900125590073069, + "learning_rate": 1.8842954907300233e-07, + "loss": 0.6475, + "step": 3888 + }, + { + "epoch": 7.201851851851852, + "grad_norm": 1.9049173073075365, + "learning_rate": 1.8842254685397397e-07, + "loss": 0.6222, + "step": 3889 + }, + { + "epoch": 7.203703703703703, + "grad_norm": 1.2899176724558696, + "learning_rate": 1.884155426469666e-07, + "loss": 0.6403, + "step": 3890 + }, + { + "epoch": 7.205555555555556, + "grad_norm": 1.6366894540785846, + "learning_rate": 1.8840853645213768e-07, + "loss": 0.6708, + "step": 3891 + }, + { + "epoch": 7.207407407407407, + "grad_norm": 1.277917052255994, + "learning_rate": 1.8840152826964474e-07, + "loss": 0.6065, + "step": 3892 + }, + { + "epoch": 7.209259259259259, + "grad_norm": 1.9130971269001955, + "learning_rate": 1.8839451809964537e-07, + "loss": 0.6875, + "step": 3893 + }, + { + "epoch": 7.211111111111111, + "grad_norm": 1.522689540944064, + "learning_rate": 1.8838750594229716e-07, + "loss": 0.6781, + "step": 3894 + }, + { + "epoch": 7.212962962962963, + "grad_norm": 1.499668882598213, + "learning_rate": 1.8838049179775774e-07, + "loss": 0.6156, + "step": 3895 + }, + { + "epoch": 7.214814814814815, + "grad_norm": 1.5689321670663572, + "learning_rate": 1.8837347566618481e-07, + "loss": 0.643, + "step": 3896 + }, + { + "epoch": 7.216666666666667, + "grad_norm": 1.4867489809770138, + "learning_rate": 1.8836645754773616e-07, + "loss": 0.6527, + "step": 3897 + }, + { + "epoch": 7.218518518518518, + "grad_norm": 1.5481088621741936, + "learning_rate": 1.8835943744256958e-07, + "loss": 0.6546, + "step": 3898 + }, + { + "epoch": 7.220370370370371, + "grad_norm": 1.6518008488001963, + "learning_rate": 1.883524153508428e-07, + "loss": 0.6025, + "step": 3899 + }, + { + "epoch": 7.222222222222222, + "grad_norm": 1.4935660192485618, + "learning_rate": 1.8834539127271377e-07, + "loss": 0.613, + "step": 3900 + }, + { + "epoch": 7.224074074074074, + "grad_norm": 1.6446163565824925, + "learning_rate": 1.8833836520834044e-07, + "loss": 0.6513, + "step": 3901 + }, + { + "epoch": 7.225925925925926, + "grad_norm": 1.6544292092216015, + "learning_rate": 1.883313371578807e-07, + "loss": 0.6419, + "step": 3902 + }, + { + "epoch": 7.227777777777778, + "grad_norm": 1.611398208604589, + "learning_rate": 1.883243071214926e-07, + "loss": 0.6145, + "step": 3903 + }, + { + "epoch": 7.229629629629629, + "grad_norm": 1.570677382116332, + "learning_rate": 1.8831727509933422e-07, + "loss": 0.6554, + "step": 3904 + }, + { + "epoch": 7.231481481481482, + "grad_norm": 1.698686244968201, + "learning_rate": 1.8831024109156362e-07, + "loss": 0.6535, + "step": 3905 + }, + { + "epoch": 7.233333333333333, + "grad_norm": 1.3183208014599064, + "learning_rate": 1.8830320509833895e-07, + "loss": 0.6301, + "step": 3906 + }, + { + "epoch": 7.235185185185185, + "grad_norm": 1.3986975525552585, + "learning_rate": 1.8829616711981837e-07, + "loss": 0.6218, + "step": 3907 + }, + { + "epoch": 7.237037037037037, + "grad_norm": 1.366057077596558, + "learning_rate": 1.882891271561602e-07, + "loss": 0.6761, + "step": 3908 + }, + { + "epoch": 7.238888888888889, + "grad_norm": 1.436074307512177, + "learning_rate": 1.882820852075226e-07, + "loss": 0.699, + "step": 3909 + }, + { + "epoch": 7.2407407407407405, + "grad_norm": 1.7941992120343393, + "learning_rate": 1.88275041274064e-07, + "loss": 0.6536, + "step": 3910 + }, + { + "epoch": 7.242592592592593, + "grad_norm": 1.4959755979091727, + "learning_rate": 1.8826799535594268e-07, + "loss": 0.6936, + "step": 3911 + }, + { + "epoch": 7.2444444444444445, + "grad_norm": 1.4669996445640352, + "learning_rate": 1.8826094745331714e-07, + "loss": 0.6804, + "step": 3912 + }, + { + "epoch": 7.246296296296296, + "grad_norm": 1.4510489047234636, + "learning_rate": 1.8825389756634577e-07, + "loss": 0.634, + "step": 3913 + }, + { + "epoch": 7.2481481481481485, + "grad_norm": 1.507667624718648, + "learning_rate": 1.8824684569518707e-07, + "loss": 0.6412, + "step": 3914 + }, + { + "epoch": 7.25, + "grad_norm": 2.26204919181382, + "learning_rate": 1.8823979183999964e-07, + "loss": 0.6346, + "step": 3915 + }, + { + "epoch": 7.2518518518518515, + "grad_norm": 2.2553876605011234, + "learning_rate": 1.8823273600094202e-07, + "loss": 0.6458, + "step": 3916 + }, + { + "epoch": 7.253703703703704, + "grad_norm": 1.4143685791332317, + "learning_rate": 1.8822567817817286e-07, + "loss": 0.6379, + "step": 3917 + }, + { + "epoch": 7.2555555555555555, + "grad_norm": 1.3817341785013824, + "learning_rate": 1.8821861837185082e-07, + "loss": 0.6528, + "step": 3918 + }, + { + "epoch": 7.257407407407407, + "grad_norm": 2.261792719811734, + "learning_rate": 1.8821155658213464e-07, + "loss": 0.6799, + "step": 3919 + }, + { + "epoch": 7.2592592592592595, + "grad_norm": 1.6073797206949885, + "learning_rate": 1.882044928091831e-07, + "loss": 0.6158, + "step": 3920 + }, + { + "epoch": 7.261111111111111, + "grad_norm": 1.4764509201538014, + "learning_rate": 1.88197427053155e-07, + "loss": 0.6289, + "step": 3921 + }, + { + "epoch": 7.262962962962963, + "grad_norm": 1.5525193173261245, + "learning_rate": 1.8819035931420923e-07, + "loss": 0.6345, + "step": 3922 + }, + { + "epoch": 7.264814814814815, + "grad_norm": 1.3764809218195233, + "learning_rate": 1.881832895925046e-07, + "loss": 0.655, + "step": 3923 + }, + { + "epoch": 7.266666666666667, + "grad_norm": 1.7909947471470726, + "learning_rate": 1.8817621788820015e-07, + "loss": 0.6961, + "step": 3924 + }, + { + "epoch": 7.268518518518518, + "grad_norm": 1.4945584351625694, + "learning_rate": 1.8816914420145484e-07, + "loss": 0.6441, + "step": 3925 + }, + { + "epoch": 7.270370370370371, + "grad_norm": 1.4756961611301094, + "learning_rate": 1.881620685324277e-07, + "loss": 0.604, + "step": 3926 + }, + { + "epoch": 7.272222222222222, + "grad_norm": 1.6034171479543946, + "learning_rate": 1.8815499088127782e-07, + "loss": 0.6652, + "step": 3927 + }, + { + "epoch": 7.274074074074074, + "grad_norm": 1.6979320194310967, + "learning_rate": 1.8814791124816431e-07, + "loss": 0.6324, + "step": 3928 + }, + { + "epoch": 7.275925925925926, + "grad_norm": 1.488771109320376, + "learning_rate": 1.881408296332464e-07, + "loss": 0.6637, + "step": 3929 + }, + { + "epoch": 7.277777777777778, + "grad_norm": 1.7298965202641352, + "learning_rate": 1.8813374603668318e-07, + "loss": 0.6605, + "step": 3930 + }, + { + "epoch": 7.279629629629629, + "grad_norm": 1.7955800628411287, + "learning_rate": 1.8812666045863402e-07, + "loss": 0.6553, + "step": 3931 + }, + { + "epoch": 7.281481481481482, + "grad_norm": 1.536856073088939, + "learning_rate": 1.8811957289925817e-07, + "loss": 0.7155, + "step": 3932 + }, + { + "epoch": 7.283333333333333, + "grad_norm": 4.672029829624969, + "learning_rate": 1.88112483358715e-07, + "loss": 0.6273, + "step": 3933 + }, + { + "epoch": 7.285185185185185, + "grad_norm": 2.058608757288798, + "learning_rate": 1.8810539183716387e-07, + "loss": 0.6399, + "step": 3934 + }, + { + "epoch": 7.287037037037037, + "grad_norm": 1.7895664318485744, + "learning_rate": 1.8809829833476427e-07, + "loss": 0.6193, + "step": 3935 + }, + { + "epoch": 7.288888888888889, + "grad_norm": 1.6735787370584698, + "learning_rate": 1.8809120285167565e-07, + "loss": 0.6822, + "step": 3936 + }, + { + "epoch": 7.290740740740741, + "grad_norm": 1.6569132458058724, + "learning_rate": 1.8808410538805752e-07, + "loss": 0.6397, + "step": 3937 + }, + { + "epoch": 7.292592592592593, + "grad_norm": 1.6423860174800282, + "learning_rate": 1.880770059440695e-07, + "loss": 0.683, + "step": 3938 + }, + { + "epoch": 7.294444444444444, + "grad_norm": 1.928869152293549, + "learning_rate": 1.8806990451987112e-07, + "loss": 0.6339, + "step": 3939 + }, + { + "epoch": 7.296296296296296, + "grad_norm": 1.9896247663473383, + "learning_rate": 1.8806280111562214e-07, + "loss": 0.6946, + "step": 3940 + }, + { + "epoch": 7.298148148148148, + "grad_norm": 1.7409612936664953, + "learning_rate": 1.880556957314822e-07, + "loss": 0.6492, + "step": 3941 + }, + { + "epoch": 7.3, + "grad_norm": 1.7510285567339543, + "learning_rate": 1.8804858836761106e-07, + "loss": 0.6583, + "step": 3942 + }, + { + "epoch": 7.301851851851852, + "grad_norm": 1.5451859824210152, + "learning_rate": 1.8804147902416855e-07, + "loss": 0.6356, + "step": 3943 + }, + { + "epoch": 7.303703703703704, + "grad_norm": 1.362482648982307, + "learning_rate": 1.8803436770131444e-07, + "loss": 0.6415, + "step": 3944 + }, + { + "epoch": 7.305555555555555, + "grad_norm": 1.4676534218636943, + "learning_rate": 1.8802725439920865e-07, + "loss": 0.6331, + "step": 3945 + }, + { + "epoch": 7.307407407407408, + "grad_norm": 1.6048907731684041, + "learning_rate": 1.8802013911801108e-07, + "loss": 0.684, + "step": 3946 + }, + { + "epoch": 7.309259259259259, + "grad_norm": 1.6965847442519473, + "learning_rate": 1.880130218578818e-07, + "loss": 0.6736, + "step": 3947 + }, + { + "epoch": 7.311111111111111, + "grad_norm": 1.6076566054676888, + "learning_rate": 1.880059026189807e-07, + "loss": 0.6086, + "step": 3948 + }, + { + "epoch": 7.312962962962963, + "grad_norm": 1.5311607599184833, + "learning_rate": 1.8799878140146785e-07, + "loss": 0.59, + "step": 3949 + }, + { + "epoch": 7.314814814814815, + "grad_norm": 1.3828818058044876, + "learning_rate": 1.8799165820550344e-07, + "loss": 0.6213, + "step": 3950 + }, + { + "epoch": 7.316666666666666, + "grad_norm": 1.5412514403839601, + "learning_rate": 1.8798453303124755e-07, + "loss": 0.6459, + "step": 3951 + }, + { + "epoch": 7.318518518518519, + "grad_norm": 1.6610813797053923, + "learning_rate": 1.8797740587886043e-07, + "loss": 0.6128, + "step": 3952 + }, + { + "epoch": 7.32037037037037, + "grad_norm": 1.9076274754200053, + "learning_rate": 1.8797027674850227e-07, + "loss": 0.6092, + "step": 3953 + }, + { + "epoch": 7.322222222222222, + "grad_norm": 1.5040723671179161, + "learning_rate": 1.8796314564033335e-07, + "loss": 0.6535, + "step": 3954 + }, + { + "epoch": 7.324074074074074, + "grad_norm": 1.8666699580720543, + "learning_rate": 1.8795601255451402e-07, + "loss": 0.6415, + "step": 3955 + }, + { + "epoch": 7.325925925925926, + "grad_norm": 1.7952492974322916, + "learning_rate": 1.8794887749120464e-07, + "loss": 0.6422, + "step": 3956 + }, + { + "epoch": 7.3277777777777775, + "grad_norm": 1.6526668294092484, + "learning_rate": 1.879417404505656e-07, + "loss": 0.6249, + "step": 3957 + }, + { + "epoch": 7.32962962962963, + "grad_norm": 1.5771963090925472, + "learning_rate": 1.8793460143275742e-07, + "loss": 0.6532, + "step": 3958 + }, + { + "epoch": 7.3314814814814815, + "grad_norm": 1.8678657727234584, + "learning_rate": 1.8792746043794058e-07, + "loss": 0.6232, + "step": 3959 + }, + { + "epoch": 7.333333333333333, + "grad_norm": 1.7589011450543741, + "learning_rate": 1.879203174662756e-07, + "loss": 0.6621, + "step": 3960 + }, + { + "epoch": 7.3351851851851855, + "grad_norm": 1.6241955707340285, + "learning_rate": 1.8791317251792312e-07, + "loss": 0.6623, + "step": 3961 + }, + { + "epoch": 7.337037037037037, + "grad_norm": 1.9852262336269557, + "learning_rate": 1.8790602559304377e-07, + "loss": 0.6543, + "step": 3962 + }, + { + "epoch": 7.338888888888889, + "grad_norm": 1.7617850997132742, + "learning_rate": 1.8789887669179815e-07, + "loss": 0.6607, + "step": 3963 + }, + { + "epoch": 7.340740740740741, + "grad_norm": 1.6416548301433183, + "learning_rate": 1.878917258143471e-07, + "loss": 0.6322, + "step": 3964 + }, + { + "epoch": 7.342592592592593, + "grad_norm": 1.4702170252549254, + "learning_rate": 1.878845729608513e-07, + "loss": 0.6448, + "step": 3965 + }, + { + "epoch": 7.344444444444444, + "grad_norm": 2.0293644914716795, + "learning_rate": 1.8787741813147168e-07, + "loss": 0.6644, + "step": 3966 + }, + { + "epoch": 7.3462962962962965, + "grad_norm": 1.5609067834681374, + "learning_rate": 1.87870261326369e-07, + "loss": 0.6344, + "step": 3967 + }, + { + "epoch": 7.348148148148148, + "grad_norm": 1.6320911450561009, + "learning_rate": 1.8786310254570417e-07, + "loss": 0.6216, + "step": 3968 + }, + { + "epoch": 7.35, + "grad_norm": 1.530596755852607, + "learning_rate": 1.8785594178963816e-07, + "loss": 0.6444, + "step": 3969 + }, + { + "epoch": 7.351851851851852, + "grad_norm": 1.4486525552744287, + "learning_rate": 1.8784877905833198e-07, + "loss": 0.6599, + "step": 3970 + }, + { + "epoch": 7.353703703703704, + "grad_norm": 1.6468175125216373, + "learning_rate": 1.8784161435194666e-07, + "loss": 0.6535, + "step": 3971 + }, + { + "epoch": 7.355555555555555, + "grad_norm": 2.515406861929086, + "learning_rate": 1.8783444767064328e-07, + "loss": 0.6162, + "step": 3972 + }, + { + "epoch": 7.357407407407408, + "grad_norm": 2.1408575222742967, + "learning_rate": 1.8782727901458295e-07, + "loss": 0.6082, + "step": 3973 + }, + { + "epoch": 7.359259259259259, + "grad_norm": 1.5767584206411043, + "learning_rate": 1.8782010838392686e-07, + "loss": 0.6621, + "step": 3974 + }, + { + "epoch": 7.361111111111111, + "grad_norm": 1.5336110860037588, + "learning_rate": 1.878129357788362e-07, + "loss": 0.6119, + "step": 3975 + }, + { + "epoch": 7.362962962962963, + "grad_norm": 1.6740254455036625, + "learning_rate": 1.8780576119947228e-07, + "loss": 0.7035, + "step": 3976 + }, + { + "epoch": 7.364814814814815, + "grad_norm": 1.5989550949179896, + "learning_rate": 1.8779858464599635e-07, + "loss": 0.6406, + "step": 3977 + }, + { + "epoch": 7.366666666666666, + "grad_norm": 1.7439363606875136, + "learning_rate": 1.8779140611856977e-07, + "loss": 0.6565, + "step": 3978 + }, + { + "epoch": 7.368518518518519, + "grad_norm": 1.8878705061997887, + "learning_rate": 1.8778422561735393e-07, + "loss": 0.6085, + "step": 3979 + }, + { + "epoch": 7.37037037037037, + "grad_norm": 1.3416696649699746, + "learning_rate": 1.877770431425103e-07, + "loss": 0.6435, + "step": 3980 + }, + { + "epoch": 7.372222222222222, + "grad_norm": 1.6151147095024616, + "learning_rate": 1.8776985869420036e-07, + "loss": 0.6295, + "step": 3981 + }, + { + "epoch": 7.374074074074074, + "grad_norm": 1.4710999831483937, + "learning_rate": 1.877626722725856e-07, + "loss": 0.6368, + "step": 3982 + }, + { + "epoch": 7.375925925925926, + "grad_norm": 1.6531219984563443, + "learning_rate": 1.877554838778276e-07, + "loss": 0.6592, + "step": 3983 + }, + { + "epoch": 7.377777777777778, + "grad_norm": 1.530917705147961, + "learning_rate": 1.87748293510088e-07, + "loss": 0.6665, + "step": 3984 + }, + { + "epoch": 7.37962962962963, + "grad_norm": 1.7138727852005868, + "learning_rate": 1.8774110116952844e-07, + "loss": 0.6704, + "step": 3985 + }, + { + "epoch": 7.381481481481481, + "grad_norm": 1.560608128273509, + "learning_rate": 1.8773390685631061e-07, + "loss": 0.6577, + "step": 3986 + }, + { + "epoch": 7.383333333333334, + "grad_norm": 1.6224083625363526, + "learning_rate": 1.8772671057059628e-07, + "loss": 0.661, + "step": 3987 + }, + { + "epoch": 7.385185185185185, + "grad_norm": 1.5726732441112203, + "learning_rate": 1.8771951231254724e-07, + "loss": 0.6482, + "step": 3988 + }, + { + "epoch": 7.387037037037037, + "grad_norm": 1.9291963480221006, + "learning_rate": 1.8771231208232533e-07, + "loss": 0.6534, + "step": 3989 + }, + { + "epoch": 7.388888888888889, + "grad_norm": 1.5343596598027045, + "learning_rate": 1.8770510988009238e-07, + "loss": 0.6709, + "step": 3990 + }, + { + "epoch": 7.390740740740741, + "grad_norm": 1.2932356018626276, + "learning_rate": 1.876979057060104e-07, + "loss": 0.6167, + "step": 3991 + }, + { + "epoch": 7.392592592592592, + "grad_norm": 1.4489345231061077, + "learning_rate": 1.8769069956024128e-07, + "loss": 0.6377, + "step": 3992 + }, + { + "epoch": 7.394444444444445, + "grad_norm": 1.7032318991020898, + "learning_rate": 1.8768349144294707e-07, + "loss": 0.6524, + "step": 3993 + }, + { + "epoch": 7.396296296296296, + "grad_norm": 1.5067205342109846, + "learning_rate": 1.8767628135428985e-07, + "loss": 0.6809, + "step": 3994 + }, + { + "epoch": 7.398148148148148, + "grad_norm": 1.5282536219355214, + "learning_rate": 1.876690692944317e-07, + "loss": 0.6629, + "step": 3995 + }, + { + "epoch": 7.4, + "grad_norm": 1.5922591850406895, + "learning_rate": 1.8766185526353477e-07, + "loss": 0.6646, + "step": 3996 + }, + { + "epoch": 7.401851851851852, + "grad_norm": 1.5105664851385068, + "learning_rate": 1.8765463926176125e-07, + "loss": 0.6506, + "step": 3997 + }, + { + "epoch": 7.4037037037037035, + "grad_norm": 1.5136841683198523, + "learning_rate": 1.8764742128927335e-07, + "loss": 0.5898, + "step": 3998 + }, + { + "epoch": 7.405555555555556, + "grad_norm": 1.550094609955071, + "learning_rate": 1.8764020134623337e-07, + "loss": 0.6511, + "step": 3999 + }, + { + "epoch": 7.407407407407407, + "grad_norm": 1.3737429684767637, + "learning_rate": 1.8763297943280366e-07, + "loss": 0.6475, + "step": 4000 + }, + { + "epoch": 7.409259259259259, + "grad_norm": 1.7435083066201096, + "learning_rate": 1.8762575554914655e-07, + "loss": 0.6383, + "step": 4001 + }, + { + "epoch": 7.411111111111111, + "grad_norm": 1.4886156360198073, + "learning_rate": 1.8761852969542447e-07, + "loss": 0.6527, + "step": 4002 + }, + { + "epoch": 7.412962962962963, + "grad_norm": 1.5659926446768353, + "learning_rate": 1.8761130187179988e-07, + "loss": 0.6434, + "step": 4003 + }, + { + "epoch": 7.4148148148148145, + "grad_norm": 1.7098065567296454, + "learning_rate": 1.8760407207843526e-07, + "loss": 0.6387, + "step": 4004 + }, + { + "epoch": 7.416666666666667, + "grad_norm": 1.6000855220329502, + "learning_rate": 1.8759684031549318e-07, + "loss": 0.6396, + "step": 4005 + }, + { + "epoch": 7.4185185185185185, + "grad_norm": 1.465118207342098, + "learning_rate": 1.8758960658313626e-07, + "loss": 0.638, + "step": 4006 + }, + { + "epoch": 7.42037037037037, + "grad_norm": 1.552743290645851, + "learning_rate": 1.8758237088152705e-07, + "loss": 0.6299, + "step": 4007 + }, + { + "epoch": 7.4222222222222225, + "grad_norm": 1.6209530980176254, + "learning_rate": 1.8757513321082826e-07, + "loss": 0.686, + "step": 4008 + }, + { + "epoch": 7.424074074074074, + "grad_norm": 1.6088353083788376, + "learning_rate": 1.8756789357120266e-07, + "loss": 0.6762, + "step": 4009 + }, + { + "epoch": 7.425925925925926, + "grad_norm": 2.0756921880937975, + "learning_rate": 1.8756065196281295e-07, + "loss": 0.6266, + "step": 4010 + }, + { + "epoch": 7.427777777777778, + "grad_norm": 1.5774188227604389, + "learning_rate": 1.8755340838582198e-07, + "loss": 0.6501, + "step": 4011 + }, + { + "epoch": 7.42962962962963, + "grad_norm": 1.9601902196906416, + "learning_rate": 1.8754616284039262e-07, + "loss": 0.6418, + "step": 4012 + }, + { + "epoch": 7.431481481481481, + "grad_norm": 1.9357120800036363, + "learning_rate": 1.8753891532668775e-07, + "loss": 0.679, + "step": 4013 + }, + { + "epoch": 7.433333333333334, + "grad_norm": 2.043317406892933, + "learning_rate": 1.8753166584487028e-07, + "loss": 0.6242, + "step": 4014 + }, + { + "epoch": 7.435185185185185, + "grad_norm": 1.5645792562007772, + "learning_rate": 1.8752441439510328e-07, + "loss": 0.6295, + "step": 4015 + }, + { + "epoch": 7.437037037037037, + "grad_norm": 1.8893870468079825, + "learning_rate": 1.8751716097754966e-07, + "loss": 0.6574, + "step": 4016 + }, + { + "epoch": 7.438888888888889, + "grad_norm": 1.4123013040658945, + "learning_rate": 1.875099055923726e-07, + "loss": 0.6534, + "step": 4017 + }, + { + "epoch": 7.440740740740741, + "grad_norm": 1.4955145167966055, + "learning_rate": 1.8750264823973523e-07, + "loss": 0.6719, + "step": 4018 + }, + { + "epoch": 7.442592592592592, + "grad_norm": 1.370447461015703, + "learning_rate": 1.8749538891980062e-07, + "loss": 0.6786, + "step": 4019 + }, + { + "epoch": 7.444444444444445, + "grad_norm": 1.8380337806617724, + "learning_rate": 1.8748812763273205e-07, + "loss": 0.6852, + "step": 4020 + }, + { + "epoch": 7.446296296296296, + "grad_norm": 1.798180985712554, + "learning_rate": 1.8748086437869275e-07, + "loss": 0.6118, + "step": 4021 + }, + { + "epoch": 7.448148148148148, + "grad_norm": 1.5041488822934745, + "learning_rate": 1.8747359915784605e-07, + "loss": 0.6671, + "step": 4022 + }, + { + "epoch": 7.45, + "grad_norm": 1.4826549827362148, + "learning_rate": 1.8746633197035523e-07, + "loss": 0.6602, + "step": 4023 + }, + { + "epoch": 7.451851851851852, + "grad_norm": 1.534158352676765, + "learning_rate": 1.8745906281638375e-07, + "loss": 0.6479, + "step": 4024 + }, + { + "epoch": 7.453703703703704, + "grad_norm": 1.5152215479266407, + "learning_rate": 1.8745179169609497e-07, + "loss": 0.6468, + "step": 4025 + }, + { + "epoch": 7.455555555555556, + "grad_norm": 1.6936004984417925, + "learning_rate": 1.8744451860965243e-07, + "loss": 0.7017, + "step": 4026 + }, + { + "epoch": 7.457407407407407, + "grad_norm": 1.5532458965125615, + "learning_rate": 1.8743724355721959e-07, + "loss": 0.6633, + "step": 4027 + }, + { + "epoch": 7.459259259259259, + "grad_norm": 2.830682580661416, + "learning_rate": 1.8742996653896004e-07, + "loss": 0.6325, + "step": 4028 + }, + { + "epoch": 7.461111111111111, + "grad_norm": 1.5554722988885161, + "learning_rate": 1.874226875550374e-07, + "loss": 0.6591, + "step": 4029 + }, + { + "epoch": 7.462962962962963, + "grad_norm": 3.4003184089125456, + "learning_rate": 1.874154066056153e-07, + "loss": 0.6578, + "step": 4030 + }, + { + "epoch": 7.464814814814815, + "grad_norm": 1.4879026736168877, + "learning_rate": 1.8740812369085744e-07, + "loss": 0.6776, + "step": 4031 + }, + { + "epoch": 7.466666666666667, + "grad_norm": 1.8577202723662751, + "learning_rate": 1.8740083881092755e-07, + "loss": 0.6808, + "step": 4032 + }, + { + "epoch": 7.468518518518518, + "grad_norm": 1.6480056299091879, + "learning_rate": 1.8739355196598944e-07, + "loss": 0.6731, + "step": 4033 + }, + { + "epoch": 7.470370370370371, + "grad_norm": 1.4266394018792556, + "learning_rate": 1.8738626315620697e-07, + "loss": 0.6829, + "step": 4034 + }, + { + "epoch": 7.472222222222222, + "grad_norm": 1.7993016395786594, + "learning_rate": 1.8737897238174388e-07, + "loss": 0.6762, + "step": 4035 + }, + { + "epoch": 7.474074074074074, + "grad_norm": 1.7744971752791796, + "learning_rate": 1.8737167964276425e-07, + "loss": 0.6435, + "step": 4036 + }, + { + "epoch": 7.475925925925926, + "grad_norm": 1.5467923767374805, + "learning_rate": 1.873643849394319e-07, + "loss": 0.634, + "step": 4037 + }, + { + "epoch": 7.477777777777778, + "grad_norm": 1.416268413536496, + "learning_rate": 1.8735708827191095e-07, + "loss": 0.6659, + "step": 4038 + }, + { + "epoch": 7.479629629629629, + "grad_norm": 1.4636895234396003, + "learning_rate": 1.8734978964036537e-07, + "loss": 0.5964, + "step": 4039 + }, + { + "epoch": 7.481481481481482, + "grad_norm": 1.5593564915151255, + "learning_rate": 1.8734248904495928e-07, + "loss": 0.644, + "step": 4040 + }, + { + "epoch": 7.483333333333333, + "grad_norm": 2.0615276535442333, + "learning_rate": 1.8733518648585682e-07, + "loss": 0.6199, + "step": 4041 + }, + { + "epoch": 7.485185185185185, + "grad_norm": 1.5671530258906192, + "learning_rate": 1.873278819632222e-07, + "loss": 0.6332, + "step": 4042 + }, + { + "epoch": 7.487037037037037, + "grad_norm": 1.523895005421646, + "learning_rate": 1.873205754772196e-07, + "loss": 0.6585, + "step": 4043 + }, + { + "epoch": 7.488888888888889, + "grad_norm": 1.538715776387948, + "learning_rate": 1.8731326702801328e-07, + "loss": 0.655, + "step": 4044 + }, + { + "epoch": 7.4907407407407405, + "grad_norm": 1.5859609127860925, + "learning_rate": 1.873059566157676e-07, + "loss": 0.6518, + "step": 4045 + }, + { + "epoch": 7.492592592592593, + "grad_norm": 1.3829374655509887, + "learning_rate": 1.8729864424064685e-07, + "loss": 0.622, + "step": 4046 + }, + { + "epoch": 7.4944444444444445, + "grad_norm": 1.7939613736001132, + "learning_rate": 1.872913299028155e-07, + "loss": 0.6557, + "step": 4047 + }, + { + "epoch": 7.496296296296296, + "grad_norm": 1.7906214805864318, + "learning_rate": 1.8728401360243801e-07, + "loss": 0.6697, + "step": 4048 + }, + { + "epoch": 7.4981481481481485, + "grad_norm": 1.6453260599093877, + "learning_rate": 1.8727669533967882e-07, + "loss": 0.6184, + "step": 4049 + }, + { + "epoch": 7.5, + "grad_norm": 1.418240400125389, + "learning_rate": 1.8726937511470244e-07, + "loss": 0.633, + "step": 4050 + }, + { + "epoch": 7.5018518518518515, + "grad_norm": 1.4344293873011618, + "learning_rate": 1.8726205292767353e-07, + "loss": 0.623, + "step": 4051 + }, + { + "epoch": 7.503703703703704, + "grad_norm": 1.5055414276790458, + "learning_rate": 1.8725472877875665e-07, + "loss": 0.7047, + "step": 4052 + }, + { + "epoch": 7.5055555555555555, + "grad_norm": 1.6205361185627871, + "learning_rate": 1.8724740266811652e-07, + "loss": 0.6511, + "step": 4053 + }, + { + "epoch": 7.507407407407407, + "grad_norm": 1.8670585084832534, + "learning_rate": 1.8724007459591778e-07, + "loss": 0.6621, + "step": 4054 + }, + { + "epoch": 7.5092592592592595, + "grad_norm": 1.3419629611741195, + "learning_rate": 1.8723274456232523e-07, + "loss": 0.6109, + "step": 4055 + }, + { + "epoch": 7.511111111111111, + "grad_norm": 2.782860920470479, + "learning_rate": 1.8722541256750368e-07, + "loss": 0.6479, + "step": 4056 + }, + { + "epoch": 7.512962962962963, + "grad_norm": 1.6492478442743528, + "learning_rate": 1.8721807861161795e-07, + "loss": 0.639, + "step": 4057 + }, + { + "epoch": 7.514814814814815, + "grad_norm": 1.6453918198706934, + "learning_rate": 1.8721074269483292e-07, + "loss": 0.6113, + "step": 4058 + }, + { + "epoch": 7.516666666666667, + "grad_norm": 2.0578754491473985, + "learning_rate": 1.8720340481731355e-07, + "loss": 0.6413, + "step": 4059 + }, + { + "epoch": 7.518518518518518, + "grad_norm": 1.596467408196277, + "learning_rate": 1.8719606497922475e-07, + "loss": 0.6481, + "step": 4060 + }, + { + "epoch": 7.520370370370371, + "grad_norm": 1.6790788253844622, + "learning_rate": 1.8718872318073166e-07, + "loss": 0.6417, + "step": 4061 + }, + { + "epoch": 7.522222222222222, + "grad_norm": 1.6472597441947012, + "learning_rate": 1.8718137942199924e-07, + "loss": 0.6134, + "step": 4062 + }, + { + "epoch": 7.524074074074074, + "grad_norm": 1.399303210350014, + "learning_rate": 1.871740337031926e-07, + "loss": 0.6688, + "step": 4063 + }, + { + "epoch": 7.525925925925926, + "grad_norm": 1.4801833157831892, + "learning_rate": 1.8716668602447697e-07, + "loss": 0.615, + "step": 4064 + }, + { + "epoch": 7.527777777777778, + "grad_norm": 1.5060934448923868, + "learning_rate": 1.8715933638601747e-07, + "loss": 0.6678, + "step": 4065 + }, + { + "epoch": 7.52962962962963, + "grad_norm": 1.4460234910383185, + "learning_rate": 1.8715198478797935e-07, + "loss": 0.6113, + "step": 4066 + }, + { + "epoch": 7.531481481481482, + "grad_norm": 1.956525870718335, + "learning_rate": 1.8714463123052794e-07, + "loss": 0.6404, + "step": 4067 + }, + { + "epoch": 7.533333333333333, + "grad_norm": 1.4309588994630034, + "learning_rate": 1.8713727571382853e-07, + "loss": 0.657, + "step": 4068 + }, + { + "epoch": 7.535185185185185, + "grad_norm": 2.568462404788745, + "learning_rate": 1.871299182380465e-07, + "loss": 0.6673, + "step": 4069 + }, + { + "epoch": 7.537037037037037, + "grad_norm": 1.4946588970444665, + "learning_rate": 1.871225588033473e-07, + "loss": 0.6267, + "step": 4070 + }, + { + "epoch": 7.538888888888889, + "grad_norm": 1.5236186631144908, + "learning_rate": 1.871151974098963e-07, + "loss": 0.6642, + "step": 4071 + }, + { + "epoch": 7.540740740740741, + "grad_norm": 1.6973899684221827, + "learning_rate": 1.871078340578591e-07, + "loss": 0.6438, + "step": 4072 + }, + { + "epoch": 7.542592592592593, + "grad_norm": 1.5599095910521932, + "learning_rate": 1.871004687474012e-07, + "loss": 0.6267, + "step": 4073 + }, + { + "epoch": 7.544444444444444, + "grad_norm": 1.5949436518509634, + "learning_rate": 1.8709310147868817e-07, + "loss": 0.6805, + "step": 4074 + }, + { + "epoch": 7.546296296296296, + "grad_norm": 1.5539873584251627, + "learning_rate": 1.870857322518857e-07, + "loss": 0.663, + "step": 4075 + }, + { + "epoch": 7.548148148148148, + "grad_norm": 1.4658836808621116, + "learning_rate": 1.8707836106715946e-07, + "loss": 0.6932, + "step": 4076 + }, + { + "epoch": 7.55, + "grad_norm": 1.7332260236788322, + "learning_rate": 1.8707098792467514e-07, + "loss": 0.6748, + "step": 4077 + }, + { + "epoch": 7.551851851851852, + "grad_norm": 1.9394567152486057, + "learning_rate": 1.8706361282459855e-07, + "loss": 0.7063, + "step": 4078 + }, + { + "epoch": 7.553703703703704, + "grad_norm": 1.574868185309456, + "learning_rate": 1.870562357670955e-07, + "loss": 0.6002, + "step": 4079 + }, + { + "epoch": 7.555555555555555, + "grad_norm": 1.5996707840386655, + "learning_rate": 1.870488567523318e-07, + "loss": 0.6647, + "step": 4080 + }, + { + "epoch": 7.557407407407408, + "grad_norm": 1.8674986538734952, + "learning_rate": 1.8704147578047338e-07, + "loss": 0.6809, + "step": 4081 + }, + { + "epoch": 7.559259259259259, + "grad_norm": 1.7562215366182878, + "learning_rate": 1.870340928516862e-07, + "loss": 0.664, + "step": 4082 + }, + { + "epoch": 7.561111111111111, + "grad_norm": 1.3540635157724146, + "learning_rate": 1.870267079661362e-07, + "loss": 0.6129, + "step": 4083 + }, + { + "epoch": 7.562962962962963, + "grad_norm": 1.3417011511580168, + "learning_rate": 1.870193211239895e-07, + "loss": 0.6198, + "step": 4084 + }, + { + "epoch": 7.564814814814815, + "grad_norm": 1.4992339490772237, + "learning_rate": 1.8701193232541206e-07, + "loss": 0.6421, + "step": 4085 + }, + { + "epoch": 7.566666666666666, + "grad_norm": 1.6067545683737503, + "learning_rate": 1.8700454157057008e-07, + "loss": 0.6315, + "step": 4086 + }, + { + "epoch": 7.568518518518519, + "grad_norm": 1.8488157713980993, + "learning_rate": 1.8699714885962973e-07, + "loss": 0.6782, + "step": 4087 + }, + { + "epoch": 7.57037037037037, + "grad_norm": 1.6190132458664848, + "learning_rate": 1.8698975419275716e-07, + "loss": 0.6534, + "step": 4088 + }, + { + "epoch": 7.572222222222222, + "grad_norm": 1.6983870556164997, + "learning_rate": 1.8698235757011863e-07, + "loss": 0.6422, + "step": 4089 + }, + { + "epoch": 7.574074074074074, + "grad_norm": 1.5614689595140214, + "learning_rate": 1.8697495899188053e-07, + "loss": 0.6724, + "step": 4090 + }, + { + "epoch": 7.575925925925926, + "grad_norm": 1.5441570201477295, + "learning_rate": 1.869675584582091e-07, + "loss": 0.6038, + "step": 4091 + }, + { + "epoch": 7.5777777777777775, + "grad_norm": 1.446960781686149, + "learning_rate": 1.8696015596927074e-07, + "loss": 0.6784, + "step": 4092 + }, + { + "epoch": 7.57962962962963, + "grad_norm": 1.7529008641910226, + "learning_rate": 1.869527515252319e-07, + "loss": 0.6176, + "step": 4093 + }, + { + "epoch": 7.5814814814814815, + "grad_norm": 1.6847089951326684, + "learning_rate": 1.8694534512625907e-07, + "loss": 0.6265, + "step": 4094 + }, + { + "epoch": 7.583333333333333, + "grad_norm": 1.4162552686373346, + "learning_rate": 1.8693793677251875e-07, + "loss": 0.6602, + "step": 4095 + }, + { + "epoch": 7.5851851851851855, + "grad_norm": 1.423651370019151, + "learning_rate": 1.8693052646417744e-07, + "loss": 0.6397, + "step": 4096 + }, + { + "epoch": 7.587037037037037, + "grad_norm": 1.5178443648516051, + "learning_rate": 1.8692311420140183e-07, + "loss": 0.654, + "step": 4097 + }, + { + "epoch": 7.588888888888889, + "grad_norm": 1.4734287879802168, + "learning_rate": 1.8691569998435853e-07, + "loss": 0.6505, + "step": 4098 + }, + { + "epoch": 7.590740740740741, + "grad_norm": 1.4322999673146037, + "learning_rate": 1.8690828381321425e-07, + "loss": 0.6345, + "step": 4099 + }, + { + "epoch": 7.592592592592593, + "grad_norm": 1.399537941239336, + "learning_rate": 1.8690086568813568e-07, + "loss": 0.6682, + "step": 4100 + }, + { + "epoch": 7.594444444444444, + "grad_norm": 1.4578955970828915, + "learning_rate": 1.8689344560928962e-07, + "loss": 0.634, + "step": 4101 + }, + { + "epoch": 7.5962962962962965, + "grad_norm": 1.3507504118187434, + "learning_rate": 1.8688602357684294e-07, + "loss": 0.6212, + "step": 4102 + }, + { + "epoch": 7.598148148148148, + "grad_norm": 1.514425950643582, + "learning_rate": 1.8687859959096246e-07, + "loss": 0.6117, + "step": 4103 + }, + { + "epoch": 7.6, + "grad_norm": 1.7487680201284967, + "learning_rate": 1.8687117365181512e-07, + "loss": 0.6439, + "step": 4104 + }, + { + "epoch": 7.601851851851852, + "grad_norm": 1.5926878763333923, + "learning_rate": 1.8686374575956782e-07, + "loss": 0.6159, + "step": 4105 + }, + { + "epoch": 7.603703703703704, + "grad_norm": 1.4476350343337396, + "learning_rate": 1.8685631591438762e-07, + "loss": 0.624, + "step": 4106 + }, + { + "epoch": 7.605555555555555, + "grad_norm": 1.4128404717725838, + "learning_rate": 1.868488841164415e-07, + "loss": 0.6421, + "step": 4107 + }, + { + "epoch": 7.607407407407408, + "grad_norm": 1.6589220959329893, + "learning_rate": 1.8684145036589665e-07, + "loss": 0.5936, + "step": 4108 + }, + { + "epoch": 7.609259259259259, + "grad_norm": 1.6706293647527808, + "learning_rate": 1.868340146629201e-07, + "loss": 0.6477, + "step": 4109 + }, + { + "epoch": 7.611111111111111, + "grad_norm": 1.3567694028042443, + "learning_rate": 1.8682657700767908e-07, + "loss": 0.7028, + "step": 4110 + }, + { + "epoch": 7.612962962962963, + "grad_norm": 1.575009583686362, + "learning_rate": 1.8681913740034077e-07, + "loss": 0.681, + "step": 4111 + }, + { + "epoch": 7.614814814814815, + "grad_norm": 1.6028449486300527, + "learning_rate": 1.8681169584107248e-07, + "loss": 0.6813, + "step": 4112 + }, + { + "epoch": 7.616666666666667, + "grad_norm": 1.6138012920798457, + "learning_rate": 1.868042523300415e-07, + "loss": 0.6642, + "step": 4113 + }, + { + "epoch": 7.618518518518519, + "grad_norm": 1.6417737668598553, + "learning_rate": 1.8679680686741511e-07, + "loss": 0.6624, + "step": 4114 + }, + { + "epoch": 7.62037037037037, + "grad_norm": 1.944018814569686, + "learning_rate": 1.867893594533608e-07, + "loss": 0.6947, + "step": 4115 + }, + { + "epoch": 7.622222222222222, + "grad_norm": 1.5583566702500589, + "learning_rate": 1.86781910088046e-07, + "loss": 0.6575, + "step": 4116 + }, + { + "epoch": 7.624074074074074, + "grad_norm": 1.6248047442394085, + "learning_rate": 1.8677445877163814e-07, + "loss": 0.6404, + "step": 4117 + }, + { + "epoch": 7.625925925925926, + "grad_norm": 1.6359576339180701, + "learning_rate": 1.8676700550430477e-07, + "loss": 0.6745, + "step": 4118 + }, + { + "epoch": 7.627777777777778, + "grad_norm": 1.3942692090006625, + "learning_rate": 1.8675955028621346e-07, + "loss": 0.6559, + "step": 4119 + }, + { + "epoch": 7.62962962962963, + "grad_norm": 1.8661747658574874, + "learning_rate": 1.8675209311753183e-07, + "loss": 0.6365, + "step": 4120 + }, + { + "epoch": 7.631481481481481, + "grad_norm": 1.7458349511732807, + "learning_rate": 1.8674463399842755e-07, + "loss": 0.63, + "step": 4121 + }, + { + "epoch": 7.633333333333333, + "grad_norm": 1.4325536750833956, + "learning_rate": 1.8673717292906828e-07, + "loss": 0.6521, + "step": 4122 + }, + { + "epoch": 7.635185185185185, + "grad_norm": 1.5807239475664143, + "learning_rate": 1.8672970990962178e-07, + "loss": 0.6401, + "step": 4123 + }, + { + "epoch": 7.637037037037037, + "grad_norm": 1.8146623204222527, + "learning_rate": 1.8672224494025587e-07, + "loss": 0.6214, + "step": 4124 + }, + { + "epoch": 7.638888888888889, + "grad_norm": 1.5746341321799902, + "learning_rate": 1.8671477802113832e-07, + "loss": 0.6383, + "step": 4125 + }, + { + "epoch": 7.640740740740741, + "grad_norm": 1.4294142776569376, + "learning_rate": 1.867073091524371e-07, + "loss": 0.6819, + "step": 4126 + }, + { + "epoch": 7.642592592592592, + "grad_norm": 1.5730825867120848, + "learning_rate": 1.8669983833432003e-07, + "loss": 0.6776, + "step": 4127 + }, + { + "epoch": 7.644444444444445, + "grad_norm": 1.6503488283002208, + "learning_rate": 1.8669236556695514e-07, + "loss": 0.6662, + "step": 4128 + }, + { + "epoch": 7.646296296296296, + "grad_norm": 1.6033761589685698, + "learning_rate": 1.8668489085051042e-07, + "loss": 0.6274, + "step": 4129 + }, + { + "epoch": 7.648148148148148, + "grad_norm": 1.5472723799331083, + "learning_rate": 1.8667741418515393e-07, + "loss": 0.6397, + "step": 4130 + }, + { + "epoch": 7.65, + "grad_norm": 1.5269383901455245, + "learning_rate": 1.8666993557105375e-07, + "loss": 0.634, + "step": 4131 + }, + { + "epoch": 7.651851851851852, + "grad_norm": 1.7745620664794373, + "learning_rate": 1.8666245500837802e-07, + "loss": 0.704, + "step": 4132 + }, + { + "epoch": 7.6537037037037035, + "grad_norm": 1.463490393733774, + "learning_rate": 1.8665497249729493e-07, + "loss": 0.6135, + "step": 4133 + }, + { + "epoch": 7.655555555555556, + "grad_norm": 1.4476115112626073, + "learning_rate": 1.866474880379727e-07, + "loss": 0.6444, + "step": 4134 + }, + { + "epoch": 7.657407407407407, + "grad_norm": 1.5550557659600037, + "learning_rate": 1.8664000163057965e-07, + "loss": 0.6334, + "step": 4135 + }, + { + "epoch": 7.659259259259259, + "grad_norm": 1.4533100743292084, + "learning_rate": 1.86632513275284e-07, + "loss": 0.6727, + "step": 4136 + }, + { + "epoch": 7.661111111111111, + "grad_norm": 1.336483695629435, + "learning_rate": 1.866250229722542e-07, + "loss": 0.6222, + "step": 4137 + }, + { + "epoch": 7.662962962962963, + "grad_norm": 1.9822843581928953, + "learning_rate": 1.866175307216586e-07, + "loss": 0.6008, + "step": 4138 + }, + { + "epoch": 7.6648148148148145, + "grad_norm": 1.5860642113070924, + "learning_rate": 1.8661003652366565e-07, + "loss": 0.6395, + "step": 4139 + }, + { + "epoch": 7.666666666666667, + "grad_norm": 1.5766500516971789, + "learning_rate": 1.8660254037844388e-07, + "loss": 0.6296, + "step": 4140 + }, + { + "epoch": 7.6685185185185185, + "grad_norm": 1.3911359736764293, + "learning_rate": 1.8659504228616175e-07, + "loss": 0.5978, + "step": 4141 + }, + { + "epoch": 7.67037037037037, + "grad_norm": 1.6035764175993417, + "learning_rate": 1.8658754224698788e-07, + "loss": 0.6507, + "step": 4142 + }, + { + "epoch": 7.6722222222222225, + "grad_norm": 1.4717955274007033, + "learning_rate": 1.8658004026109092e-07, + "loss": 0.6542, + "step": 4143 + }, + { + "epoch": 7.674074074074074, + "grad_norm": 1.469327684153371, + "learning_rate": 1.865725363286395e-07, + "loss": 0.644, + "step": 4144 + }, + { + "epoch": 7.675925925925926, + "grad_norm": 1.5363057272052651, + "learning_rate": 1.8656503044980235e-07, + "loss": 0.6804, + "step": 4145 + }, + { + "epoch": 7.677777777777778, + "grad_norm": 1.4531300272551757, + "learning_rate": 1.865575226247482e-07, + "loss": 0.6775, + "step": 4146 + }, + { + "epoch": 7.67962962962963, + "grad_norm": 1.5383319069582633, + "learning_rate": 1.8655001285364584e-07, + "loss": 0.6039, + "step": 4147 + }, + { + "epoch": 7.681481481481481, + "grad_norm": 1.7909824422740164, + "learning_rate": 1.8654250113666415e-07, + "loss": 0.6171, + "step": 4148 + }, + { + "epoch": 7.683333333333334, + "grad_norm": 1.3492724670530756, + "learning_rate": 1.8653498747397195e-07, + "loss": 0.6335, + "step": 4149 + }, + { + "epoch": 7.685185185185185, + "grad_norm": 1.553135203701596, + "learning_rate": 1.8652747186573822e-07, + "loss": 0.6299, + "step": 4150 + }, + { + "epoch": 7.687037037037037, + "grad_norm": 1.4691030321047163, + "learning_rate": 1.8651995431213196e-07, + "loss": 0.6378, + "step": 4151 + }, + { + "epoch": 7.688888888888889, + "grad_norm": 1.8708028071404206, + "learning_rate": 1.865124348133221e-07, + "loss": 0.6234, + "step": 4152 + }, + { + "epoch": 7.690740740740741, + "grad_norm": 1.3929958372307631, + "learning_rate": 1.8650491336947775e-07, + "loss": 0.6259, + "step": 4153 + }, + { + "epoch": 7.692592592592592, + "grad_norm": 1.420316610570055, + "learning_rate": 1.8649738998076798e-07, + "loss": 0.6443, + "step": 4154 + }, + { + "epoch": 7.694444444444445, + "grad_norm": 1.749485863276849, + "learning_rate": 1.86489864647362e-07, + "loss": 0.6598, + "step": 4155 + }, + { + "epoch": 7.696296296296296, + "grad_norm": 1.347132867413065, + "learning_rate": 1.8648233736942894e-07, + "loss": 0.6407, + "step": 4156 + }, + { + "epoch": 7.698148148148148, + "grad_norm": 1.9319174794968108, + "learning_rate": 1.8647480814713804e-07, + "loss": 0.6305, + "step": 4157 + }, + { + "epoch": 7.7, + "grad_norm": 1.588122532870714, + "learning_rate": 1.8646727698065863e-07, + "loss": 0.6831, + "step": 4158 + }, + { + "epoch": 7.701851851851852, + "grad_norm": 1.5731503694414677, + "learning_rate": 1.8645974387015998e-07, + "loss": 0.595, + "step": 4159 + }, + { + "epoch": 7.703703703703704, + "grad_norm": 1.703346002954575, + "learning_rate": 1.8645220881581144e-07, + "loss": 0.6737, + "step": 4160 + }, + { + "epoch": 7.705555555555556, + "grad_norm": 1.549629897948105, + "learning_rate": 1.8644467181778246e-07, + "loss": 0.6653, + "step": 4161 + }, + { + "epoch": 7.707407407407407, + "grad_norm": 1.5190009048795332, + "learning_rate": 1.8643713287624247e-07, + "loss": 0.6204, + "step": 4162 + }, + { + "epoch": 7.709259259259259, + "grad_norm": 2.102605853425928, + "learning_rate": 1.86429591991361e-07, + "loss": 0.6703, + "step": 4163 + }, + { + "epoch": 7.711111111111111, + "grad_norm": 2.224338451369923, + "learning_rate": 1.8642204916330755e-07, + "loss": 0.6727, + "step": 4164 + }, + { + "epoch": 7.712962962962963, + "grad_norm": 1.446373386823586, + "learning_rate": 1.8641450439225169e-07, + "loss": 0.6802, + "step": 4165 + }, + { + "epoch": 7.714814814814815, + "grad_norm": 1.437228662435052, + "learning_rate": 1.864069576783631e-07, + "loss": 0.6529, + "step": 4166 + }, + { + "epoch": 7.716666666666667, + "grad_norm": 1.6400586302159657, + "learning_rate": 1.8639940902181142e-07, + "loss": 0.6227, + "step": 4167 + }, + { + "epoch": 7.718518518518518, + "grad_norm": 1.4350685004590364, + "learning_rate": 1.8639185842276635e-07, + "loss": 0.66, + "step": 4168 + }, + { + "epoch": 7.72037037037037, + "grad_norm": 1.4091877959516466, + "learning_rate": 1.863843058813977e-07, + "loss": 0.6844, + "step": 4169 + }, + { + "epoch": 7.722222222222222, + "grad_norm": 1.362957417411261, + "learning_rate": 1.863767513978752e-07, + "loss": 0.6774, + "step": 4170 + }, + { + "epoch": 7.724074074074074, + "grad_norm": 1.9436180705373733, + "learning_rate": 1.8636919497236876e-07, + "loss": 0.6374, + "step": 4171 + }, + { + "epoch": 7.725925925925926, + "grad_norm": 1.530064974317294, + "learning_rate": 1.8636163660504822e-07, + "loss": 0.6205, + "step": 4172 + }, + { + "epoch": 7.727777777777778, + "grad_norm": 3.3478188353432223, + "learning_rate": 1.8635407629608358e-07, + "loss": 0.6163, + "step": 4173 + }, + { + "epoch": 7.729629629629629, + "grad_norm": 1.537655692465855, + "learning_rate": 1.8634651404564473e-07, + "loss": 0.6487, + "step": 4174 + }, + { + "epoch": 7.731481481481482, + "grad_norm": 1.4478956761419441, + "learning_rate": 1.8633894985390173e-07, + "loss": 0.6518, + "step": 4175 + }, + { + "epoch": 7.733333333333333, + "grad_norm": 1.5665318667251331, + "learning_rate": 1.8633138372102466e-07, + "loss": 0.6714, + "step": 4176 + }, + { + "epoch": 7.735185185185185, + "grad_norm": 1.561653964888794, + "learning_rate": 1.863238156471836e-07, + "loss": 0.648, + "step": 4177 + }, + { + "epoch": 7.737037037037037, + "grad_norm": 1.6655212599700395, + "learning_rate": 1.863162456325487e-07, + "loss": 0.6267, + "step": 4178 + }, + { + "epoch": 7.738888888888889, + "grad_norm": 1.8084827258756873, + "learning_rate": 1.863086736772902e-07, + "loss": 0.6411, + "step": 4179 + }, + { + "epoch": 7.7407407407407405, + "grad_norm": 1.59021261787782, + "learning_rate": 1.8630109978157827e-07, + "loss": 0.6472, + "step": 4180 + }, + { + "epoch": 7.742592592592593, + "grad_norm": 1.7005023157939847, + "learning_rate": 1.8629352394558328e-07, + "loss": 0.6206, + "step": 4181 + }, + { + "epoch": 7.7444444444444445, + "grad_norm": 1.5556903789202785, + "learning_rate": 1.8628594616947546e-07, + "loss": 0.6476, + "step": 4182 + }, + { + "epoch": 7.746296296296296, + "grad_norm": 1.6136594489433103, + "learning_rate": 1.8627836645342523e-07, + "loss": 0.6869, + "step": 4183 + }, + { + "epoch": 7.7481481481481485, + "grad_norm": 1.8900510714546466, + "learning_rate": 1.86270784797603e-07, + "loss": 0.6542, + "step": 4184 + }, + { + "epoch": 7.75, + "grad_norm": 1.593568532823711, + "learning_rate": 1.862632012021792e-07, + "loss": 0.6513, + "step": 4185 + }, + { + "epoch": 7.7518518518518515, + "grad_norm": 1.4748305020417718, + "learning_rate": 1.8625561566732438e-07, + "loss": 0.6759, + "step": 4186 + }, + { + "epoch": 7.753703703703704, + "grad_norm": 1.6213285537756228, + "learning_rate": 1.8624802819320902e-07, + "loss": 0.672, + "step": 4187 + }, + { + "epoch": 7.7555555555555555, + "grad_norm": 1.403214412600631, + "learning_rate": 1.8624043878000374e-07, + "loss": 0.6347, + "step": 4188 + }, + { + "epoch": 7.757407407407407, + "grad_norm": 1.7317484034809942, + "learning_rate": 1.862328474278792e-07, + "loss": 0.6617, + "step": 4189 + }, + { + "epoch": 7.7592592592592595, + "grad_norm": 1.4312230393163905, + "learning_rate": 1.8622525413700603e-07, + "loss": 0.639, + "step": 4190 + }, + { + "epoch": 7.761111111111111, + "grad_norm": 1.5467294525945965, + "learning_rate": 1.8621765890755497e-07, + "loss": 0.6641, + "step": 4191 + }, + { + "epoch": 7.762962962962963, + "grad_norm": 1.7597578980312507, + "learning_rate": 1.8621006173969675e-07, + "loss": 0.6333, + "step": 4192 + }, + { + "epoch": 7.764814814814815, + "grad_norm": 1.5385857974725132, + "learning_rate": 1.8620246263360224e-07, + "loss": 0.6303, + "step": 4193 + }, + { + "epoch": 7.766666666666667, + "grad_norm": 1.5603199575634885, + "learning_rate": 1.861948615894422e-07, + "loss": 0.6104, + "step": 4194 + }, + { + "epoch": 7.768518518518518, + "grad_norm": 1.7077788971930847, + "learning_rate": 1.8618725860738761e-07, + "loss": 0.616, + "step": 4195 + }, + { + "epoch": 7.770370370370371, + "grad_norm": 1.4075230011896642, + "learning_rate": 1.8617965368760936e-07, + "loss": 0.6915, + "step": 4196 + }, + { + "epoch": 7.772222222222222, + "grad_norm": 1.429929144872628, + "learning_rate": 1.861720468302784e-07, + "loss": 0.633, + "step": 4197 + }, + { + "epoch": 7.774074074074074, + "grad_norm": 1.6918948829595555, + "learning_rate": 1.8616443803556583e-07, + "loss": 0.6243, + "step": 4198 + }, + { + "epoch": 7.775925925925926, + "grad_norm": 1.4922432742782052, + "learning_rate": 1.8615682730364265e-07, + "loss": 0.6492, + "step": 4199 + }, + { + "epoch": 7.777777777777778, + "grad_norm": 1.4295920482476536, + "learning_rate": 1.8614921463468e-07, + "loss": 0.6366, + "step": 4200 + }, + { + "epoch": 7.77962962962963, + "grad_norm": 1.5561707759060341, + "learning_rate": 1.8614160002884902e-07, + "loss": 0.6569, + "step": 4201 + }, + { + "epoch": 7.781481481481482, + "grad_norm": 1.5595983391384, + "learning_rate": 1.8613398348632093e-07, + "loss": 0.6683, + "step": 4202 + }, + { + "epoch": 7.783333333333333, + "grad_norm": 1.453895198698467, + "learning_rate": 1.8612636500726694e-07, + "loss": 0.6776, + "step": 4203 + }, + { + "epoch": 7.785185185185185, + "grad_norm": 1.4813169379502018, + "learning_rate": 1.8611874459185837e-07, + "loss": 0.6284, + "step": 4204 + }, + { + "epoch": 7.787037037037037, + "grad_norm": 2.7805780326579446, + "learning_rate": 1.8611112224026652e-07, + "loss": 0.6448, + "step": 4205 + }, + { + "epoch": 7.788888888888889, + "grad_norm": 1.4077861802135274, + "learning_rate": 1.8610349795266273e-07, + "loss": 0.6505, + "step": 4206 + }, + { + "epoch": 7.790740740740741, + "grad_norm": 1.5090737509015753, + "learning_rate": 1.8609587172921854e-07, + "loss": 0.6621, + "step": 4207 + }, + { + "epoch": 7.792592592592593, + "grad_norm": 1.5741265115904144, + "learning_rate": 1.860882435701052e-07, + "loss": 0.634, + "step": 4208 + }, + { + "epoch": 7.794444444444444, + "grad_norm": 1.5084332309374784, + "learning_rate": 1.8608061347549442e-07, + "loss": 0.6274, + "step": 4209 + }, + { + "epoch": 7.796296296296296, + "grad_norm": 1.622029858928403, + "learning_rate": 1.8607298144555763e-07, + "loss": 0.5935, + "step": 4210 + }, + { + "epoch": 7.798148148148148, + "grad_norm": 1.5772689872271521, + "learning_rate": 1.8606534748046645e-07, + "loss": 0.6524, + "step": 4211 + }, + { + "epoch": 7.8, + "grad_norm": 3.2961944151240847, + "learning_rate": 1.8605771158039253e-07, + "loss": 0.6624, + "step": 4212 + }, + { + "epoch": 7.801851851851852, + "grad_norm": 1.9150620798645273, + "learning_rate": 1.8605007374550747e-07, + "loss": 0.6264, + "step": 4213 + }, + { + "epoch": 7.803703703703704, + "grad_norm": 1.8630508993229984, + "learning_rate": 1.860424339759831e-07, + "loss": 0.6749, + "step": 4214 + }, + { + "epoch": 7.805555555555555, + "grad_norm": 1.5282530718257328, + "learning_rate": 1.860347922719911e-07, + "loss": 0.6011, + "step": 4215 + }, + { + "epoch": 7.807407407407408, + "grad_norm": 1.712442421556992, + "learning_rate": 1.860271486337033e-07, + "loss": 0.6377, + "step": 4216 + }, + { + "epoch": 7.809259259259259, + "grad_norm": 2.4504626785639467, + "learning_rate": 1.8601950306129156e-07, + "loss": 0.6016, + "step": 4217 + }, + { + "epoch": 7.811111111111111, + "grad_norm": 1.7471682745128487, + "learning_rate": 1.8601185555492775e-07, + "loss": 0.6012, + "step": 4218 + }, + { + "epoch": 7.812962962962963, + "grad_norm": 1.5842894188431227, + "learning_rate": 1.8600420611478388e-07, + "loss": 0.666, + "step": 4219 + }, + { + "epoch": 7.814814814814815, + "grad_norm": 1.6603301316866073, + "learning_rate": 1.8599655474103182e-07, + "loss": 0.6541, + "step": 4220 + }, + { + "epoch": 7.816666666666666, + "grad_norm": 1.8708674570862722, + "learning_rate": 1.8598890143384366e-07, + "loss": 0.6478, + "step": 4221 + }, + { + "epoch": 7.818518518518519, + "grad_norm": 1.368562142897557, + "learning_rate": 1.8598124619339143e-07, + "loss": 0.6239, + "step": 4222 + }, + { + "epoch": 7.82037037037037, + "grad_norm": 1.4363596719162366, + "learning_rate": 1.8597358901984727e-07, + "loss": 0.6646, + "step": 4223 + }, + { + "epoch": 7.822222222222222, + "grad_norm": 1.775958451072795, + "learning_rate": 1.8596592991338333e-07, + "loss": 0.6412, + "step": 4224 + }, + { + "epoch": 7.824074074074074, + "grad_norm": 1.7252118115311468, + "learning_rate": 1.8595826887417183e-07, + "loss": 0.6202, + "step": 4225 + }, + { + "epoch": 7.825925925925926, + "grad_norm": 1.6836867453631728, + "learning_rate": 1.8595060590238498e-07, + "loss": 0.6523, + "step": 4226 + }, + { + "epoch": 7.8277777777777775, + "grad_norm": 1.4806842244711003, + "learning_rate": 1.8594294099819506e-07, + "loss": 0.6228, + "step": 4227 + }, + { + "epoch": 7.82962962962963, + "grad_norm": 3.662140315928878, + "learning_rate": 1.8593527416177442e-07, + "loss": 0.6542, + "step": 4228 + }, + { + "epoch": 7.8314814814814815, + "grad_norm": 1.7823423846598296, + "learning_rate": 1.8592760539329542e-07, + "loss": 0.6491, + "step": 4229 + }, + { + "epoch": 7.833333333333333, + "grad_norm": 2.4833834280045037, + "learning_rate": 1.8591993469293046e-07, + "loss": 0.6531, + "step": 4230 + }, + { + "epoch": 7.8351851851851855, + "grad_norm": 1.5381319220752652, + "learning_rate": 1.85912262060852e-07, + "loss": 0.6186, + "step": 4231 + }, + { + "epoch": 7.837037037037037, + "grad_norm": 1.5617546543737226, + "learning_rate": 1.8590458749723262e-07, + "loss": 0.5944, + "step": 4232 + }, + { + "epoch": 7.838888888888889, + "grad_norm": 1.7299643375012799, + "learning_rate": 1.8589691100224475e-07, + "loss": 0.6389, + "step": 4233 + }, + { + "epoch": 7.840740740740741, + "grad_norm": 1.5588826698334022, + "learning_rate": 1.8588923257606104e-07, + "loss": 0.6073, + "step": 4234 + }, + { + "epoch": 7.842592592592593, + "grad_norm": 1.707479621324081, + "learning_rate": 1.8588155221885408e-07, + "loss": 0.6309, + "step": 4235 + }, + { + "epoch": 7.844444444444444, + "grad_norm": 1.4316085111407069, + "learning_rate": 1.858738699307966e-07, + "loss": 0.6845, + "step": 4236 + }, + { + "epoch": 7.8462962962962965, + "grad_norm": 1.458372035120529, + "learning_rate": 1.8586618571206133e-07, + "loss": 0.6422, + "step": 4237 + }, + { + "epoch": 7.848148148148148, + "grad_norm": 2.0739420093556915, + "learning_rate": 1.8585849956282095e-07, + "loss": 0.6129, + "step": 4238 + }, + { + "epoch": 7.85, + "grad_norm": 1.7042284260106428, + "learning_rate": 1.858508114832483e-07, + "loss": 0.6199, + "step": 4239 + }, + { + "epoch": 7.851851851851852, + "grad_norm": 1.7217152096944712, + "learning_rate": 1.858431214735163e-07, + "loss": 0.6698, + "step": 4240 + }, + { + "epoch": 7.853703703703704, + "grad_norm": 1.5269613222959182, + "learning_rate": 1.8583542953379772e-07, + "loss": 0.6076, + "step": 4241 + }, + { + "epoch": 7.855555555555555, + "grad_norm": 1.6024419616940813, + "learning_rate": 1.8582773566426558e-07, + "loss": 0.6239, + "step": 4242 + }, + { + "epoch": 7.857407407407408, + "grad_norm": 1.792100159409612, + "learning_rate": 1.8582003986509282e-07, + "loss": 0.6887, + "step": 4243 + }, + { + "epoch": 7.859259259259259, + "grad_norm": 1.5524411399251183, + "learning_rate": 1.8581234213645253e-07, + "loss": 0.6009, + "step": 4244 + }, + { + "epoch": 7.861111111111111, + "grad_norm": 1.7138923463648534, + "learning_rate": 1.858046424785177e-07, + "loss": 0.6131, + "step": 4245 + }, + { + "epoch": 7.862962962962963, + "grad_norm": 1.6579334318768093, + "learning_rate": 1.8579694089146144e-07, + "loss": 0.6829, + "step": 4246 + }, + { + "epoch": 7.864814814814815, + "grad_norm": 1.7584239444790144, + "learning_rate": 1.8578923737545695e-07, + "loss": 0.6217, + "step": 4247 + }, + { + "epoch": 7.866666666666667, + "grad_norm": 1.658315787124268, + "learning_rate": 1.8578153193067743e-07, + "loss": 0.6696, + "step": 4248 + }, + { + "epoch": 7.868518518518519, + "grad_norm": 1.7134472877783422, + "learning_rate": 1.8577382455729604e-07, + "loss": 0.6494, + "step": 4249 + }, + { + "epoch": 7.87037037037037, + "grad_norm": 1.624983905525824, + "learning_rate": 1.8576611525548614e-07, + "loss": 0.6346, + "step": 4250 + }, + { + "epoch": 7.872222222222222, + "grad_norm": 8.312117548713308, + "learning_rate": 1.8575840402542107e-07, + "loss": 0.6388, + "step": 4251 + }, + { + "epoch": 7.874074074074074, + "grad_norm": 1.5945072361428139, + "learning_rate": 1.8575069086727412e-07, + "loss": 0.613, + "step": 4252 + }, + { + "epoch": 7.875925925925926, + "grad_norm": 1.5040493394371637, + "learning_rate": 1.8574297578121874e-07, + "loss": 0.6301, + "step": 4253 + }, + { + "epoch": 7.877777777777778, + "grad_norm": 1.3848543515523097, + "learning_rate": 1.857352587674284e-07, + "loss": 0.6116, + "step": 4254 + }, + { + "epoch": 7.87962962962963, + "grad_norm": 3.4059485554226328, + "learning_rate": 1.8572753982607661e-07, + "loss": 0.621, + "step": 4255 + }, + { + "epoch": 7.881481481481481, + "grad_norm": 1.380778060377783, + "learning_rate": 1.8571981895733686e-07, + "loss": 0.6302, + "step": 4256 + }, + { + "epoch": 7.883333333333333, + "grad_norm": 2.5197301367682257, + "learning_rate": 1.8571209616138277e-07, + "loss": 0.6529, + "step": 4257 + }, + { + "epoch": 7.885185185185185, + "grad_norm": 1.394644740304966, + "learning_rate": 1.8570437143838797e-07, + "loss": 0.6744, + "step": 4258 + }, + { + "epoch": 7.887037037037037, + "grad_norm": 1.6460818344196817, + "learning_rate": 1.8569664478852616e-07, + "loss": 0.6105, + "step": 4259 + }, + { + "epoch": 7.888888888888889, + "grad_norm": 1.5927364343792982, + "learning_rate": 1.85688916211971e-07, + "loss": 0.6865, + "step": 4260 + }, + { + "epoch": 7.890740740740741, + "grad_norm": 2.0963265659207875, + "learning_rate": 1.856811857088963e-07, + "loss": 0.6509, + "step": 4261 + }, + { + "epoch": 7.892592592592592, + "grad_norm": 1.9794374700612483, + "learning_rate": 1.856734532794758e-07, + "loss": 0.6298, + "step": 4262 + }, + { + "epoch": 7.894444444444445, + "grad_norm": 1.5525811133556973, + "learning_rate": 1.856657189238834e-07, + "loss": 0.5925, + "step": 4263 + }, + { + "epoch": 7.896296296296296, + "grad_norm": 2.0709331003014464, + "learning_rate": 1.85657982642293e-07, + "loss": 0.6437, + "step": 4264 + }, + { + "epoch": 7.898148148148148, + "grad_norm": 1.6976029855130077, + "learning_rate": 1.8565024443487848e-07, + "loss": 0.6645, + "step": 4265 + }, + { + "epoch": 7.9, + "grad_norm": 1.5529385899318902, + "learning_rate": 1.8564250430181384e-07, + "loss": 0.6513, + "step": 4266 + }, + { + "epoch": 7.901851851851852, + "grad_norm": 1.4414187730073065, + "learning_rate": 1.856347622432731e-07, + "loss": 0.6116, + "step": 4267 + }, + { + "epoch": 7.9037037037037035, + "grad_norm": 1.7384404136145046, + "learning_rate": 1.8562701825943035e-07, + "loss": 0.6572, + "step": 4268 + }, + { + "epoch": 7.905555555555556, + "grad_norm": 1.5401282393535152, + "learning_rate": 1.8561927235045964e-07, + "loss": 0.6364, + "step": 4269 + }, + { + "epoch": 7.907407407407407, + "grad_norm": 1.8818522707456258, + "learning_rate": 1.8561152451653517e-07, + "loss": 0.6609, + "step": 4270 + }, + { + "epoch": 7.909259259259259, + "grad_norm": 2.0577859345367076, + "learning_rate": 1.856037747578311e-07, + "loss": 0.6427, + "step": 4271 + }, + { + "epoch": 7.911111111111111, + "grad_norm": 1.6308335595786876, + "learning_rate": 1.855960230745217e-07, + "loss": 0.6258, + "step": 4272 + }, + { + "epoch": 7.912962962962963, + "grad_norm": 1.5851849895190326, + "learning_rate": 1.855882694667812e-07, + "loss": 0.654, + "step": 4273 + }, + { + "epoch": 7.9148148148148145, + "grad_norm": 1.4676589101905246, + "learning_rate": 1.8558051393478395e-07, + "loss": 0.632, + "step": 4274 + }, + { + "epoch": 7.916666666666667, + "grad_norm": 1.4516158575923812, + "learning_rate": 1.8557275647870433e-07, + "loss": 0.6338, + "step": 4275 + }, + { + "epoch": 7.9185185185185185, + "grad_norm": 1.6025343622657053, + "learning_rate": 1.855649970987167e-07, + "loss": 0.6355, + "step": 4276 + }, + { + "epoch": 7.92037037037037, + "grad_norm": 1.7740244786284893, + "learning_rate": 1.855572357949956e-07, + "loss": 0.6474, + "step": 4277 + }, + { + "epoch": 7.9222222222222225, + "grad_norm": 1.6227287514844153, + "learning_rate": 1.855494725677154e-07, + "loss": 0.6409, + "step": 4278 + }, + { + "epoch": 7.924074074074074, + "grad_norm": 1.5908595654365214, + "learning_rate": 1.8554170741705076e-07, + "loss": 0.6613, + "step": 4279 + }, + { + "epoch": 7.925925925925926, + "grad_norm": 1.567252608698113, + "learning_rate": 1.8553394034317619e-07, + "loss": 0.6723, + "step": 4280 + }, + { + "epoch": 7.927777777777778, + "grad_norm": 1.453275447335963, + "learning_rate": 1.8552617134626633e-07, + "loss": 0.6268, + "step": 4281 + }, + { + "epoch": 7.92962962962963, + "grad_norm": 1.4753947705197545, + "learning_rate": 1.8551840042649586e-07, + "loss": 0.6487, + "step": 4282 + }, + { + "epoch": 7.931481481481481, + "grad_norm": 1.5814961888000902, + "learning_rate": 1.855106275840395e-07, + "loss": 0.648, + "step": 4283 + }, + { + "epoch": 7.933333333333334, + "grad_norm": 1.45632632111028, + "learning_rate": 1.8550285281907198e-07, + "loss": 0.6605, + "step": 4284 + }, + { + "epoch": 7.935185185185185, + "grad_norm": 1.6794029166643385, + "learning_rate": 1.854950761317681e-07, + "loss": 0.6601, + "step": 4285 + }, + { + "epoch": 7.937037037037037, + "grad_norm": 1.395438030656467, + "learning_rate": 1.854872975223027e-07, + "loss": 0.598, + "step": 4286 + }, + { + "epoch": 7.938888888888889, + "grad_norm": 1.238066686594357, + "learning_rate": 1.8547951699085068e-07, + "loss": 0.608, + "step": 4287 + }, + { + "epoch": 7.940740740740741, + "grad_norm": 1.6229941168801458, + "learning_rate": 1.8547173453758695e-07, + "loss": 0.6941, + "step": 4288 + }, + { + "epoch": 7.942592592592592, + "grad_norm": 1.5292300190384138, + "learning_rate": 1.854639501626865e-07, + "loss": 0.631, + "step": 4289 + }, + { + "epoch": 7.944444444444445, + "grad_norm": 1.4496658060766403, + "learning_rate": 1.8545616386632433e-07, + "loss": 0.6459, + "step": 4290 + }, + { + "epoch": 7.946296296296296, + "grad_norm": 1.6171495851246431, + "learning_rate": 1.854483756486755e-07, + "loss": 0.6509, + "step": 4291 + }, + { + "epoch": 7.948148148148148, + "grad_norm": 1.6253871821133663, + "learning_rate": 1.8544058550991513e-07, + "loss": 0.6456, + "step": 4292 + }, + { + "epoch": 7.95, + "grad_norm": 1.3464307100900252, + "learning_rate": 1.8543279345021832e-07, + "loss": 0.5869, + "step": 4293 + }, + { + "epoch": 7.951851851851852, + "grad_norm": 1.7335300007136392, + "learning_rate": 1.8542499946976027e-07, + "loss": 0.6586, + "step": 4294 + }, + { + "epoch": 7.953703703703704, + "grad_norm": 2.6870364919059506, + "learning_rate": 1.8541720356871626e-07, + "loss": 0.6322, + "step": 4295 + }, + { + "epoch": 7.955555555555556, + "grad_norm": 1.5670737690153551, + "learning_rate": 1.854094057472615e-07, + "loss": 0.6173, + "step": 4296 + }, + { + "epoch": 7.957407407407407, + "grad_norm": 1.3526343505502112, + "learning_rate": 1.8540160600557134e-07, + "loss": 0.6194, + "step": 4297 + }, + { + "epoch": 7.959259259259259, + "grad_norm": 1.448531816474373, + "learning_rate": 1.853938043438211e-07, + "loss": 0.664, + "step": 4298 + }, + { + "epoch": 7.961111111111111, + "grad_norm": 2.4246803155012606, + "learning_rate": 1.8538600076218626e-07, + "loss": 0.6319, + "step": 4299 + }, + { + "epoch": 7.962962962962963, + "grad_norm": 1.7017338196614473, + "learning_rate": 1.853781952608422e-07, + "loss": 0.6544, + "step": 4300 + }, + { + "epoch": 7.964814814814815, + "grad_norm": 1.4700864334407744, + "learning_rate": 1.853703878399644e-07, + "loss": 0.6532, + "step": 4301 + }, + { + "epoch": 7.966666666666667, + "grad_norm": 1.6545321919887692, + "learning_rate": 1.8536257849972844e-07, + "loss": 0.6462, + "step": 4302 + }, + { + "epoch": 7.968518518518518, + "grad_norm": 1.565082343945817, + "learning_rate": 1.8535476724030987e-07, + "loss": 0.6448, + "step": 4303 + }, + { + "epoch": 7.97037037037037, + "grad_norm": 1.6067404281562736, + "learning_rate": 1.8534695406188435e-07, + "loss": 0.6263, + "step": 4304 + }, + { + "epoch": 7.972222222222222, + "grad_norm": 1.4730438201168865, + "learning_rate": 1.8533913896462745e-07, + "loss": 0.6265, + "step": 4305 + }, + { + "epoch": 7.974074074074074, + "grad_norm": 1.5164473424735692, + "learning_rate": 1.8533132194871497e-07, + "loss": 0.6417, + "step": 4306 + }, + { + "epoch": 7.975925925925926, + "grad_norm": 1.7426257340424622, + "learning_rate": 1.8532350301432258e-07, + "loss": 0.6304, + "step": 4307 + }, + { + "epoch": 7.977777777777778, + "grad_norm": 1.9292112457906772, + "learning_rate": 1.8531568216162614e-07, + "loss": 0.6251, + "step": 4308 + }, + { + "epoch": 7.979629629629629, + "grad_norm": 1.8134625828924371, + "learning_rate": 1.853078593908014e-07, + "loss": 0.6264, + "step": 4309 + }, + { + "epoch": 7.981481481481482, + "grad_norm": 1.6079926089118437, + "learning_rate": 1.8530003470202434e-07, + "loss": 0.6583, + "step": 4310 + }, + { + "epoch": 7.983333333333333, + "grad_norm": 1.7299718911925903, + "learning_rate": 1.852922080954708e-07, + "loss": 0.6604, + "step": 4311 + }, + { + "epoch": 7.985185185185185, + "grad_norm": 1.7540557514425157, + "learning_rate": 1.852843795713168e-07, + "loss": 0.7104, + "step": 4312 + }, + { + "epoch": 7.987037037037037, + "grad_norm": 1.5018343152050084, + "learning_rate": 1.852765491297383e-07, + "loss": 0.6364, + "step": 4313 + }, + { + "epoch": 7.988888888888889, + "grad_norm": 1.5380650953221953, + "learning_rate": 1.8526871677091136e-07, + "loss": 0.6474, + "step": 4314 + }, + { + "epoch": 7.9907407407407405, + "grad_norm": 1.879917967560461, + "learning_rate": 1.8526088249501212e-07, + "loss": 0.6191, + "step": 4315 + }, + { + "epoch": 7.992592592592593, + "grad_norm": 1.6924644344447572, + "learning_rate": 1.852530463022166e-07, + "loss": 0.6649, + "step": 4316 + }, + { + "epoch": 7.9944444444444445, + "grad_norm": 1.5033886068092717, + "learning_rate": 1.8524520819270114e-07, + "loss": 0.6652, + "step": 4317 + }, + { + "epoch": 7.996296296296296, + "grad_norm": 1.3640533603143965, + "learning_rate": 1.8523736816664184e-07, + "loss": 0.6334, + "step": 4318 + }, + { + "epoch": 7.9981481481481485, + "grad_norm": 1.4770085813986176, + "learning_rate": 1.85229526224215e-07, + "loss": 0.6362, + "step": 4319 + }, + { + "epoch": 8.0, + "grad_norm": 1.6997209249669618, + "learning_rate": 1.8522168236559692e-07, + "loss": 0.6934, + "step": 4320 + }, + { + "epoch": 8.001851851851852, + "grad_norm": 1.6463305880589907, + "learning_rate": 1.85213836590964e-07, + "loss": 0.6503, + "step": 4321 + }, + { + "epoch": 8.003703703703703, + "grad_norm": 1.8935318491018889, + "learning_rate": 1.8520598890049254e-07, + "loss": 0.6494, + "step": 4322 + }, + { + "epoch": 8.005555555555556, + "grad_norm": 1.66853566427282, + "learning_rate": 1.8519813929435908e-07, + "loss": 0.6382, + "step": 4323 + }, + { + "epoch": 8.007407407407408, + "grad_norm": 1.7978058810590847, + "learning_rate": 1.8519028777274005e-07, + "loss": 0.6659, + "step": 4324 + }, + { + "epoch": 8.00925925925926, + "grad_norm": 1.320906984219989, + "learning_rate": 1.8518243433581194e-07, + "loss": 0.6796, + "step": 4325 + }, + { + "epoch": 8.011111111111111, + "grad_norm": 1.5965518375485208, + "learning_rate": 1.8517457898375136e-07, + "loss": 0.6353, + "step": 4326 + }, + { + "epoch": 8.012962962962963, + "grad_norm": 1.5946415135397438, + "learning_rate": 1.8516672171673495e-07, + "loss": 0.6317, + "step": 4327 + }, + { + "epoch": 8.014814814814814, + "grad_norm": 1.6117001839716567, + "learning_rate": 1.8515886253493932e-07, + "loss": 0.6004, + "step": 4328 + }, + { + "epoch": 8.016666666666667, + "grad_norm": 1.5769063563174675, + "learning_rate": 1.8515100143854116e-07, + "loss": 0.7085, + "step": 4329 + }, + { + "epoch": 8.018518518518519, + "grad_norm": 1.5131430070108751, + "learning_rate": 1.851431384277172e-07, + "loss": 0.6534, + "step": 4330 + }, + { + "epoch": 8.02037037037037, + "grad_norm": 1.3425115448660485, + "learning_rate": 1.851352735026443e-07, + "loss": 0.6744, + "step": 4331 + }, + { + "epoch": 8.022222222222222, + "grad_norm": 1.649345172673442, + "learning_rate": 1.8512740666349918e-07, + "loss": 0.6335, + "step": 4332 + }, + { + "epoch": 8.024074074074074, + "grad_norm": 1.768822983071256, + "learning_rate": 1.8511953791045877e-07, + "loss": 0.6469, + "step": 4333 + }, + { + "epoch": 8.025925925925925, + "grad_norm": 1.7247692819257618, + "learning_rate": 1.8511166724369996e-07, + "loss": 0.6824, + "step": 4334 + }, + { + "epoch": 8.027777777777779, + "grad_norm": 1.5123442165031116, + "learning_rate": 1.851037946633997e-07, + "loss": 0.6182, + "step": 4335 + }, + { + "epoch": 8.02962962962963, + "grad_norm": 1.4697323300137064, + "learning_rate": 1.8509592016973503e-07, + "loss": 0.7144, + "step": 4336 + }, + { + "epoch": 8.031481481481482, + "grad_norm": 1.8044024070931437, + "learning_rate": 1.8508804376288297e-07, + "loss": 0.6558, + "step": 4337 + }, + { + "epoch": 8.033333333333333, + "grad_norm": 1.665857768710196, + "learning_rate": 1.8508016544302055e-07, + "loss": 0.6221, + "step": 4338 + }, + { + "epoch": 8.035185185185185, + "grad_norm": 1.5675589029891257, + "learning_rate": 1.8507228521032493e-07, + "loss": 0.6831, + "step": 4339 + }, + { + "epoch": 8.037037037037036, + "grad_norm": 1.8239402150067279, + "learning_rate": 1.8506440306497333e-07, + "loss": 0.646, + "step": 4340 + }, + { + "epoch": 8.03888888888889, + "grad_norm": 3.501799224046458, + "learning_rate": 1.8505651900714289e-07, + "loss": 0.7225, + "step": 4341 + }, + { + "epoch": 8.040740740740741, + "grad_norm": 1.5963086775963384, + "learning_rate": 1.850486330370109e-07, + "loss": 0.6809, + "step": 4342 + }, + { + "epoch": 8.042592592592593, + "grad_norm": 1.9526566130968028, + "learning_rate": 1.8504074515475464e-07, + "loss": 0.6606, + "step": 4343 + }, + { + "epoch": 8.044444444444444, + "grad_norm": 1.6290485135049857, + "learning_rate": 1.8503285536055145e-07, + "loss": 0.6437, + "step": 4344 + }, + { + "epoch": 8.046296296296296, + "grad_norm": 1.657413998377113, + "learning_rate": 1.8502496365457873e-07, + "loss": 0.6435, + "step": 4345 + }, + { + "epoch": 8.048148148148147, + "grad_norm": 1.465077365534314, + "learning_rate": 1.8501707003701395e-07, + "loss": 0.6362, + "step": 4346 + }, + { + "epoch": 8.05, + "grad_norm": 2.0106985575683387, + "learning_rate": 1.850091745080345e-07, + "loss": 0.6334, + "step": 4347 + }, + { + "epoch": 8.051851851851852, + "grad_norm": 1.5505533501509736, + "learning_rate": 1.850012770678179e-07, + "loss": 0.6722, + "step": 4348 + }, + { + "epoch": 8.053703703703704, + "grad_norm": 1.6036963048630817, + "learning_rate": 1.8499337771654176e-07, + "loss": 0.6396, + "step": 4349 + }, + { + "epoch": 8.055555555555555, + "grad_norm": 1.6011496495248434, + "learning_rate": 1.8498547645438364e-07, + "loss": 0.6683, + "step": 4350 + }, + { + "epoch": 8.057407407407407, + "grad_norm": 1.3797810921372278, + "learning_rate": 1.849775732815212e-07, + "loss": 0.6333, + "step": 4351 + }, + { + "epoch": 8.059259259259258, + "grad_norm": 1.5533664317223292, + "learning_rate": 1.8496966819813213e-07, + "loss": 0.6564, + "step": 4352 + }, + { + "epoch": 8.061111111111112, + "grad_norm": 1.9275002885753574, + "learning_rate": 1.8496176120439412e-07, + "loss": 0.6313, + "step": 4353 + }, + { + "epoch": 8.062962962962963, + "grad_norm": 1.6328122279662558, + "learning_rate": 1.8495385230048499e-07, + "loss": 0.6389, + "step": 4354 + }, + { + "epoch": 8.064814814814815, + "grad_norm": 1.4991029471086874, + "learning_rate": 1.8494594148658252e-07, + "loss": 0.6653, + "step": 4355 + }, + { + "epoch": 8.066666666666666, + "grad_norm": 1.534128662119345, + "learning_rate": 1.8493802876286456e-07, + "loss": 0.6774, + "step": 4356 + }, + { + "epoch": 8.068518518518518, + "grad_norm": 1.9999009197071358, + "learning_rate": 1.8493011412950903e-07, + "loss": 0.6065, + "step": 4357 + }, + { + "epoch": 8.07037037037037, + "grad_norm": 1.5220124854008472, + "learning_rate": 1.8492219758669387e-07, + "loss": 0.6619, + "step": 4358 + }, + { + "epoch": 8.072222222222223, + "grad_norm": 1.9095483502041495, + "learning_rate": 1.8491427913459707e-07, + "loss": 0.6492, + "step": 4359 + }, + { + "epoch": 8.074074074074074, + "grad_norm": 1.3001501103484336, + "learning_rate": 1.8490635877339664e-07, + "loss": 0.6673, + "step": 4360 + }, + { + "epoch": 8.075925925925926, + "grad_norm": 1.4017387089671602, + "learning_rate": 1.8489843650327067e-07, + "loss": 0.6746, + "step": 4361 + }, + { + "epoch": 8.077777777777778, + "grad_norm": 1.6264662484582797, + "learning_rate": 1.848905123243973e-07, + "loss": 0.6783, + "step": 4362 + }, + { + "epoch": 8.079629629629629, + "grad_norm": 1.4885869832802863, + "learning_rate": 1.8488258623695462e-07, + "loss": 0.6114, + "step": 4363 + }, + { + "epoch": 8.081481481481482, + "grad_norm": 1.6734729503633872, + "learning_rate": 1.8487465824112084e-07, + "loss": 0.6397, + "step": 4364 + }, + { + "epoch": 8.083333333333334, + "grad_norm": 1.3722899036518412, + "learning_rate": 1.8486672833707425e-07, + "loss": 0.6543, + "step": 4365 + }, + { + "epoch": 8.085185185185185, + "grad_norm": 1.4160090447514315, + "learning_rate": 1.8485879652499311e-07, + "loss": 0.6575, + "step": 4366 + }, + { + "epoch": 8.087037037037037, + "grad_norm": 1.638746924366951, + "learning_rate": 1.8485086280505576e-07, + "loss": 0.6325, + "step": 4367 + }, + { + "epoch": 8.088888888888889, + "grad_norm": 2.07357267133277, + "learning_rate": 1.848429271774406e-07, + "loss": 0.6514, + "step": 4368 + }, + { + "epoch": 8.09074074074074, + "grad_norm": 1.388452779040256, + "learning_rate": 1.8483498964232593e-07, + "loss": 0.6956, + "step": 4369 + }, + { + "epoch": 8.092592592592593, + "grad_norm": 2.2146826901188694, + "learning_rate": 1.8482705019989034e-07, + "loss": 0.6657, + "step": 4370 + }, + { + "epoch": 8.094444444444445, + "grad_norm": 1.3897073864275475, + "learning_rate": 1.8481910885031224e-07, + "loss": 0.6304, + "step": 4371 + }, + { + "epoch": 8.096296296296297, + "grad_norm": 1.8588323917875276, + "learning_rate": 1.8481116559377024e-07, + "loss": 0.6121, + "step": 4372 + }, + { + "epoch": 8.098148148148148, + "grad_norm": 1.5909449451188151, + "learning_rate": 1.8480322043044287e-07, + "loss": 0.6391, + "step": 4373 + }, + { + "epoch": 8.1, + "grad_norm": 1.6987182691635367, + "learning_rate": 1.8479527336050878e-07, + "loss": 0.6344, + "step": 4374 + }, + { + "epoch": 8.101851851851851, + "grad_norm": 1.5486197353695765, + "learning_rate": 1.8478732438414666e-07, + "loss": 0.67, + "step": 4375 + }, + { + "epoch": 8.103703703703705, + "grad_norm": 1.3226929125277187, + "learning_rate": 1.847793735015352e-07, + "loss": 0.6706, + "step": 4376 + }, + { + "epoch": 8.105555555555556, + "grad_norm": 1.3837574224367704, + "learning_rate": 1.8477142071285316e-07, + "loss": 0.6257, + "step": 4377 + }, + { + "epoch": 8.107407407407408, + "grad_norm": 1.5315962693672416, + "learning_rate": 1.8476346601827936e-07, + "loss": 0.6336, + "step": 4378 + }, + { + "epoch": 8.10925925925926, + "grad_norm": 1.4767475905610052, + "learning_rate": 1.8475550941799264e-07, + "loss": 0.652, + "step": 4379 + }, + { + "epoch": 8.11111111111111, + "grad_norm": 1.5308153588879692, + "learning_rate": 1.8474755091217184e-07, + "loss": 0.6577, + "step": 4380 + }, + { + "epoch": 8.112962962962962, + "grad_norm": 1.2693184323832511, + "learning_rate": 1.8473959050099595e-07, + "loss": 0.6165, + "step": 4381 + }, + { + "epoch": 8.114814814814816, + "grad_norm": 1.3508101952097766, + "learning_rate": 1.847316281846439e-07, + "loss": 0.6329, + "step": 4382 + }, + { + "epoch": 8.116666666666667, + "grad_norm": 1.5248127101440063, + "learning_rate": 1.8472366396329476e-07, + "loss": 0.6677, + "step": 4383 + }, + { + "epoch": 8.118518518518519, + "grad_norm": 1.4675346734606156, + "learning_rate": 1.8471569783712752e-07, + "loss": 0.6897, + "step": 4384 + }, + { + "epoch": 8.12037037037037, + "grad_norm": 1.452606606369142, + "learning_rate": 1.847077298063213e-07, + "loss": 0.6795, + "step": 4385 + }, + { + "epoch": 8.122222222222222, + "grad_norm": 1.3186620871342463, + "learning_rate": 1.8469975987105526e-07, + "loss": 0.6273, + "step": 4386 + }, + { + "epoch": 8.124074074074073, + "grad_norm": 1.4605072670996249, + "learning_rate": 1.846917880315086e-07, + "loss": 0.6737, + "step": 4387 + }, + { + "epoch": 8.125925925925927, + "grad_norm": 1.60525991633902, + "learning_rate": 1.846838142878605e-07, + "loss": 0.5899, + "step": 4388 + }, + { + "epoch": 8.127777777777778, + "grad_norm": 1.5295405089725784, + "learning_rate": 1.8467583864029023e-07, + "loss": 0.6407, + "step": 4389 + }, + { + "epoch": 8.12962962962963, + "grad_norm": 1.8837356310494495, + "learning_rate": 1.8466786108897718e-07, + "loss": 0.6047, + "step": 4390 + }, + { + "epoch": 8.131481481481481, + "grad_norm": 1.7326375202580129, + "learning_rate": 1.8465988163410065e-07, + "loss": 0.6584, + "step": 4391 + }, + { + "epoch": 8.133333333333333, + "grad_norm": 1.9541442422159079, + "learning_rate": 1.8465190027584002e-07, + "loss": 0.6528, + "step": 4392 + }, + { + "epoch": 8.135185185185184, + "grad_norm": 1.634809818560585, + "learning_rate": 1.846439170143748e-07, + "loss": 0.6796, + "step": 4393 + }, + { + "epoch": 8.137037037037038, + "grad_norm": 1.4625602427882987, + "learning_rate": 1.846359318498844e-07, + "loss": 0.658, + "step": 4394 + }, + { + "epoch": 8.13888888888889, + "grad_norm": 1.293210269158073, + "learning_rate": 1.846279447825484e-07, + "loss": 0.6044, + "step": 4395 + }, + { + "epoch": 8.14074074074074, + "grad_norm": 1.4174705384243207, + "learning_rate": 1.8461995581254634e-07, + "loss": 0.6474, + "step": 4396 + }, + { + "epoch": 8.142592592592592, + "grad_norm": 1.565633661072717, + "learning_rate": 1.8461196494005788e-07, + "loss": 0.6497, + "step": 4397 + }, + { + "epoch": 8.144444444444444, + "grad_norm": 1.4862483121816472, + "learning_rate": 1.8460397216526262e-07, + "loss": 0.6921, + "step": 4398 + }, + { + "epoch": 8.146296296296295, + "grad_norm": 1.4742831036198445, + "learning_rate": 1.845959774883403e-07, + "loss": 0.6603, + "step": 4399 + }, + { + "epoch": 8.148148148148149, + "grad_norm": 1.5245190999193128, + "learning_rate": 1.8458798090947062e-07, + "loss": 0.6346, + "step": 4400 + }, + { + "epoch": 8.15, + "grad_norm": 2.6289947895979795, + "learning_rate": 1.8457998242883345e-07, + "loss": 0.6121, + "step": 4401 + }, + { + "epoch": 8.151851851851852, + "grad_norm": 1.5753656994842067, + "learning_rate": 1.845719820466085e-07, + "loss": 0.657, + "step": 4402 + }, + { + "epoch": 8.153703703703703, + "grad_norm": 1.6369488831914727, + "learning_rate": 1.8456397976297574e-07, + "loss": 0.6802, + "step": 4403 + }, + { + "epoch": 8.155555555555555, + "grad_norm": 1.5602197748074216, + "learning_rate": 1.84555975578115e-07, + "loss": 0.7119, + "step": 4404 + }, + { + "epoch": 8.157407407407407, + "grad_norm": 1.419753530834734, + "learning_rate": 1.845479694922063e-07, + "loss": 0.6218, + "step": 4405 + }, + { + "epoch": 8.15925925925926, + "grad_norm": 1.5965928258490798, + "learning_rate": 1.8453996150542964e-07, + "loss": 0.6473, + "step": 4406 + }, + { + "epoch": 8.161111111111111, + "grad_norm": 1.4252535019427055, + "learning_rate": 1.84531951617965e-07, + "loss": 0.6442, + "step": 4407 + }, + { + "epoch": 8.162962962962963, + "grad_norm": 1.7328641861437968, + "learning_rate": 1.8452393982999253e-07, + "loss": 0.6223, + "step": 4408 + }, + { + "epoch": 8.164814814814815, + "grad_norm": 1.4578094338045657, + "learning_rate": 1.845159261416923e-07, + "loss": 0.6682, + "step": 4409 + }, + { + "epoch": 8.166666666666666, + "grad_norm": 1.566527952702924, + "learning_rate": 1.8450791055324454e-07, + "loss": 0.6322, + "step": 4410 + }, + { + "epoch": 8.168518518518518, + "grad_norm": 1.4457926849917626, + "learning_rate": 1.8449989306482941e-07, + "loss": 0.6522, + "step": 4411 + }, + { + "epoch": 8.170370370370371, + "grad_norm": 1.5107172795680537, + "learning_rate": 1.8449187367662723e-07, + "loss": 0.6493, + "step": 4412 + }, + { + "epoch": 8.172222222222222, + "grad_norm": 1.5901504504824933, + "learning_rate": 1.8448385238881818e-07, + "loss": 0.6378, + "step": 4413 + }, + { + "epoch": 8.174074074074074, + "grad_norm": 1.6261408486277367, + "learning_rate": 1.8447582920158273e-07, + "loss": 0.6503, + "step": 4414 + }, + { + "epoch": 8.175925925925926, + "grad_norm": 1.4731146018208923, + "learning_rate": 1.844678041151012e-07, + "loss": 0.6382, + "step": 4415 + }, + { + "epoch": 8.177777777777777, + "grad_norm": 1.650804041661046, + "learning_rate": 1.8445977712955402e-07, + "loss": 0.632, + "step": 4416 + }, + { + "epoch": 8.17962962962963, + "grad_norm": 1.8744121055056089, + "learning_rate": 1.8445174824512165e-07, + "loss": 0.6274, + "step": 4417 + }, + { + "epoch": 8.181481481481482, + "grad_norm": 1.520125284346202, + "learning_rate": 1.844437174619846e-07, + "loss": 0.6348, + "step": 4418 + }, + { + "epoch": 8.183333333333334, + "grad_norm": 2.0812536694644, + "learning_rate": 1.8443568478032345e-07, + "loss": 0.6571, + "step": 4419 + }, + { + "epoch": 8.185185185185185, + "grad_norm": 1.5159981249232075, + "learning_rate": 1.8442765020031876e-07, + "loss": 0.6687, + "step": 4420 + }, + { + "epoch": 8.187037037037037, + "grad_norm": 1.5200485704478703, + "learning_rate": 1.8441961372215123e-07, + "loss": 0.6174, + "step": 4421 + }, + { + "epoch": 8.188888888888888, + "grad_norm": 1.6702124172076425, + "learning_rate": 1.8441157534600146e-07, + "loss": 0.6214, + "step": 4422 + }, + { + "epoch": 8.190740740740742, + "grad_norm": 1.558946158227551, + "learning_rate": 1.8440353507205026e-07, + "loss": 0.6494, + "step": 4423 + }, + { + "epoch": 8.192592592592593, + "grad_norm": 1.3178292131506788, + "learning_rate": 1.8439549290047833e-07, + "loss": 0.6148, + "step": 4424 + }, + { + "epoch": 8.194444444444445, + "grad_norm": 1.6258181046546634, + "learning_rate": 1.8438744883146653e-07, + "loss": 0.6137, + "step": 4425 + }, + { + "epoch": 8.196296296296296, + "grad_norm": 1.933349907683464, + "learning_rate": 1.8437940286519568e-07, + "loss": 0.6616, + "step": 4426 + }, + { + "epoch": 8.198148148148148, + "grad_norm": 1.550809557818013, + "learning_rate": 1.8437135500184666e-07, + "loss": 0.6444, + "step": 4427 + }, + { + "epoch": 8.2, + "grad_norm": 1.7348700374583959, + "learning_rate": 1.8436330524160044e-07, + "loss": 0.6476, + "step": 4428 + }, + { + "epoch": 8.201851851851853, + "grad_norm": 1.5675733841164587, + "learning_rate": 1.84355253584638e-07, + "loss": 0.6734, + "step": 4429 + }, + { + "epoch": 8.203703703703704, + "grad_norm": 1.4643498703928413, + "learning_rate": 1.8434720003114036e-07, + "loss": 0.6048, + "step": 4430 + }, + { + "epoch": 8.205555555555556, + "grad_norm": 1.7356581217539526, + "learning_rate": 1.8433914458128858e-07, + "loss": 0.6501, + "step": 4431 + }, + { + "epoch": 8.207407407407407, + "grad_norm": 1.749789735311238, + "learning_rate": 1.8433108723526373e-07, + "loss": 0.6533, + "step": 4432 + }, + { + "epoch": 8.209259259259259, + "grad_norm": 1.4760663658099382, + "learning_rate": 1.8432302799324703e-07, + "loss": 0.6839, + "step": 4433 + }, + { + "epoch": 8.21111111111111, + "grad_norm": 1.3915985742078312, + "learning_rate": 1.8431496685541966e-07, + "loss": 0.6382, + "step": 4434 + }, + { + "epoch": 8.212962962962964, + "grad_norm": 1.5337320469283482, + "learning_rate": 1.843069038219628e-07, + "loss": 0.6657, + "step": 4435 + }, + { + "epoch": 8.214814814814815, + "grad_norm": 1.9392122524922977, + "learning_rate": 1.842988388930578e-07, + "loss": 0.648, + "step": 4436 + }, + { + "epoch": 8.216666666666667, + "grad_norm": 2.6080664065494013, + "learning_rate": 1.8429077206888592e-07, + "loss": 0.598, + "step": 4437 + }, + { + "epoch": 8.218518518518518, + "grad_norm": 1.4176907829478624, + "learning_rate": 1.842827033496286e-07, + "loss": 0.604, + "step": 4438 + }, + { + "epoch": 8.22037037037037, + "grad_norm": 1.4349506000776107, + "learning_rate": 1.8427463273546716e-07, + "loss": 0.6145, + "step": 4439 + }, + { + "epoch": 8.222222222222221, + "grad_norm": 1.4518955441655228, + "learning_rate": 1.842665602265831e-07, + "loss": 0.6179, + "step": 4440 + }, + { + "epoch": 8.224074074074075, + "grad_norm": 1.4432296280533177, + "learning_rate": 1.842584858231579e-07, + "loss": 0.6196, + "step": 4441 + }, + { + "epoch": 8.225925925925926, + "grad_norm": 1.5930920736653553, + "learning_rate": 1.8425040952537309e-07, + "loss": 0.6222, + "step": 4442 + }, + { + "epoch": 8.227777777777778, + "grad_norm": 1.5437189024522038, + "learning_rate": 1.8424233133341025e-07, + "loss": 0.6446, + "step": 4443 + }, + { + "epoch": 8.22962962962963, + "grad_norm": 1.584257659466699, + "learning_rate": 1.8423425124745103e-07, + "loss": 0.6779, + "step": 4444 + }, + { + "epoch": 8.231481481481481, + "grad_norm": 2.6045755456545057, + "learning_rate": 1.8422616926767702e-07, + "loss": 0.656, + "step": 4445 + }, + { + "epoch": 8.233333333333333, + "grad_norm": 1.5696173877055712, + "learning_rate": 1.8421808539427002e-07, + "loss": 0.6494, + "step": 4446 + }, + { + "epoch": 8.235185185185186, + "grad_norm": 1.7944084779019387, + "learning_rate": 1.842099996274117e-07, + "loss": 0.6719, + "step": 4447 + }, + { + "epoch": 8.237037037037037, + "grad_norm": 1.4192688076247182, + "learning_rate": 1.8420191196728388e-07, + "loss": 0.6431, + "step": 4448 + }, + { + "epoch": 8.238888888888889, + "grad_norm": 1.615805838148867, + "learning_rate": 1.841938224140684e-07, + "loss": 0.6343, + "step": 4449 + }, + { + "epoch": 8.24074074074074, + "grad_norm": 1.4622559081780984, + "learning_rate": 1.8418573096794712e-07, + "loss": 0.6213, + "step": 4450 + }, + { + "epoch": 8.242592592592592, + "grad_norm": 1.4235125550651881, + "learning_rate": 1.8417763762910196e-07, + "loss": 0.6888, + "step": 4451 + }, + { + "epoch": 8.244444444444444, + "grad_norm": 1.5886904843159426, + "learning_rate": 1.841695423977149e-07, + "loss": 0.6277, + "step": 4452 + }, + { + "epoch": 8.246296296296297, + "grad_norm": 1.5425377616543834, + "learning_rate": 1.8416144527396789e-07, + "loss": 0.6399, + "step": 4453 + }, + { + "epoch": 8.248148148148148, + "grad_norm": 1.5471698447424052, + "learning_rate": 1.8415334625804303e-07, + "loss": 0.671, + "step": 4454 + }, + { + "epoch": 8.25, + "grad_norm": 1.4217648391956874, + "learning_rate": 1.8414524535012242e-07, + "loss": 0.6142, + "step": 4455 + }, + { + "epoch": 8.251851851851852, + "grad_norm": 1.5084265828577654, + "learning_rate": 1.841371425503881e-07, + "loss": 0.6833, + "step": 4456 + }, + { + "epoch": 8.253703703703703, + "grad_norm": 1.399182612384619, + "learning_rate": 1.8412903785902235e-07, + "loss": 0.635, + "step": 4457 + }, + { + "epoch": 8.255555555555556, + "grad_norm": 1.9357521076138207, + "learning_rate": 1.841209312762073e-07, + "loss": 0.6235, + "step": 4458 + }, + { + "epoch": 8.257407407407408, + "grad_norm": 1.5735292173773687, + "learning_rate": 1.841128228021253e-07, + "loss": 0.6613, + "step": 4459 + }, + { + "epoch": 8.25925925925926, + "grad_norm": 1.8104192216203563, + "learning_rate": 1.8410471243695855e-07, + "loss": 0.616, + "step": 4460 + }, + { + "epoch": 8.261111111111111, + "grad_norm": 1.5368353609788117, + "learning_rate": 1.8409660018088946e-07, + "loss": 0.6311, + "step": 4461 + }, + { + "epoch": 8.262962962962963, + "grad_norm": 1.7106857496061911, + "learning_rate": 1.840884860341004e-07, + "loss": 0.6851, + "step": 4462 + }, + { + "epoch": 8.264814814814814, + "grad_norm": 1.5679887869269284, + "learning_rate": 1.8408036999677377e-07, + "loss": 0.6266, + "step": 4463 + }, + { + "epoch": 8.266666666666667, + "grad_norm": 7.226378031845, + "learning_rate": 1.8407225206909208e-07, + "loss": 0.6169, + "step": 4464 + }, + { + "epoch": 8.268518518518519, + "grad_norm": 1.4476282976606532, + "learning_rate": 1.8406413225123784e-07, + "loss": 0.6375, + "step": 4465 + }, + { + "epoch": 8.27037037037037, + "grad_norm": 1.5106928677491207, + "learning_rate": 1.8405601054339355e-07, + "loss": 0.6639, + "step": 4466 + }, + { + "epoch": 8.272222222222222, + "grad_norm": 2.573263763803964, + "learning_rate": 1.8404788694574187e-07, + "loss": 0.632, + "step": 4467 + }, + { + "epoch": 8.274074074074074, + "grad_norm": 1.3784290685138725, + "learning_rate": 1.8403976145846543e-07, + "loss": 0.5972, + "step": 4468 + }, + { + "epoch": 8.275925925925925, + "grad_norm": 1.4775133270272727, + "learning_rate": 1.840316340817469e-07, + "loss": 0.6518, + "step": 4469 + }, + { + "epoch": 8.277777777777779, + "grad_norm": 1.4429689298292858, + "learning_rate": 1.84023504815769e-07, + "loss": 0.6619, + "step": 4470 + }, + { + "epoch": 8.27962962962963, + "grad_norm": 1.5634488793950807, + "learning_rate": 1.8401537366071452e-07, + "loss": 0.6402, + "step": 4471 + }, + { + "epoch": 8.281481481481482, + "grad_norm": 1.3815683476118397, + "learning_rate": 1.8400724061676626e-07, + "loss": 0.6083, + "step": 4472 + }, + { + "epoch": 8.283333333333333, + "grad_norm": 1.4814211980123844, + "learning_rate": 1.8399910568410705e-07, + "loss": 0.6206, + "step": 4473 + }, + { + "epoch": 8.285185185185185, + "grad_norm": 1.5126234557080462, + "learning_rate": 1.8399096886291983e-07, + "loss": 0.6569, + "step": 4474 + }, + { + "epoch": 8.287037037037036, + "grad_norm": 1.4905478081848376, + "learning_rate": 1.8398283015338748e-07, + "loss": 0.6752, + "step": 4475 + }, + { + "epoch": 8.28888888888889, + "grad_norm": 1.6238189415991262, + "learning_rate": 1.8397468955569307e-07, + "loss": 0.6879, + "step": 4476 + }, + { + "epoch": 8.290740740740741, + "grad_norm": 1.7677577704848157, + "learning_rate": 1.8396654707001952e-07, + "loss": 0.6402, + "step": 4477 + }, + { + "epoch": 8.292592592592593, + "grad_norm": 1.4088138604229454, + "learning_rate": 1.8395840269654996e-07, + "loss": 0.6259, + "step": 4478 + }, + { + "epoch": 8.294444444444444, + "grad_norm": 1.5362515791971183, + "learning_rate": 1.8395025643546748e-07, + "loss": 0.6472, + "step": 4479 + }, + { + "epoch": 8.296296296296296, + "grad_norm": 1.546211815592026, + "learning_rate": 1.8394210828695522e-07, + "loss": 0.6076, + "step": 4480 + }, + { + "epoch": 8.298148148148147, + "grad_norm": 1.5885515198620532, + "learning_rate": 1.839339582511964e-07, + "loss": 0.6363, + "step": 4481 + }, + { + "epoch": 8.3, + "grad_norm": 1.4788680858177672, + "learning_rate": 1.8392580632837423e-07, + "loss": 0.6375, + "step": 4482 + }, + { + "epoch": 8.301851851851852, + "grad_norm": 1.5580058305481075, + "learning_rate": 1.83917652518672e-07, + "loss": 0.6377, + "step": 4483 + }, + { + "epoch": 8.303703703703704, + "grad_norm": 1.691032618517879, + "learning_rate": 1.8390949682227302e-07, + "loss": 0.6499, + "step": 4484 + }, + { + "epoch": 8.305555555555555, + "grad_norm": 1.3978341668575072, + "learning_rate": 1.8390133923936064e-07, + "loss": 0.6636, + "step": 4485 + }, + { + "epoch": 8.307407407407407, + "grad_norm": 1.6007950023712536, + "learning_rate": 1.8389317977011828e-07, + "loss": 0.6402, + "step": 4486 + }, + { + "epoch": 8.309259259259258, + "grad_norm": 1.8494485784292016, + "learning_rate": 1.838850184147294e-07, + "loss": 0.6699, + "step": 4487 + }, + { + "epoch": 8.311111111111112, + "grad_norm": 1.8585086606793517, + "learning_rate": 1.838768551733775e-07, + "loss": 0.6507, + "step": 4488 + }, + { + "epoch": 8.312962962962963, + "grad_norm": 1.5945178824609725, + "learning_rate": 1.8386869004624604e-07, + "loss": 0.6489, + "step": 4489 + }, + { + "epoch": 8.314814814814815, + "grad_norm": 1.4984619488016442, + "learning_rate": 1.838605230335187e-07, + "loss": 0.6517, + "step": 4490 + }, + { + "epoch": 8.316666666666666, + "grad_norm": 1.9315032752479253, + "learning_rate": 1.83852354135379e-07, + "loss": 0.6131, + "step": 4491 + }, + { + "epoch": 8.318518518518518, + "grad_norm": 1.3195694125447324, + "learning_rate": 1.8384418335201066e-07, + "loss": 0.6392, + "step": 4492 + }, + { + "epoch": 8.32037037037037, + "grad_norm": 1.6595583653607129, + "learning_rate": 1.8383601068359735e-07, + "loss": 0.641, + "step": 4493 + }, + { + "epoch": 8.322222222222223, + "grad_norm": 1.3498339808571895, + "learning_rate": 1.8382783613032283e-07, + "loss": 0.6274, + "step": 4494 + }, + { + "epoch": 8.324074074074074, + "grad_norm": 1.5288211928657378, + "learning_rate": 1.838196596923709e-07, + "loss": 0.6629, + "step": 4495 + }, + { + "epoch": 8.325925925925926, + "grad_norm": 1.521044331787283, + "learning_rate": 1.8381148136992534e-07, + "loss": 0.5925, + "step": 4496 + }, + { + "epoch": 8.327777777777778, + "grad_norm": 2.1594079249317666, + "learning_rate": 1.8380330116317007e-07, + "loss": 0.6169, + "step": 4497 + }, + { + "epoch": 8.329629629629629, + "grad_norm": 1.9142829518684596, + "learning_rate": 1.8379511907228896e-07, + "loss": 0.6249, + "step": 4498 + }, + { + "epoch": 8.331481481481482, + "grad_norm": 2.190524264767541, + "learning_rate": 1.8378693509746602e-07, + "loss": 0.6115, + "step": 4499 + }, + { + "epoch": 8.333333333333334, + "grad_norm": 1.5447064729566953, + "learning_rate": 1.837787492388852e-07, + "loss": 0.6323, + "step": 4500 + }, + { + "epoch": 8.335185185185185, + "grad_norm": 1.6328414471817385, + "learning_rate": 1.8377056149673057e-07, + "loss": 0.6525, + "step": 4501 + }, + { + "epoch": 8.337037037037037, + "grad_norm": 1.3366391145866388, + "learning_rate": 1.8376237187118617e-07, + "loss": 0.6197, + "step": 4502 + }, + { + "epoch": 8.338888888888889, + "grad_norm": 1.370917703692105, + "learning_rate": 1.8375418036243617e-07, + "loss": 0.6291, + "step": 4503 + }, + { + "epoch": 8.34074074074074, + "grad_norm": 1.8541410270394698, + "learning_rate": 1.8374598697066475e-07, + "loss": 0.6369, + "step": 4504 + }, + { + "epoch": 8.342592592592593, + "grad_norm": 1.4542215035162216, + "learning_rate": 1.8373779169605603e-07, + "loss": 0.6619, + "step": 4505 + }, + { + "epoch": 8.344444444444445, + "grad_norm": 1.6191045063866245, + "learning_rate": 1.8372959453879438e-07, + "loss": 0.6086, + "step": 4506 + }, + { + "epoch": 8.346296296296297, + "grad_norm": 1.824311218026256, + "learning_rate": 1.8372139549906402e-07, + "loss": 0.6554, + "step": 4507 + }, + { + "epoch": 8.348148148148148, + "grad_norm": 1.530900501725902, + "learning_rate": 1.8371319457704934e-07, + "loss": 0.6776, + "step": 4508 + }, + { + "epoch": 8.35, + "grad_norm": 1.6540338128844, + "learning_rate": 1.8370499177293462e-07, + "loss": 0.664, + "step": 4509 + }, + { + "epoch": 8.351851851851851, + "grad_norm": 1.6417826859270062, + "learning_rate": 1.8369678708690436e-07, + "loss": 0.6686, + "step": 4510 + }, + { + "epoch": 8.353703703703705, + "grad_norm": 1.9617097612692143, + "learning_rate": 1.8368858051914304e-07, + "loss": 0.6486, + "step": 4511 + }, + { + "epoch": 8.355555555555556, + "grad_norm": 1.702950181609952, + "learning_rate": 1.8368037206983513e-07, + "loss": 0.645, + "step": 4512 + }, + { + "epoch": 8.357407407407408, + "grad_norm": 1.4557490552542212, + "learning_rate": 1.836721617391652e-07, + "loss": 0.6388, + "step": 4513 + }, + { + "epoch": 8.35925925925926, + "grad_norm": 1.5668711873375605, + "learning_rate": 1.8366394952731776e-07, + "loss": 0.6276, + "step": 4514 + }, + { + "epoch": 8.36111111111111, + "grad_norm": 1.4112367621654067, + "learning_rate": 1.8365573543447754e-07, + "loss": 0.6653, + "step": 4515 + }, + { + "epoch": 8.362962962962962, + "grad_norm": 1.604709948789727, + "learning_rate": 1.8364751946082918e-07, + "loss": 0.6596, + "step": 4516 + }, + { + "epoch": 8.364814814814816, + "grad_norm": 1.4546142157849635, + "learning_rate": 1.8363930160655744e-07, + "loss": 0.6644, + "step": 4517 + }, + { + "epoch": 8.366666666666667, + "grad_norm": 1.6605625604766467, + "learning_rate": 1.83631081871847e-07, + "loss": 0.6201, + "step": 4518 + }, + { + "epoch": 8.368518518518519, + "grad_norm": 1.4296544642826179, + "learning_rate": 1.836228602568827e-07, + "loss": 0.6216, + "step": 4519 + }, + { + "epoch": 8.37037037037037, + "grad_norm": 1.5080770207724064, + "learning_rate": 1.8361463676184938e-07, + "loss": 0.5978, + "step": 4520 + }, + { + "epoch": 8.372222222222222, + "grad_norm": 1.4510967381213307, + "learning_rate": 1.8360641138693195e-07, + "loss": 0.6804, + "step": 4521 + }, + { + "epoch": 8.374074074074073, + "grad_norm": 1.6864678245257798, + "learning_rate": 1.8359818413231531e-07, + "loss": 0.6138, + "step": 4522 + }, + { + "epoch": 8.375925925925927, + "grad_norm": 1.8309270085825362, + "learning_rate": 1.8358995499818442e-07, + "loss": 0.6183, + "step": 4523 + }, + { + "epoch": 8.377777777777778, + "grad_norm": 1.581548293298123, + "learning_rate": 1.8358172398472436e-07, + "loss": 0.6269, + "step": 4524 + }, + { + "epoch": 8.37962962962963, + "grad_norm": 1.5912797255674556, + "learning_rate": 1.8357349109212013e-07, + "loss": 0.6608, + "step": 4525 + }, + { + "epoch": 8.381481481481481, + "grad_norm": 1.4686004244078337, + "learning_rate": 1.8356525632055683e-07, + "loss": 0.6687, + "step": 4526 + }, + { + "epoch": 8.383333333333333, + "grad_norm": 1.476903957886199, + "learning_rate": 1.8355701967021958e-07, + "loss": 0.6605, + "step": 4527 + }, + { + "epoch": 8.385185185185184, + "grad_norm": 1.4229082800769943, + "learning_rate": 1.8354878114129364e-07, + "loss": 0.6205, + "step": 4528 + }, + { + "epoch": 8.387037037037038, + "grad_norm": 1.6011482064276765, + "learning_rate": 1.8354054073396417e-07, + "loss": 0.6494, + "step": 4529 + }, + { + "epoch": 8.38888888888889, + "grad_norm": 1.4849614245967997, + "learning_rate": 1.8353229844841644e-07, + "loss": 0.6178, + "step": 4530 + }, + { + "epoch": 8.39074074074074, + "grad_norm": 1.4415165403682586, + "learning_rate": 1.8352405428483575e-07, + "loss": 0.6778, + "step": 4531 + }, + { + "epoch": 8.392592592592592, + "grad_norm": 1.4330827623318043, + "learning_rate": 1.8351580824340752e-07, + "loss": 0.6649, + "step": 4532 + }, + { + "epoch": 8.394444444444444, + "grad_norm": 1.562137986180405, + "learning_rate": 1.8350756032431708e-07, + "loss": 0.7017, + "step": 4533 + }, + { + "epoch": 8.396296296296295, + "grad_norm": 1.491515292892923, + "learning_rate": 1.8349931052774984e-07, + "loss": 0.6398, + "step": 4534 + }, + { + "epoch": 8.398148148148149, + "grad_norm": 1.4577729697546422, + "learning_rate": 1.8349105885389138e-07, + "loss": 0.6289, + "step": 4535 + }, + { + "epoch": 8.4, + "grad_norm": 1.7012731865284316, + "learning_rate": 1.834828053029271e-07, + "loss": 0.6305, + "step": 4536 + }, + { + "epoch": 8.401851851851852, + "grad_norm": 1.5534133877230325, + "learning_rate": 1.8347454987504266e-07, + "loss": 0.6549, + "step": 4537 + }, + { + "epoch": 8.403703703703703, + "grad_norm": 2.06982233609931, + "learning_rate": 1.834662925704236e-07, + "loss": 0.5926, + "step": 4538 + }, + { + "epoch": 8.405555555555555, + "grad_norm": 1.3814469697110232, + "learning_rate": 1.834580333892556e-07, + "loss": 0.6392, + "step": 4539 + }, + { + "epoch": 8.407407407407407, + "grad_norm": 1.382295643754722, + "learning_rate": 1.8344977233172437e-07, + "loss": 0.665, + "step": 4540 + }, + { + "epoch": 8.40925925925926, + "grad_norm": 1.3990804594408262, + "learning_rate": 1.834415093980156e-07, + "loss": 0.6354, + "step": 4541 + }, + { + "epoch": 8.411111111111111, + "grad_norm": 1.363437766916677, + "learning_rate": 1.8343324458831506e-07, + "loss": 0.6374, + "step": 4542 + }, + { + "epoch": 8.412962962962963, + "grad_norm": 1.6314452720963881, + "learning_rate": 1.8342497790280857e-07, + "loss": 0.6608, + "step": 4543 + }, + { + "epoch": 8.414814814814815, + "grad_norm": 1.3630394580862388, + "learning_rate": 1.8341670934168204e-07, + "loss": 0.6233, + "step": 4544 + }, + { + "epoch": 8.416666666666666, + "grad_norm": 1.9049982870144648, + "learning_rate": 1.8340843890512127e-07, + "loss": 0.5932, + "step": 4545 + }, + { + "epoch": 8.418518518518518, + "grad_norm": 1.874352471556863, + "learning_rate": 1.834001665933123e-07, + "loss": 0.6537, + "step": 4546 + }, + { + "epoch": 8.420370370370371, + "grad_norm": 1.6993176651426807, + "learning_rate": 1.833918924064411e-07, + "loss": 0.6541, + "step": 4547 + }, + { + "epoch": 8.422222222222222, + "grad_norm": 1.564222617365321, + "learning_rate": 1.833836163446936e-07, + "loss": 0.631, + "step": 4548 + }, + { + "epoch": 8.424074074074074, + "grad_norm": 1.8963502602083337, + "learning_rate": 1.8337533840825595e-07, + "loss": 0.6853, + "step": 4549 + }, + { + "epoch": 8.425925925925926, + "grad_norm": 1.4688978684591096, + "learning_rate": 1.833670585973143e-07, + "loss": 0.6844, + "step": 4550 + }, + { + "epoch": 8.427777777777777, + "grad_norm": 1.595942662172928, + "learning_rate": 1.8335877691205472e-07, + "loss": 0.6047, + "step": 4551 + }, + { + "epoch": 8.42962962962963, + "grad_norm": 1.4500602902578772, + "learning_rate": 1.8335049335266344e-07, + "loss": 0.6364, + "step": 4552 + }, + { + "epoch": 8.431481481481482, + "grad_norm": 1.3893302144834807, + "learning_rate": 1.833422079193267e-07, + "loss": 0.6477, + "step": 4553 + }, + { + "epoch": 8.433333333333334, + "grad_norm": 1.4838805938301063, + "learning_rate": 1.8333392061223078e-07, + "loss": 0.6639, + "step": 4554 + }, + { + "epoch": 8.435185185185185, + "grad_norm": 1.4699031024385159, + "learning_rate": 1.8332563143156193e-07, + "loss": 0.635, + "step": 4555 + }, + { + "epoch": 8.437037037037037, + "grad_norm": 1.4374806778754223, + "learning_rate": 1.8331734037750664e-07, + "loss": 0.5841, + "step": 4556 + }, + { + "epoch": 8.438888888888888, + "grad_norm": 1.5296147639887085, + "learning_rate": 1.8330904745025122e-07, + "loss": 0.6831, + "step": 4557 + }, + { + "epoch": 8.440740740740742, + "grad_norm": 1.5381406069873516, + "learning_rate": 1.8330075264998218e-07, + "loss": 0.583, + "step": 4558 + }, + { + "epoch": 8.442592592592593, + "grad_norm": 1.7699048857280157, + "learning_rate": 1.8329245597688594e-07, + "loss": 0.6491, + "step": 4559 + }, + { + "epoch": 8.444444444444445, + "grad_norm": 1.4577225336014208, + "learning_rate": 1.8328415743114912e-07, + "loss": 0.6554, + "step": 4560 + }, + { + "epoch": 8.446296296296296, + "grad_norm": 1.5996410280927722, + "learning_rate": 1.832758570129582e-07, + "loss": 0.6051, + "step": 4561 + }, + { + "epoch": 8.448148148148148, + "grad_norm": 1.434765292793164, + "learning_rate": 1.8326755472249985e-07, + "loss": 0.7148, + "step": 4562 + }, + { + "epoch": 8.45, + "grad_norm": 1.4282866469760362, + "learning_rate": 1.8325925055996074e-07, + "loss": 0.6291, + "step": 4563 + }, + { + "epoch": 8.451851851851853, + "grad_norm": 1.6708832412222447, + "learning_rate": 1.8325094452552755e-07, + "loss": 0.6581, + "step": 4564 + }, + { + "epoch": 8.453703703703704, + "grad_norm": 1.7123286697631361, + "learning_rate": 1.8324263661938702e-07, + "loss": 0.6125, + "step": 4565 + }, + { + "epoch": 8.455555555555556, + "grad_norm": 1.4812133951355995, + "learning_rate": 1.832343268417259e-07, + "loss": 0.6699, + "step": 4566 + }, + { + "epoch": 8.457407407407407, + "grad_norm": 1.5158201209553643, + "learning_rate": 1.8322601519273107e-07, + "loss": 0.6646, + "step": 4567 + }, + { + "epoch": 8.459259259259259, + "grad_norm": 1.7434690183775805, + "learning_rate": 1.832177016725894e-07, + "loss": 0.6511, + "step": 4568 + }, + { + "epoch": 8.46111111111111, + "grad_norm": 1.8785515829188772, + "learning_rate": 1.8320938628148777e-07, + "loss": 0.653, + "step": 4569 + }, + { + "epoch": 8.462962962962964, + "grad_norm": 1.4281991813971242, + "learning_rate": 1.8320106901961311e-07, + "loss": 0.6911, + "step": 4570 + }, + { + "epoch": 8.464814814814815, + "grad_norm": 1.3679988205675195, + "learning_rate": 1.8319274988715248e-07, + "loss": 0.6357, + "step": 4571 + }, + { + "epoch": 8.466666666666667, + "grad_norm": 1.7775941656739307, + "learning_rate": 1.831844288842929e-07, + "loss": 0.6462, + "step": 4572 + }, + { + "epoch": 8.468518518518518, + "grad_norm": 1.5231018287887081, + "learning_rate": 1.831761060112214e-07, + "loss": 0.6409, + "step": 4573 + }, + { + "epoch": 8.47037037037037, + "grad_norm": 1.4144579142939049, + "learning_rate": 1.8316778126812515e-07, + "loss": 0.5929, + "step": 4574 + }, + { + "epoch": 8.472222222222221, + "grad_norm": 1.3834680242035378, + "learning_rate": 1.831594546551913e-07, + "loss": 0.6291, + "step": 4575 + }, + { + "epoch": 8.474074074074075, + "grad_norm": 1.57796722508518, + "learning_rate": 1.8315112617260703e-07, + "loss": 0.617, + "step": 4576 + }, + { + "epoch": 8.475925925925926, + "grad_norm": 1.6181525456398926, + "learning_rate": 1.8314279582055963e-07, + "loss": 0.6773, + "step": 4577 + }, + { + "epoch": 8.477777777777778, + "grad_norm": 1.5919046856276704, + "learning_rate": 1.8313446359923636e-07, + "loss": 0.6523, + "step": 4578 + }, + { + "epoch": 8.47962962962963, + "grad_norm": 1.5173886834138257, + "learning_rate": 1.8312612950882456e-07, + "loss": 0.6245, + "step": 4579 + }, + { + "epoch": 8.481481481481481, + "grad_norm": 1.371609178734613, + "learning_rate": 1.831177935495116e-07, + "loss": 0.6377, + "step": 4580 + }, + { + "epoch": 8.483333333333333, + "grad_norm": 1.7343617199354378, + "learning_rate": 1.8310945572148491e-07, + "loss": 0.6296, + "step": 4581 + }, + { + "epoch": 8.485185185185186, + "grad_norm": 1.6157766312139887, + "learning_rate": 1.8310111602493193e-07, + "loss": 0.645, + "step": 4582 + }, + { + "epoch": 8.487037037037037, + "grad_norm": 2.2192004255214353, + "learning_rate": 1.8309277446004012e-07, + "loss": 0.6336, + "step": 4583 + }, + { + "epoch": 8.488888888888889, + "grad_norm": 1.5177882417896689, + "learning_rate": 1.830844310269971e-07, + "loss": 0.6259, + "step": 4584 + }, + { + "epoch": 8.49074074074074, + "grad_norm": 2.2876322684769095, + "learning_rate": 1.830760857259904e-07, + "loss": 0.6529, + "step": 4585 + }, + { + "epoch": 8.492592592592592, + "grad_norm": 1.6443906210680563, + "learning_rate": 1.830677385572077e-07, + "loss": 0.6066, + "step": 4586 + }, + { + "epoch": 8.494444444444444, + "grad_norm": 1.6620671081429497, + "learning_rate": 1.830593895208366e-07, + "loss": 0.605, + "step": 4587 + }, + { + "epoch": 8.496296296296297, + "grad_norm": 1.6237177762667632, + "learning_rate": 1.8305103861706485e-07, + "loss": 0.6037, + "step": 4588 + }, + { + "epoch": 8.498148148148148, + "grad_norm": 1.5424428006789714, + "learning_rate": 1.8304268584608017e-07, + "loss": 0.5897, + "step": 4589 + }, + { + "epoch": 8.5, + "grad_norm": 1.4428500929496997, + "learning_rate": 1.830343312080704e-07, + "loss": 0.6281, + "step": 4590 + }, + { + "epoch": 8.501851851851852, + "grad_norm": 1.4226147571645114, + "learning_rate": 1.8302597470322332e-07, + "loss": 0.6322, + "step": 4591 + }, + { + "epoch": 8.503703703703703, + "grad_norm": 3.66458925985423, + "learning_rate": 1.8301761633172684e-07, + "loss": 0.6583, + "step": 4592 + }, + { + "epoch": 8.505555555555556, + "grad_norm": 1.4721840927227268, + "learning_rate": 1.8300925609376887e-07, + "loss": 0.6681, + "step": 4593 + }, + { + "epoch": 8.507407407407408, + "grad_norm": 2.296984133904764, + "learning_rate": 1.8300089398953737e-07, + "loss": 0.6372, + "step": 4594 + }, + { + "epoch": 8.50925925925926, + "grad_norm": 1.6193845328258152, + "learning_rate": 1.8299253001922036e-07, + "loss": 0.6082, + "step": 4595 + }, + { + "epoch": 8.511111111111111, + "grad_norm": 1.7125968009672943, + "learning_rate": 1.829841641830059e-07, + "loss": 0.6505, + "step": 4596 + }, + { + "epoch": 8.512962962962963, + "grad_norm": 1.4994943326885157, + "learning_rate": 1.8297579648108198e-07, + "loss": 0.6387, + "step": 4597 + }, + { + "epoch": 8.514814814814814, + "grad_norm": 1.4561646082662318, + "learning_rate": 1.8296742691363685e-07, + "loss": 0.6573, + "step": 4598 + }, + { + "epoch": 8.516666666666667, + "grad_norm": 1.6034150146411446, + "learning_rate": 1.829590554808586e-07, + "loss": 0.6316, + "step": 4599 + }, + { + "epoch": 8.518518518518519, + "grad_norm": 1.6917244304149643, + "learning_rate": 1.8295068218293545e-07, + "loss": 0.6308, + "step": 4600 + }, + { + "epoch": 8.52037037037037, + "grad_norm": 1.2741846009401572, + "learning_rate": 1.829423070200557e-07, + "loss": 0.6452, + "step": 4601 + }, + { + "epoch": 8.522222222222222, + "grad_norm": 1.7116703651677392, + "learning_rate": 1.8293392999240765e-07, + "loss": 0.6435, + "step": 4602 + }, + { + "epoch": 8.524074074074074, + "grad_norm": 1.5014199062805065, + "learning_rate": 1.8292555110017958e-07, + "loss": 0.6613, + "step": 4603 + }, + { + "epoch": 8.525925925925925, + "grad_norm": 1.5460828042744386, + "learning_rate": 1.8291717034355988e-07, + "loss": 0.6739, + "step": 4604 + }, + { + "epoch": 8.527777777777779, + "grad_norm": 1.5500219760823115, + "learning_rate": 1.82908787722737e-07, + "loss": 0.6254, + "step": 4605 + }, + { + "epoch": 8.52962962962963, + "grad_norm": 1.570950037911747, + "learning_rate": 1.8290040323789943e-07, + "loss": 0.6436, + "step": 4606 + }, + { + "epoch": 8.531481481481482, + "grad_norm": 1.4662886803866015, + "learning_rate": 1.8289201688923562e-07, + "loss": 0.6453, + "step": 4607 + }, + { + "epoch": 8.533333333333333, + "grad_norm": 2.140665127568465, + "learning_rate": 1.8288362867693413e-07, + "loss": 0.6457, + "step": 4608 + }, + { + "epoch": 8.535185185185185, + "grad_norm": 1.5446819930546778, + "learning_rate": 1.8287523860118356e-07, + "loss": 0.6621, + "step": 4609 + }, + { + "epoch": 8.537037037037036, + "grad_norm": 2.2355785679658857, + "learning_rate": 1.8286684666217255e-07, + "loss": 0.6489, + "step": 4610 + }, + { + "epoch": 8.53888888888889, + "grad_norm": 1.545796072142865, + "learning_rate": 1.8285845286008973e-07, + "loss": 0.6373, + "step": 4611 + }, + { + "epoch": 8.540740740740741, + "grad_norm": 1.5205906408171752, + "learning_rate": 1.8285005719512388e-07, + "loss": 0.6166, + "step": 4612 + }, + { + "epoch": 8.542592592592593, + "grad_norm": 1.3964947578471048, + "learning_rate": 1.8284165966746373e-07, + "loss": 0.6373, + "step": 4613 + }, + { + "epoch": 8.544444444444444, + "grad_norm": 1.56279393515347, + "learning_rate": 1.828332602772981e-07, + "loss": 0.5994, + "step": 4614 + }, + { + "epoch": 8.546296296296296, + "grad_norm": 1.5192694019191217, + "learning_rate": 1.8282485902481575e-07, + "loss": 0.6423, + "step": 4615 + }, + { + "epoch": 8.548148148148147, + "grad_norm": 1.560031159362777, + "learning_rate": 1.8281645591020564e-07, + "loss": 0.6416, + "step": 4616 + }, + { + "epoch": 8.55, + "grad_norm": 1.3911895632401228, + "learning_rate": 1.828080509336567e-07, + "loss": 0.6522, + "step": 4617 + }, + { + "epoch": 8.551851851851852, + "grad_norm": 1.5042651037844108, + "learning_rate": 1.8279964409535783e-07, + "loss": 0.6435, + "step": 4618 + }, + { + "epoch": 8.553703703703704, + "grad_norm": 1.7811036677197682, + "learning_rate": 1.8279123539549812e-07, + "loss": 0.6001, + "step": 4619 + }, + { + "epoch": 8.555555555555555, + "grad_norm": 1.5563950588343298, + "learning_rate": 1.8278282483426657e-07, + "loss": 0.656, + "step": 4620 + }, + { + "epoch": 8.557407407407407, + "grad_norm": 1.6224035928703437, + "learning_rate": 1.8277441241185228e-07, + "loss": 0.6884, + "step": 4621 + }, + { + "epoch": 8.559259259259258, + "grad_norm": 3.539910385848698, + "learning_rate": 1.8276599812844437e-07, + "loss": 0.5912, + "step": 4622 + }, + { + "epoch": 8.561111111111112, + "grad_norm": 1.4926072753544442, + "learning_rate": 1.8275758198423205e-07, + "loss": 0.6165, + "step": 4623 + }, + { + "epoch": 8.562962962962963, + "grad_norm": 1.6971106085024001, + "learning_rate": 1.8274916397940453e-07, + "loss": 0.6497, + "step": 4624 + }, + { + "epoch": 8.564814814814815, + "grad_norm": 1.445082588868154, + "learning_rate": 1.8274074411415104e-07, + "loss": 0.6245, + "step": 4625 + }, + { + "epoch": 8.566666666666666, + "grad_norm": 1.6964333323163376, + "learning_rate": 1.827323223886609e-07, + "loss": 0.6466, + "step": 4626 + }, + { + "epoch": 8.568518518518518, + "grad_norm": 1.37633204202434, + "learning_rate": 1.827238988031235e-07, + "loss": 0.6411, + "step": 4627 + }, + { + "epoch": 8.57037037037037, + "grad_norm": 1.5430830872250902, + "learning_rate": 1.8271547335772811e-07, + "loss": 0.6436, + "step": 4628 + }, + { + "epoch": 8.572222222222223, + "grad_norm": 1.9230183394575824, + "learning_rate": 1.827070460526643e-07, + "loss": 0.6266, + "step": 4629 + }, + { + "epoch": 8.574074074074074, + "grad_norm": 1.4166430563025159, + "learning_rate": 1.826986168881214e-07, + "loss": 0.6583, + "step": 4630 + }, + { + "epoch": 8.575925925925926, + "grad_norm": 1.5076059966983177, + "learning_rate": 1.8269018586428903e-07, + "loss": 0.6081, + "step": 4631 + }, + { + "epoch": 8.577777777777778, + "grad_norm": 1.4557491288992066, + "learning_rate": 1.826817529813567e-07, + "loss": 0.6049, + "step": 4632 + }, + { + "epoch": 8.579629629629629, + "grad_norm": 1.4809285355161672, + "learning_rate": 1.8267331823951397e-07, + "loss": 0.6781, + "step": 4633 + }, + { + "epoch": 8.581481481481482, + "grad_norm": 1.4212208772258825, + "learning_rate": 1.8266488163895054e-07, + "loss": 0.6933, + "step": 4634 + }, + { + "epoch": 8.583333333333334, + "grad_norm": 1.5799924915930879, + "learning_rate": 1.8265644317985602e-07, + "loss": 0.646, + "step": 4635 + }, + { + "epoch": 8.585185185185185, + "grad_norm": 1.5573144804872525, + "learning_rate": 1.8264800286242022e-07, + "loss": 0.6869, + "step": 4636 + }, + { + "epoch": 8.587037037037037, + "grad_norm": 1.452921149434807, + "learning_rate": 1.826395606868328e-07, + "loss": 0.6283, + "step": 4637 + }, + { + "epoch": 8.588888888888889, + "grad_norm": 1.7743806641822972, + "learning_rate": 1.8263111665328362e-07, + "loss": 0.6298, + "step": 4638 + }, + { + "epoch": 8.59074074074074, + "grad_norm": 1.5208365051868078, + "learning_rate": 1.8262267076196252e-07, + "loss": 0.5782, + "step": 4639 + }, + { + "epoch": 8.592592592592592, + "grad_norm": 1.4851290397447496, + "learning_rate": 1.826142230130594e-07, + "loss": 0.6403, + "step": 4640 + }, + { + "epoch": 8.594444444444445, + "grad_norm": 1.409981785128526, + "learning_rate": 1.8260577340676413e-07, + "loss": 0.6583, + "step": 4641 + }, + { + "epoch": 8.596296296296297, + "grad_norm": 1.9323666449845667, + "learning_rate": 1.8259732194326675e-07, + "loss": 0.6623, + "step": 4642 + }, + { + "epoch": 8.598148148148148, + "grad_norm": 1.6448795921633081, + "learning_rate": 1.8258886862275725e-07, + "loss": 0.6429, + "step": 4643 + }, + { + "epoch": 8.6, + "grad_norm": 1.4409766365499421, + "learning_rate": 1.8258041344542562e-07, + "loss": 0.6624, + "step": 4644 + }, + { + "epoch": 8.601851851851851, + "grad_norm": 1.848940669302376, + "learning_rate": 1.8257195641146206e-07, + "loss": 0.59, + "step": 4645 + }, + { + "epoch": 8.603703703703705, + "grad_norm": 1.645895531465644, + "learning_rate": 1.8256349752105665e-07, + "loss": 0.6406, + "step": 4646 + }, + { + "epoch": 8.605555555555556, + "grad_norm": 1.5332580219964758, + "learning_rate": 1.825550367743996e-07, + "loss": 0.6083, + "step": 4647 + }, + { + "epoch": 8.607407407407408, + "grad_norm": 1.572436611417677, + "learning_rate": 1.8254657417168107e-07, + "loss": 0.6334, + "step": 4648 + }, + { + "epoch": 8.60925925925926, + "grad_norm": 1.4398395258793437, + "learning_rate": 1.8253810971309135e-07, + "loss": 0.6532, + "step": 4649 + }, + { + "epoch": 8.61111111111111, + "grad_norm": 1.642290473024365, + "learning_rate": 1.825296433988208e-07, + "loss": 0.6629, + "step": 4650 + }, + { + "epoch": 8.612962962962962, + "grad_norm": 1.3984362963954995, + "learning_rate": 1.8252117522905968e-07, + "loss": 0.6596, + "step": 4651 + }, + { + "epoch": 8.614814814814816, + "grad_norm": 1.6116596754796, + "learning_rate": 1.8251270520399844e-07, + "loss": 0.6489, + "step": 4652 + }, + { + "epoch": 8.616666666666667, + "grad_norm": 2.69604900841089, + "learning_rate": 1.8250423332382747e-07, + "loss": 0.6542, + "step": 4653 + }, + { + "epoch": 8.618518518518519, + "grad_norm": 1.4288185593949385, + "learning_rate": 1.8249575958873726e-07, + "loss": 0.6453, + "step": 4654 + }, + { + "epoch": 8.62037037037037, + "grad_norm": 1.647910173806169, + "learning_rate": 1.8248728399891832e-07, + "loss": 0.6345, + "step": 4655 + }, + { + "epoch": 8.622222222222222, + "grad_norm": 1.3281289359551343, + "learning_rate": 1.8247880655456122e-07, + "loss": 0.6835, + "step": 4656 + }, + { + "epoch": 8.624074074074073, + "grad_norm": 1.4710730760143462, + "learning_rate": 1.8247032725585653e-07, + "loss": 0.6275, + "step": 4657 + }, + { + "epoch": 8.625925925925927, + "grad_norm": 1.564139242155735, + "learning_rate": 1.824618461029949e-07, + "loss": 0.646, + "step": 4658 + }, + { + "epoch": 8.627777777777778, + "grad_norm": 1.5376627191948549, + "learning_rate": 1.8245336309616698e-07, + "loss": 0.5937, + "step": 4659 + }, + { + "epoch": 8.62962962962963, + "grad_norm": 1.4797688334607642, + "learning_rate": 1.8244487823556357e-07, + "loss": 0.6373, + "step": 4660 + }, + { + "epoch": 8.631481481481481, + "grad_norm": 1.5830035686800359, + "learning_rate": 1.824363915213753e-07, + "loss": 0.6417, + "step": 4661 + }, + { + "epoch": 8.633333333333333, + "grad_norm": 1.6981934717298532, + "learning_rate": 1.8242790295379312e-07, + "loss": 0.6655, + "step": 4662 + }, + { + "epoch": 8.635185185185184, + "grad_norm": 2.054784709026285, + "learning_rate": 1.8241941253300782e-07, + "loss": 0.6896, + "step": 4663 + }, + { + "epoch": 8.637037037037038, + "grad_norm": 1.598474932463622, + "learning_rate": 1.8241092025921025e-07, + "loss": 0.6477, + "step": 4664 + }, + { + "epoch": 8.63888888888889, + "grad_norm": 1.61900113925755, + "learning_rate": 1.8240242613259134e-07, + "loss": 0.6107, + "step": 4665 + }, + { + "epoch": 8.64074074074074, + "grad_norm": 1.6328890300869303, + "learning_rate": 1.8239393015334216e-07, + "loss": 0.667, + "step": 4666 + }, + { + "epoch": 8.642592592592592, + "grad_norm": 1.485073056746911, + "learning_rate": 1.8238543232165357e-07, + "loss": 0.6064, + "step": 4667 + }, + { + "epoch": 8.644444444444444, + "grad_norm": 1.6212773723414033, + "learning_rate": 1.8237693263771674e-07, + "loss": 0.6109, + "step": 4668 + }, + { + "epoch": 8.646296296296295, + "grad_norm": 1.554389555194729, + "learning_rate": 1.8236843110172276e-07, + "loss": 0.6528, + "step": 4669 + }, + { + "epoch": 8.648148148148149, + "grad_norm": 1.6280280529034825, + "learning_rate": 1.8235992771386272e-07, + "loss": 0.6346, + "step": 4670 + }, + { + "epoch": 8.65, + "grad_norm": 1.7458232937442923, + "learning_rate": 1.823514224743278e-07, + "loss": 0.6223, + "step": 4671 + }, + { + "epoch": 8.651851851851852, + "grad_norm": 1.585324461122069, + "learning_rate": 1.823429153833093e-07, + "loss": 0.6516, + "step": 4672 + }, + { + "epoch": 8.653703703703703, + "grad_norm": 1.8145245840940134, + "learning_rate": 1.8233440644099835e-07, + "loss": 0.6666, + "step": 4673 + }, + { + "epoch": 8.655555555555555, + "grad_norm": 1.647551951510653, + "learning_rate": 1.8232589564758634e-07, + "loss": 0.6099, + "step": 4674 + }, + { + "epoch": 8.657407407407408, + "grad_norm": 1.7238778767416179, + "learning_rate": 1.8231738300326461e-07, + "loss": 0.6231, + "step": 4675 + }, + { + "epoch": 8.65925925925926, + "grad_norm": 1.5130845337692778, + "learning_rate": 1.8230886850822458e-07, + "loss": 0.6302, + "step": 4676 + }, + { + "epoch": 8.661111111111111, + "grad_norm": 1.6029840751467057, + "learning_rate": 1.823003521626576e-07, + "loss": 0.672, + "step": 4677 + }, + { + "epoch": 8.662962962962963, + "grad_norm": 1.4008070921977813, + "learning_rate": 1.8229183396675518e-07, + "loss": 0.6595, + "step": 4678 + }, + { + "epoch": 8.664814814814815, + "grad_norm": 1.5119462980495688, + "learning_rate": 1.8228331392070881e-07, + "loss": 0.673, + "step": 4679 + }, + { + "epoch": 8.666666666666666, + "grad_norm": 1.498123257492758, + "learning_rate": 1.8227479202471012e-07, + "loss": 0.6271, + "step": 4680 + }, + { + "epoch": 8.668518518518518, + "grad_norm": 1.8742693532260817, + "learning_rate": 1.8226626827895063e-07, + "loss": 0.6118, + "step": 4681 + }, + { + "epoch": 8.670370370370371, + "grad_norm": 2.63094956801755, + "learning_rate": 1.8225774268362198e-07, + "loss": 0.6411, + "step": 4682 + }, + { + "epoch": 8.672222222222222, + "grad_norm": 1.5429727271348872, + "learning_rate": 1.8224921523891585e-07, + "loss": 0.6299, + "step": 4683 + }, + { + "epoch": 8.674074074074074, + "grad_norm": 1.8526143298357611, + "learning_rate": 1.82240685945024e-07, + "loss": 0.6999, + "step": 4684 + }, + { + "epoch": 8.675925925925926, + "grad_norm": 1.702897015000899, + "learning_rate": 1.8223215480213818e-07, + "loss": 0.6128, + "step": 4685 + }, + { + "epoch": 8.677777777777777, + "grad_norm": 1.4412285634416815, + "learning_rate": 1.8222362181045016e-07, + "loss": 0.664, + "step": 4686 + }, + { + "epoch": 8.67962962962963, + "grad_norm": 1.667561604277166, + "learning_rate": 1.8221508697015182e-07, + "loss": 0.6524, + "step": 4687 + }, + { + "epoch": 8.681481481481482, + "grad_norm": 2.2931596283003164, + "learning_rate": 1.8220655028143503e-07, + "loss": 0.6119, + "step": 4688 + }, + { + "epoch": 8.683333333333334, + "grad_norm": 1.6072472899378067, + "learning_rate": 1.821980117444917e-07, + "loss": 0.6313, + "step": 4689 + }, + { + "epoch": 8.685185185185185, + "grad_norm": 1.6181307696646436, + "learning_rate": 1.8218947135951385e-07, + "loss": 0.6181, + "step": 4690 + }, + { + "epoch": 8.687037037037037, + "grad_norm": 1.3917733867377007, + "learning_rate": 1.8218092912669346e-07, + "loss": 0.6825, + "step": 4691 + }, + { + "epoch": 8.688888888888888, + "grad_norm": 1.54247117128281, + "learning_rate": 1.8217238504622255e-07, + "loss": 0.6575, + "step": 4692 + }, + { + "epoch": 8.690740740740742, + "grad_norm": 1.5563706340018573, + "learning_rate": 1.8216383911829327e-07, + "loss": 0.6197, + "step": 4693 + }, + { + "epoch": 8.692592592592593, + "grad_norm": 1.5874580072919338, + "learning_rate": 1.8215529134309774e-07, + "loss": 0.6482, + "step": 4694 + }, + { + "epoch": 8.694444444444445, + "grad_norm": 1.7687437304554279, + "learning_rate": 1.8214674172082814e-07, + "loss": 0.6855, + "step": 4695 + }, + { + "epoch": 8.696296296296296, + "grad_norm": 1.6009761158263907, + "learning_rate": 1.8213819025167664e-07, + "loss": 0.637, + "step": 4696 + }, + { + "epoch": 8.698148148148148, + "grad_norm": 1.4795538809233624, + "learning_rate": 1.8212963693583557e-07, + "loss": 0.6721, + "step": 4697 + }, + { + "epoch": 8.7, + "grad_norm": 1.719666487858422, + "learning_rate": 1.821210817734972e-07, + "loss": 0.6213, + "step": 4698 + }, + { + "epoch": 8.701851851851853, + "grad_norm": 1.488734611142881, + "learning_rate": 1.8211252476485384e-07, + "loss": 0.6431, + "step": 4699 + }, + { + "epoch": 8.703703703703704, + "grad_norm": 1.5299407513110628, + "learning_rate": 1.8210396591009794e-07, + "loss": 0.6551, + "step": 4700 + }, + { + "epoch": 8.705555555555556, + "grad_norm": 1.8923040523516026, + "learning_rate": 1.8209540520942189e-07, + "loss": 0.6621, + "step": 4701 + }, + { + "epoch": 8.707407407407407, + "grad_norm": 1.5644709130386947, + "learning_rate": 1.8208684266301818e-07, + "loss": 0.6582, + "step": 4702 + }, + { + "epoch": 8.709259259259259, + "grad_norm": 1.5950998731019763, + "learning_rate": 1.8207827827107924e-07, + "loss": 0.6127, + "step": 4703 + }, + { + "epoch": 8.71111111111111, + "grad_norm": 1.5045411998125697, + "learning_rate": 1.8206971203379773e-07, + "loss": 0.6344, + "step": 4704 + }, + { + "epoch": 8.712962962962964, + "grad_norm": 1.7444162133434356, + "learning_rate": 1.8206114395136622e-07, + "loss": 0.6268, + "step": 4705 + }, + { + "epoch": 8.714814814814815, + "grad_norm": 1.6332213368195259, + "learning_rate": 1.820525740239773e-07, + "loss": 0.6272, + "step": 4706 + }, + { + "epoch": 8.716666666666667, + "grad_norm": 1.4707964202906618, + "learning_rate": 1.8204400225182363e-07, + "loss": 0.6771, + "step": 4707 + }, + { + "epoch": 8.718518518518518, + "grad_norm": 1.4141728248616752, + "learning_rate": 1.82035428635098e-07, + "loss": 0.6369, + "step": 4708 + }, + { + "epoch": 8.72037037037037, + "grad_norm": 1.6191809489828612, + "learning_rate": 1.820268531739931e-07, + "loss": 0.6402, + "step": 4709 + }, + { + "epoch": 8.722222222222221, + "grad_norm": 1.5002667373306158, + "learning_rate": 1.820182758687018e-07, + "loss": 0.6276, + "step": 4710 + }, + { + "epoch": 8.724074074074075, + "grad_norm": 1.6404742064137752, + "learning_rate": 1.8200969671941685e-07, + "loss": 0.6476, + "step": 4711 + }, + { + "epoch": 8.725925925925926, + "grad_norm": 1.5602978649137849, + "learning_rate": 1.820011157263312e-07, + "loss": 0.6384, + "step": 4712 + }, + { + "epoch": 8.727777777777778, + "grad_norm": 1.6241076608343443, + "learning_rate": 1.8199253288963778e-07, + "loss": 0.6191, + "step": 4713 + }, + { + "epoch": 8.72962962962963, + "grad_norm": 1.5924767719461121, + "learning_rate": 1.819839482095295e-07, + "loss": 0.622, + "step": 4714 + }, + { + "epoch": 8.731481481481481, + "grad_norm": 1.568942150358102, + "learning_rate": 1.8197536168619944e-07, + "loss": 0.6352, + "step": 4715 + }, + { + "epoch": 8.733333333333333, + "grad_norm": 1.5984664339770507, + "learning_rate": 1.819667733198406e-07, + "loss": 0.6301, + "step": 4716 + }, + { + "epoch": 8.735185185185186, + "grad_norm": 1.941038959656708, + "learning_rate": 1.8195818311064603e-07, + "loss": 0.5882, + "step": 4717 + }, + { + "epoch": 8.737037037037037, + "grad_norm": 1.549290053544421, + "learning_rate": 1.8194959105880895e-07, + "loss": 0.6326, + "step": 4718 + }, + { + "epoch": 8.738888888888889, + "grad_norm": 1.3673495020605901, + "learning_rate": 1.8194099716452248e-07, + "loss": 0.6238, + "step": 4719 + }, + { + "epoch": 8.74074074074074, + "grad_norm": 1.3991381699312113, + "learning_rate": 1.8193240142797985e-07, + "loss": 0.6163, + "step": 4720 + }, + { + "epoch": 8.742592592592592, + "grad_norm": 2.000034218585617, + "learning_rate": 1.8192380384937431e-07, + "loss": 0.6271, + "step": 4721 + }, + { + "epoch": 8.744444444444444, + "grad_norm": 1.5598029480816074, + "learning_rate": 1.8191520442889918e-07, + "loss": 0.6042, + "step": 4722 + }, + { + "epoch": 8.746296296296297, + "grad_norm": 1.478153850381044, + "learning_rate": 1.8190660316674773e-07, + "loss": 0.6362, + "step": 4723 + }, + { + "epoch": 8.748148148148148, + "grad_norm": 1.4748774850924247, + "learning_rate": 1.818980000631134e-07, + "loss": 0.6448, + "step": 4724 + }, + { + "epoch": 8.75, + "grad_norm": 1.7754842521417857, + "learning_rate": 1.8188939511818963e-07, + "loss": 0.6413, + "step": 4725 + }, + { + "epoch": 8.751851851851852, + "grad_norm": 1.4849536229583662, + "learning_rate": 1.8188078833216983e-07, + "loss": 0.6098, + "step": 4726 + }, + { + "epoch": 8.753703703703703, + "grad_norm": 1.5398261095887904, + "learning_rate": 1.818721797052475e-07, + "loss": 0.6302, + "step": 4727 + }, + { + "epoch": 8.755555555555556, + "grad_norm": 1.5000368681911562, + "learning_rate": 1.8186356923761626e-07, + "loss": 0.6418, + "step": 4728 + }, + { + "epoch": 8.757407407407408, + "grad_norm": 1.5013206321264958, + "learning_rate": 1.818549569294696e-07, + "loss": 0.6709, + "step": 4729 + }, + { + "epoch": 8.75925925925926, + "grad_norm": 2.6487533430746613, + "learning_rate": 1.8184634278100122e-07, + "loss": 0.6348, + "step": 4730 + }, + { + "epoch": 8.761111111111111, + "grad_norm": 1.4688366605716632, + "learning_rate": 1.8183772679240472e-07, + "loss": 0.6015, + "step": 4731 + }, + { + "epoch": 8.762962962962963, + "grad_norm": 1.2984630671770157, + "learning_rate": 1.818291089638739e-07, + "loss": 0.6261, + "step": 4732 + }, + { + "epoch": 8.764814814814814, + "grad_norm": 1.5969695983099004, + "learning_rate": 1.8182048929560242e-07, + "loss": 0.6475, + "step": 4733 + }, + { + "epoch": 8.766666666666667, + "grad_norm": 1.7820378959978886, + "learning_rate": 1.8181186778778416e-07, + "loss": 0.6313, + "step": 4734 + }, + { + "epoch": 8.768518518518519, + "grad_norm": 1.4622881840933173, + "learning_rate": 1.8180324444061292e-07, + "loss": 0.6233, + "step": 4735 + }, + { + "epoch": 8.77037037037037, + "grad_norm": 1.4612519905828116, + "learning_rate": 1.8179461925428254e-07, + "loss": 0.6488, + "step": 4736 + }, + { + "epoch": 8.772222222222222, + "grad_norm": 1.5163673743922017, + "learning_rate": 1.81785992228987e-07, + "loss": 0.6919, + "step": 4737 + }, + { + "epoch": 8.774074074074074, + "grad_norm": 1.6644282014852592, + "learning_rate": 1.817773633649202e-07, + "loss": 0.6741, + "step": 4738 + }, + { + "epoch": 8.775925925925925, + "grad_norm": 1.4647295128471252, + "learning_rate": 1.817687326622762e-07, + "loss": 0.648, + "step": 4739 + }, + { + "epoch": 8.777777777777779, + "grad_norm": 1.9867233597039433, + "learning_rate": 1.8176010012124895e-07, + "loss": 0.6546, + "step": 4740 + }, + { + "epoch": 8.77962962962963, + "grad_norm": 1.4490874637681297, + "learning_rate": 1.8175146574203264e-07, + "loss": 0.6658, + "step": 4741 + }, + { + "epoch": 8.781481481481482, + "grad_norm": 1.701503399174617, + "learning_rate": 1.8174282952482133e-07, + "loss": 0.6899, + "step": 4742 + }, + { + "epoch": 8.783333333333333, + "grad_norm": 1.5454750252062397, + "learning_rate": 1.8173419146980922e-07, + "loss": 0.6478, + "step": 4743 + }, + { + "epoch": 8.785185185185185, + "grad_norm": 1.5586226965585541, + "learning_rate": 1.817255515771905e-07, + "loss": 0.6438, + "step": 4744 + }, + { + "epoch": 8.787037037037036, + "grad_norm": 1.5067538618749443, + "learning_rate": 1.817169098471594e-07, + "loss": 0.6252, + "step": 4745 + }, + { + "epoch": 8.78888888888889, + "grad_norm": 1.3632548071742268, + "learning_rate": 1.8170826627991023e-07, + "loss": 0.614, + "step": 4746 + }, + { + "epoch": 8.790740740740741, + "grad_norm": 1.5590647594041502, + "learning_rate": 1.816996208756373e-07, + "loss": 0.6555, + "step": 4747 + }, + { + "epoch": 8.792592592592593, + "grad_norm": 1.5168107560316604, + "learning_rate": 1.8169097363453503e-07, + "loss": 0.6203, + "step": 4748 + }, + { + "epoch": 8.794444444444444, + "grad_norm": 1.7194309072778642, + "learning_rate": 1.8168232455679778e-07, + "loss": 0.6388, + "step": 4749 + }, + { + "epoch": 8.796296296296296, + "grad_norm": 1.5391918004481664, + "learning_rate": 1.8167367364262004e-07, + "loss": 0.6447, + "step": 4750 + }, + { + "epoch": 8.798148148148147, + "grad_norm": 1.8542247574495883, + "learning_rate": 1.816650208921963e-07, + "loss": 0.6345, + "step": 4751 + }, + { + "epoch": 8.8, + "grad_norm": 3.1963687310357507, + "learning_rate": 1.8165636630572108e-07, + "loss": 0.6004, + "step": 4752 + }, + { + "epoch": 8.801851851851852, + "grad_norm": 1.9949250538162673, + "learning_rate": 1.8164770988338897e-07, + "loss": 0.662, + "step": 4753 + }, + { + "epoch": 8.803703703703704, + "grad_norm": 1.880751629648909, + "learning_rate": 1.816390516253946e-07, + "loss": 0.6191, + "step": 4754 + }, + { + "epoch": 8.805555555555555, + "grad_norm": 1.7011827412918006, + "learning_rate": 1.8163039153193262e-07, + "loss": 0.6, + "step": 4755 + }, + { + "epoch": 8.807407407407407, + "grad_norm": 1.4842433890205666, + "learning_rate": 1.8162172960319774e-07, + "loss": 0.6321, + "step": 4756 + }, + { + "epoch": 8.809259259259258, + "grad_norm": 1.4332477888658655, + "learning_rate": 1.8161306583938466e-07, + "loss": 0.638, + "step": 4757 + }, + { + "epoch": 8.811111111111112, + "grad_norm": 1.5498554110002054, + "learning_rate": 1.8160440024068823e-07, + "loss": 0.6473, + "step": 4758 + }, + { + "epoch": 8.812962962962963, + "grad_norm": 1.537076131959102, + "learning_rate": 1.8159573280730328e-07, + "loss": 0.6513, + "step": 4759 + }, + { + "epoch": 8.814814814814815, + "grad_norm": 1.6855792368040095, + "learning_rate": 1.815870635394246e-07, + "loss": 0.6044, + "step": 4760 + }, + { + "epoch": 8.816666666666666, + "grad_norm": 1.5394131205208132, + "learning_rate": 1.8157839243724715e-07, + "loss": 0.6375, + "step": 4761 + }, + { + "epoch": 8.818518518518518, + "grad_norm": 1.5437336959923607, + "learning_rate": 1.815697195009659e-07, + "loss": 0.6287, + "step": 4762 + }, + { + "epoch": 8.82037037037037, + "grad_norm": 1.5997878235485652, + "learning_rate": 1.815610447307758e-07, + "loss": 0.6449, + "step": 4763 + }, + { + "epoch": 8.822222222222223, + "grad_norm": 2.0988768306639973, + "learning_rate": 1.815523681268719e-07, + "loss": 0.6324, + "step": 4764 + }, + { + "epoch": 8.824074074074074, + "grad_norm": 2.1354272439864985, + "learning_rate": 1.8154368968944925e-07, + "loss": 0.6275, + "step": 4765 + }, + { + "epoch": 8.825925925925926, + "grad_norm": 1.5347030507272574, + "learning_rate": 1.8153500941870302e-07, + "loss": 0.6508, + "step": 4766 + }, + { + "epoch": 8.827777777777778, + "grad_norm": 1.661853429257589, + "learning_rate": 1.8152632731482833e-07, + "loss": 0.6192, + "step": 4767 + }, + { + "epoch": 8.829629629629629, + "grad_norm": 1.3367538919049962, + "learning_rate": 1.8151764337802032e-07, + "loss": 0.6488, + "step": 4768 + }, + { + "epoch": 8.831481481481482, + "grad_norm": 1.3774919246792268, + "learning_rate": 1.8150895760847437e-07, + "loss": 0.6194, + "step": 4769 + }, + { + "epoch": 8.833333333333334, + "grad_norm": 1.630670597639851, + "learning_rate": 1.8150027000638563e-07, + "loss": 0.6709, + "step": 4770 + }, + { + "epoch": 8.835185185185185, + "grad_norm": 1.4075565066123874, + "learning_rate": 1.814915805719495e-07, + "loss": 0.6062, + "step": 4771 + }, + { + "epoch": 8.837037037037037, + "grad_norm": 1.6812371369625405, + "learning_rate": 1.8148288930536129e-07, + "loss": 0.6346, + "step": 4772 + }, + { + "epoch": 8.838888888888889, + "grad_norm": 1.8649683367785135, + "learning_rate": 1.8147419620681643e-07, + "loss": 0.6228, + "step": 4773 + }, + { + "epoch": 8.84074074074074, + "grad_norm": 1.4885054629074472, + "learning_rate": 1.8146550127651037e-07, + "loss": 0.6511, + "step": 4774 + }, + { + "epoch": 8.842592592592592, + "grad_norm": 1.6236803099433943, + "learning_rate": 1.8145680451463857e-07, + "loss": 0.6328, + "step": 4775 + }, + { + "epoch": 8.844444444444445, + "grad_norm": 1.940368928585568, + "learning_rate": 1.8144810592139655e-07, + "loss": 0.6136, + "step": 4776 + }, + { + "epoch": 8.846296296296297, + "grad_norm": 1.4758065614773443, + "learning_rate": 1.8143940549697993e-07, + "loss": 0.6577, + "step": 4777 + }, + { + "epoch": 8.848148148148148, + "grad_norm": 1.4390984157305504, + "learning_rate": 1.8143070324158427e-07, + "loss": 0.665, + "step": 4778 + }, + { + "epoch": 8.85, + "grad_norm": 1.7184382133170344, + "learning_rate": 1.8142199915540526e-07, + "loss": 0.6602, + "step": 4779 + }, + { + "epoch": 8.851851851851851, + "grad_norm": 1.6384462288972075, + "learning_rate": 1.8141329323863858e-07, + "loss": 0.661, + "step": 4780 + }, + { + "epoch": 8.853703703703705, + "grad_norm": 2.314472491481749, + "learning_rate": 1.814045854914799e-07, + "loss": 0.627, + "step": 4781 + }, + { + "epoch": 8.855555555555556, + "grad_norm": 1.5186661668525334, + "learning_rate": 1.813958759141251e-07, + "loss": 0.615, + "step": 4782 + }, + { + "epoch": 8.857407407407408, + "grad_norm": 1.4791592421566457, + "learning_rate": 1.813871645067699e-07, + "loss": 0.6223, + "step": 4783 + }, + { + "epoch": 8.85925925925926, + "grad_norm": 1.5437100243956041, + "learning_rate": 1.813784512696102e-07, + "loss": 0.6025, + "step": 4784 + }, + { + "epoch": 8.86111111111111, + "grad_norm": 1.643807220976132, + "learning_rate": 1.8136973620284193e-07, + "loss": 0.666, + "step": 4785 + }, + { + "epoch": 8.862962962962962, + "grad_norm": 1.6739861880512772, + "learning_rate": 1.81361019306661e-07, + "loss": 0.6507, + "step": 4786 + }, + { + "epoch": 8.864814814814816, + "grad_norm": 1.5952673058975069, + "learning_rate": 1.8135230058126333e-07, + "loss": 0.6434, + "step": 4787 + }, + { + "epoch": 8.866666666666667, + "grad_norm": 1.6437273328889874, + "learning_rate": 1.81343580026845e-07, + "loss": 0.6612, + "step": 4788 + }, + { + "epoch": 8.868518518518519, + "grad_norm": 1.622596914954736, + "learning_rate": 1.813348576436021e-07, + "loss": 0.6538, + "step": 4789 + }, + { + "epoch": 8.87037037037037, + "grad_norm": 1.3060592873421302, + "learning_rate": 1.8132613343173068e-07, + "loss": 0.6221, + "step": 4790 + }, + { + "epoch": 8.872222222222222, + "grad_norm": 1.5810597563113369, + "learning_rate": 1.8131740739142692e-07, + "loss": 0.6429, + "step": 4791 + }, + { + "epoch": 8.874074074074073, + "grad_norm": 1.6395330414936504, + "learning_rate": 1.8130867952288696e-07, + "loss": 0.5895, + "step": 4792 + }, + { + "epoch": 8.875925925925927, + "grad_norm": 1.5921565627648888, + "learning_rate": 1.8129994982630707e-07, + "loss": 0.6883, + "step": 4793 + }, + { + "epoch": 8.877777777777778, + "grad_norm": 1.5861378481811457, + "learning_rate": 1.8129121830188352e-07, + "loss": 0.6483, + "step": 4794 + }, + { + "epoch": 8.87962962962963, + "grad_norm": 2.0366477235989655, + "learning_rate": 1.8128248494981256e-07, + "loss": 0.6239, + "step": 4795 + }, + { + "epoch": 8.881481481481481, + "grad_norm": 1.4555855106498405, + "learning_rate": 1.8127374977029063e-07, + "loss": 0.6278, + "step": 4796 + }, + { + "epoch": 8.883333333333333, + "grad_norm": 1.428021416975343, + "learning_rate": 1.8126501276351404e-07, + "loss": 0.6176, + "step": 4797 + }, + { + "epoch": 8.885185185185184, + "grad_norm": 1.4028919477392796, + "learning_rate": 1.8125627392967923e-07, + "loss": 0.675, + "step": 4798 + }, + { + "epoch": 8.887037037037038, + "grad_norm": 1.6754221304722121, + "learning_rate": 1.812475332689827e-07, + "loss": 0.6281, + "step": 4799 + }, + { + "epoch": 8.88888888888889, + "grad_norm": 1.7550058280564609, + "learning_rate": 1.8123879078162097e-07, + "loss": 0.6626, + "step": 4800 + }, + { + "epoch": 8.89074074074074, + "grad_norm": 1.615829983650522, + "learning_rate": 1.812300464677906e-07, + "loss": 0.6121, + "step": 4801 + }, + { + "epoch": 8.892592592592592, + "grad_norm": 1.7515681734317055, + "learning_rate": 1.8122130032768812e-07, + "loss": 0.6196, + "step": 4802 + }, + { + "epoch": 8.894444444444444, + "grad_norm": 1.4298392742036237, + "learning_rate": 1.8121255236151023e-07, + "loss": 0.6614, + "step": 4803 + }, + { + "epoch": 8.896296296296295, + "grad_norm": 1.5809941412851993, + "learning_rate": 1.812038025694536e-07, + "loss": 0.6592, + "step": 4804 + }, + { + "epoch": 8.898148148148149, + "grad_norm": 1.514028225205355, + "learning_rate": 1.8119505095171497e-07, + "loss": 0.6342, + "step": 4805 + }, + { + "epoch": 8.9, + "grad_norm": 1.4341172218221372, + "learning_rate": 1.8118629750849104e-07, + "loss": 0.6172, + "step": 4806 + }, + { + "epoch": 8.901851851851852, + "grad_norm": 1.525011857098194, + "learning_rate": 1.8117754223997866e-07, + "loss": 0.6262, + "step": 4807 + }, + { + "epoch": 8.903703703703703, + "grad_norm": 1.6771318069019463, + "learning_rate": 1.8116878514637464e-07, + "loss": 0.6525, + "step": 4808 + }, + { + "epoch": 8.905555555555555, + "grad_norm": 1.7982890756177687, + "learning_rate": 1.811600262278759e-07, + "loss": 0.6123, + "step": 4809 + }, + { + "epoch": 8.907407407407408, + "grad_norm": 1.5055494196298373, + "learning_rate": 1.811512654846793e-07, + "loss": 0.6593, + "step": 4810 + }, + { + "epoch": 8.90925925925926, + "grad_norm": 1.7677444419090338, + "learning_rate": 1.8114250291698187e-07, + "loss": 0.6361, + "step": 4811 + }, + { + "epoch": 8.911111111111111, + "grad_norm": 1.3275828657322852, + "learning_rate": 1.811337385249806e-07, + "loss": 0.6298, + "step": 4812 + }, + { + "epoch": 8.912962962962963, + "grad_norm": 1.5191138290104935, + "learning_rate": 1.8112497230887256e-07, + "loss": 0.6709, + "step": 4813 + }, + { + "epoch": 8.914814814814815, + "grad_norm": 1.7926323308926118, + "learning_rate": 1.8111620426885476e-07, + "loss": 0.6325, + "step": 4814 + }, + { + "epoch": 8.916666666666666, + "grad_norm": 1.4347454040040613, + "learning_rate": 1.8110743440512444e-07, + "loss": 0.6121, + "step": 4815 + }, + { + "epoch": 8.918518518518518, + "grad_norm": 1.4312511052912815, + "learning_rate": 1.8109866271787865e-07, + "loss": 0.6374, + "step": 4816 + }, + { + "epoch": 8.920370370370371, + "grad_norm": 1.4237599959113545, + "learning_rate": 1.8108988920731469e-07, + "loss": 0.6318, + "step": 4817 + }, + { + "epoch": 8.922222222222222, + "grad_norm": 1.7623743746650833, + "learning_rate": 1.810811138736298e-07, + "loss": 0.6133, + "step": 4818 + }, + { + "epoch": 8.924074074074074, + "grad_norm": 1.7303734268128559, + "learning_rate": 1.8107233671702122e-07, + "loss": 0.6009, + "step": 4819 + }, + { + "epoch": 8.925925925925926, + "grad_norm": 2.0789529512237945, + "learning_rate": 1.8106355773768636e-07, + "loss": 0.6569, + "step": 4820 + }, + { + "epoch": 8.927777777777777, + "grad_norm": 1.4620603123364848, + "learning_rate": 1.8105477693582252e-07, + "loss": 0.6797, + "step": 4821 + }, + { + "epoch": 8.92962962962963, + "grad_norm": 1.5127297967909374, + "learning_rate": 1.8104599431162715e-07, + "loss": 0.6371, + "step": 4822 + }, + { + "epoch": 8.931481481481482, + "grad_norm": 2.3423121292169826, + "learning_rate": 1.8103720986529773e-07, + "loss": 0.6451, + "step": 4823 + }, + { + "epoch": 8.933333333333334, + "grad_norm": 1.5080718483619968, + "learning_rate": 1.8102842359703175e-07, + "loss": 0.6062, + "step": 4824 + }, + { + "epoch": 8.935185185185185, + "grad_norm": 1.560177617698358, + "learning_rate": 1.8101963550702674e-07, + "loss": 0.6392, + "step": 4825 + }, + { + "epoch": 8.937037037037037, + "grad_norm": 1.3734722786351372, + "learning_rate": 1.8101084559548023e-07, + "loss": 0.6425, + "step": 4826 + }, + { + "epoch": 8.938888888888888, + "grad_norm": 1.5953822169552891, + "learning_rate": 1.8100205386258995e-07, + "loss": 0.6275, + "step": 4827 + }, + { + "epoch": 8.940740740740742, + "grad_norm": 1.6082485311037924, + "learning_rate": 1.8099326030855346e-07, + "loss": 0.6271, + "step": 4828 + }, + { + "epoch": 8.942592592592593, + "grad_norm": 1.4954702487737823, + "learning_rate": 1.8098446493356852e-07, + "loss": 0.6376, + "step": 4829 + }, + { + "epoch": 8.944444444444445, + "grad_norm": 1.6226568298952204, + "learning_rate": 1.8097566773783285e-07, + "loss": 0.5922, + "step": 4830 + }, + { + "epoch": 8.946296296296296, + "grad_norm": 1.922238094573983, + "learning_rate": 1.8096686872154425e-07, + "loss": 0.6186, + "step": 4831 + }, + { + "epoch": 8.948148148148148, + "grad_norm": 1.5965934037450158, + "learning_rate": 1.8095806788490054e-07, + "loss": 0.6601, + "step": 4832 + }, + { + "epoch": 8.95, + "grad_norm": 2.3586461110550383, + "learning_rate": 1.8094926522809956e-07, + "loss": 0.6373, + "step": 4833 + }, + { + "epoch": 8.951851851851853, + "grad_norm": 1.503350121104173, + "learning_rate": 1.8094046075133927e-07, + "loss": 0.6279, + "step": 4834 + }, + { + "epoch": 8.953703703703704, + "grad_norm": 1.5018749230431285, + "learning_rate": 1.809316544548176e-07, + "loss": 0.6511, + "step": 4835 + }, + { + "epoch": 8.955555555555556, + "grad_norm": 1.5233438064006162, + "learning_rate": 1.8092284633873248e-07, + "loss": 0.6484, + "step": 4836 + }, + { + "epoch": 8.957407407407407, + "grad_norm": 1.475330476269923, + "learning_rate": 1.8091403640328205e-07, + "loss": 0.6559, + "step": 4837 + }, + { + "epoch": 8.959259259259259, + "grad_norm": 2.1607204789682855, + "learning_rate": 1.809052246486643e-07, + "loss": 0.6151, + "step": 4838 + }, + { + "epoch": 8.96111111111111, + "grad_norm": 1.5687888980311786, + "learning_rate": 1.808964110750774e-07, + "loss": 0.6497, + "step": 4839 + }, + { + "epoch": 8.962962962962964, + "grad_norm": 1.3948814041499955, + "learning_rate": 1.808875956827194e-07, + "loss": 0.6475, + "step": 4840 + }, + { + "epoch": 8.964814814814815, + "grad_norm": 1.519589769064513, + "learning_rate": 1.808787784717886e-07, + "loss": 0.6629, + "step": 4841 + }, + { + "epoch": 8.966666666666667, + "grad_norm": 2.110036887270788, + "learning_rate": 1.8086995944248317e-07, + "loss": 0.6949, + "step": 4842 + }, + { + "epoch": 8.968518518518518, + "grad_norm": 1.691389530965149, + "learning_rate": 1.8086113859500146e-07, + "loss": 0.6476, + "step": 4843 + }, + { + "epoch": 8.97037037037037, + "grad_norm": 1.6704512669708718, + "learning_rate": 1.808523159295417e-07, + "loss": 0.6792, + "step": 4844 + }, + { + "epoch": 8.972222222222221, + "grad_norm": 1.4432593127391118, + "learning_rate": 1.8084349144630228e-07, + "loss": 0.6008, + "step": 4845 + }, + { + "epoch": 8.974074074074075, + "grad_norm": 1.6026957206222825, + "learning_rate": 1.8083466514548163e-07, + "loss": 0.6528, + "step": 4846 + }, + { + "epoch": 8.975925925925926, + "grad_norm": 1.7981805280056289, + "learning_rate": 1.8082583702727817e-07, + "loss": 0.6343, + "step": 4847 + }, + { + "epoch": 8.977777777777778, + "grad_norm": 1.9680978528804036, + "learning_rate": 1.8081700709189037e-07, + "loss": 0.6557, + "step": 4848 + }, + { + "epoch": 8.97962962962963, + "grad_norm": 1.6198450599379675, + "learning_rate": 1.808081753395167e-07, + "loss": 0.6605, + "step": 4849 + }, + { + "epoch": 8.981481481481481, + "grad_norm": 1.8288338170075198, + "learning_rate": 1.8079934177035584e-07, + "loss": 0.6544, + "step": 4850 + }, + { + "epoch": 8.983333333333333, + "grad_norm": 1.5415661748127376, + "learning_rate": 1.8079050638460632e-07, + "loss": 0.6511, + "step": 4851 + }, + { + "epoch": 8.985185185185186, + "grad_norm": 1.429545976687254, + "learning_rate": 1.8078166918246678e-07, + "loss": 0.6494, + "step": 4852 + }, + { + "epoch": 8.987037037037037, + "grad_norm": 1.773454452391686, + "learning_rate": 1.8077283016413592e-07, + "loss": 0.6387, + "step": 4853 + }, + { + "epoch": 8.988888888888889, + "grad_norm": 1.8215141516909366, + "learning_rate": 1.8076398932981248e-07, + "loss": 0.6013, + "step": 4854 + }, + { + "epoch": 8.99074074074074, + "grad_norm": 1.6802766249658658, + "learning_rate": 1.807551466796952e-07, + "loss": 0.6298, + "step": 4855 + }, + { + "epoch": 8.992592592592592, + "grad_norm": 2.727235777158391, + "learning_rate": 1.8074630221398287e-07, + "loss": 0.6386, + "step": 4856 + }, + { + "epoch": 8.994444444444444, + "grad_norm": 1.702439565294306, + "learning_rate": 1.8073745593287438e-07, + "loss": 0.6381, + "step": 4857 + }, + { + "epoch": 8.996296296296297, + "grad_norm": 1.8383892174425474, + "learning_rate": 1.8072860783656858e-07, + "loss": 0.6769, + "step": 4858 + }, + { + "epoch": 8.998148148148148, + "grad_norm": 1.618941278882836, + "learning_rate": 1.8071975792526445e-07, + "loss": 0.6035, + "step": 4859 + }, + { + "epoch": 9.0, + "grad_norm": 1.5401358476940168, + "learning_rate": 1.807109061991609e-07, + "loss": 0.6431, + "step": 4860 + }, + { + "epoch": 9.001851851851852, + "grad_norm": 1.670582670894463, + "learning_rate": 1.8070205265845703e-07, + "loss": 0.629, + "step": 4861 + }, + { + "epoch": 9.003703703703703, + "grad_norm": 1.5090899879571136, + "learning_rate": 1.8069319730335176e-07, + "loss": 0.5943, + "step": 4862 + }, + { + "epoch": 9.005555555555556, + "grad_norm": 1.6877111795180013, + "learning_rate": 1.806843401340443e-07, + "loss": 0.628, + "step": 4863 + }, + { + "epoch": 9.007407407407408, + "grad_norm": 1.5596527361672607, + "learning_rate": 1.8067548115073374e-07, + "loss": 0.6689, + "step": 4864 + }, + { + "epoch": 9.00925925925926, + "grad_norm": 2.143374943886906, + "learning_rate": 1.8066662035361925e-07, + "loss": 0.6323, + "step": 4865 + }, + { + "epoch": 9.011111111111111, + "grad_norm": 1.5731065339124362, + "learning_rate": 1.806577577429e-07, + "loss": 0.6204, + "step": 4866 + }, + { + "epoch": 9.012962962962963, + "grad_norm": 1.732070680461305, + "learning_rate": 1.8064889331877534e-07, + "loss": 0.6269, + "step": 4867 + }, + { + "epoch": 9.014814814814814, + "grad_norm": 1.577489965214457, + "learning_rate": 1.8064002708144452e-07, + "loss": 0.6388, + "step": 4868 + }, + { + "epoch": 9.016666666666667, + "grad_norm": 1.492399754931415, + "learning_rate": 1.8063115903110688e-07, + "loss": 0.6491, + "step": 4869 + }, + { + "epoch": 9.018518518518519, + "grad_norm": 1.3124732044189318, + "learning_rate": 1.8062228916796177e-07, + "loss": 0.6721, + "step": 4870 + }, + { + "epoch": 9.02037037037037, + "grad_norm": 1.7090493040093824, + "learning_rate": 1.8061341749220863e-07, + "loss": 0.6171, + "step": 4871 + }, + { + "epoch": 9.022222222222222, + "grad_norm": 1.8996995053563839, + "learning_rate": 1.8060454400404695e-07, + "loss": 0.6251, + "step": 4872 + }, + { + "epoch": 9.024074074074074, + "grad_norm": 1.4563608758878375, + "learning_rate": 1.8059566870367617e-07, + "loss": 0.6134, + "step": 4873 + }, + { + "epoch": 9.025925925925925, + "grad_norm": 1.5721457100028966, + "learning_rate": 1.805867915912959e-07, + "loss": 0.5932, + "step": 4874 + }, + { + "epoch": 9.027777777777779, + "grad_norm": 1.286492788291641, + "learning_rate": 1.8057791266710564e-07, + "loss": 0.6048, + "step": 4875 + }, + { + "epoch": 9.02962962962963, + "grad_norm": 1.8518693533491049, + "learning_rate": 1.805690319313051e-07, + "loss": 0.6124, + "step": 4876 + }, + { + "epoch": 9.031481481481482, + "grad_norm": 1.6898112425993053, + "learning_rate": 1.805601493840939e-07, + "loss": 0.6081, + "step": 4877 + }, + { + "epoch": 9.033333333333333, + "grad_norm": 1.5024791483925741, + "learning_rate": 1.805512650256717e-07, + "loss": 0.6136, + "step": 4878 + }, + { + "epoch": 9.035185185185185, + "grad_norm": 1.683033488553737, + "learning_rate": 1.8054237885623831e-07, + "loss": 0.6685, + "step": 4879 + }, + { + "epoch": 9.037037037037036, + "grad_norm": 1.7062090724179861, + "learning_rate": 1.805334908759935e-07, + "loss": 0.6563, + "step": 4880 + }, + { + "epoch": 9.03888888888889, + "grad_norm": 1.5285265723010917, + "learning_rate": 1.805246010851371e-07, + "loss": 0.6713, + "step": 4881 + }, + { + "epoch": 9.040740740740741, + "grad_norm": 1.5768796010668769, + "learning_rate": 1.8051570948386895e-07, + "loss": 0.6405, + "step": 4882 + }, + { + "epoch": 9.042592592592593, + "grad_norm": 1.6988096391799035, + "learning_rate": 1.8050681607238895e-07, + "loss": 0.6159, + "step": 4883 + }, + { + "epoch": 9.044444444444444, + "grad_norm": 1.5203602139755228, + "learning_rate": 1.8049792085089712e-07, + "loss": 0.6489, + "step": 4884 + }, + { + "epoch": 9.046296296296296, + "grad_norm": 1.6562165501659378, + "learning_rate": 1.804890238195934e-07, + "loss": 0.6007, + "step": 4885 + }, + { + "epoch": 9.048148148148147, + "grad_norm": 1.6500159216629178, + "learning_rate": 1.8048012497867776e-07, + "loss": 0.6326, + "step": 4886 + }, + { + "epoch": 9.05, + "grad_norm": 1.7431341341622981, + "learning_rate": 1.8047122432835038e-07, + "loss": 0.6403, + "step": 4887 + }, + { + "epoch": 9.051851851851852, + "grad_norm": 1.5421787446841935, + "learning_rate": 1.8046232186881132e-07, + "loss": 0.6467, + "step": 4888 + }, + { + "epoch": 9.053703703703704, + "grad_norm": 1.7387780891120124, + "learning_rate": 1.804534176002607e-07, + "loss": 0.6492, + "step": 4889 + }, + { + "epoch": 9.055555555555555, + "grad_norm": 1.421758201398186, + "learning_rate": 1.8044451152289877e-07, + "loss": 0.6306, + "step": 4890 + }, + { + "epoch": 9.057407407407407, + "grad_norm": 1.6624817408623478, + "learning_rate": 1.8043560363692574e-07, + "loss": 0.6303, + "step": 4891 + }, + { + "epoch": 9.059259259259258, + "grad_norm": 1.538891195740592, + "learning_rate": 1.804266939425419e-07, + "loss": 0.6219, + "step": 4892 + }, + { + "epoch": 9.061111111111112, + "grad_norm": 1.9771516523804664, + "learning_rate": 1.804177824399475e-07, + "loss": 0.6916, + "step": 4893 + }, + { + "epoch": 9.062962962962963, + "grad_norm": 1.6812721833699744, + "learning_rate": 1.8040886912934292e-07, + "loss": 0.6621, + "step": 4894 + }, + { + "epoch": 9.064814814814815, + "grad_norm": 1.431847059574879, + "learning_rate": 1.8039995401092862e-07, + "loss": 0.6211, + "step": 4895 + }, + { + "epoch": 9.066666666666666, + "grad_norm": 1.7768114597845408, + "learning_rate": 1.80391037084905e-07, + "loss": 0.6569, + "step": 4896 + }, + { + "epoch": 9.068518518518518, + "grad_norm": 1.6598328035460053, + "learning_rate": 1.803821183514725e-07, + "loss": 0.6667, + "step": 4897 + }, + { + "epoch": 9.07037037037037, + "grad_norm": 1.592453738061996, + "learning_rate": 1.8037319781083166e-07, + "loss": 0.6268, + "step": 4898 + }, + { + "epoch": 9.072222222222223, + "grad_norm": 1.7487959339946386, + "learning_rate": 1.8036427546318306e-07, + "loss": 0.6231, + "step": 4899 + }, + { + "epoch": 9.074074074074074, + "grad_norm": 1.6713807234040476, + "learning_rate": 1.803553513087273e-07, + "loss": 0.6492, + "step": 4900 + }, + { + "epoch": 9.075925925925926, + "grad_norm": 1.4033205965529207, + "learning_rate": 1.80346425347665e-07, + "loss": 0.659, + "step": 4901 + }, + { + "epoch": 9.077777777777778, + "grad_norm": 1.8068211291080067, + "learning_rate": 1.8033749758019678e-07, + "loss": 0.6495, + "step": 4902 + }, + { + "epoch": 9.079629629629629, + "grad_norm": 1.5516555526865257, + "learning_rate": 1.8032856800652344e-07, + "loss": 0.6396, + "step": 4903 + }, + { + "epoch": 9.081481481481482, + "grad_norm": 1.6807042891880808, + "learning_rate": 1.8031963662684577e-07, + "loss": 0.6317, + "step": 4904 + }, + { + "epoch": 9.083333333333334, + "grad_norm": 1.522403961214487, + "learning_rate": 1.803107034413645e-07, + "loss": 0.6484, + "step": 4905 + }, + { + "epoch": 9.085185185185185, + "grad_norm": 1.8348439508929184, + "learning_rate": 1.803017684502805e-07, + "loss": 0.7465, + "step": 4906 + }, + { + "epoch": 9.087037037037037, + "grad_norm": 1.8591497885796593, + "learning_rate": 1.8029283165379464e-07, + "loss": 0.6329, + "step": 4907 + }, + { + "epoch": 9.088888888888889, + "grad_norm": 1.6794104133695151, + "learning_rate": 1.8028389305210784e-07, + "loss": 0.623, + "step": 4908 + }, + { + "epoch": 9.09074074074074, + "grad_norm": 1.534546021159242, + "learning_rate": 1.802749526454211e-07, + "loss": 0.646, + "step": 4909 + }, + { + "epoch": 9.092592592592593, + "grad_norm": 2.2076579395792537, + "learning_rate": 1.802660104339354e-07, + "loss": 0.6616, + "step": 4910 + }, + { + "epoch": 9.094444444444445, + "grad_norm": 1.735066846237029, + "learning_rate": 1.8025706641785175e-07, + "loss": 0.6561, + "step": 4911 + }, + { + "epoch": 9.096296296296297, + "grad_norm": 1.7863350317501159, + "learning_rate": 1.802481205973713e-07, + "loss": 0.6028, + "step": 4912 + }, + { + "epoch": 9.098148148148148, + "grad_norm": 1.4377673643273938, + "learning_rate": 1.8023917297269517e-07, + "loss": 0.6386, + "step": 4913 + }, + { + "epoch": 9.1, + "grad_norm": 1.3457356985117823, + "learning_rate": 1.802302235440245e-07, + "loss": 0.6775, + "step": 4914 + }, + { + "epoch": 9.101851851851851, + "grad_norm": 2.089901384560452, + "learning_rate": 1.8022127231156047e-07, + "loss": 0.6159, + "step": 4915 + }, + { + "epoch": 9.103703703703705, + "grad_norm": 1.7794981059506234, + "learning_rate": 1.8021231927550437e-07, + "loss": 0.6239, + "step": 4916 + }, + { + "epoch": 9.105555555555556, + "grad_norm": 1.4997174730793863, + "learning_rate": 1.8020336443605747e-07, + "loss": 0.6546, + "step": 4917 + }, + { + "epoch": 9.107407407407408, + "grad_norm": 1.6246675824622723, + "learning_rate": 1.8019440779342113e-07, + "loss": 0.6521, + "step": 4918 + }, + { + "epoch": 9.10925925925926, + "grad_norm": 1.6028425939016664, + "learning_rate": 1.8018544934779671e-07, + "loss": 0.6563, + "step": 4919 + }, + { + "epoch": 9.11111111111111, + "grad_norm": 3.1309177845584384, + "learning_rate": 1.801764890993856e-07, + "loss": 0.6563, + "step": 4920 + }, + { + "epoch": 9.112962962962962, + "grad_norm": 1.7354720790564453, + "learning_rate": 1.8016752704838925e-07, + "loss": 0.6146, + "step": 4921 + }, + { + "epoch": 9.114814814814816, + "grad_norm": 1.5981173872314567, + "learning_rate": 1.8015856319500916e-07, + "loss": 0.6294, + "step": 4922 + }, + { + "epoch": 9.116666666666667, + "grad_norm": 1.6175541542904597, + "learning_rate": 1.8014959753944686e-07, + "loss": 0.6288, + "step": 4923 + }, + { + "epoch": 9.118518518518519, + "grad_norm": 1.466344210031735, + "learning_rate": 1.801406300819039e-07, + "loss": 0.5894, + "step": 4924 + }, + { + "epoch": 9.12037037037037, + "grad_norm": 2.490831660922919, + "learning_rate": 1.8013166082258196e-07, + "loss": 0.6833, + "step": 4925 + }, + { + "epoch": 9.122222222222222, + "grad_norm": 1.5062822291987519, + "learning_rate": 1.8012268976168264e-07, + "loss": 0.6362, + "step": 4926 + }, + { + "epoch": 9.124074074074073, + "grad_norm": 1.3442409120015848, + "learning_rate": 1.801137168994076e-07, + "loss": 0.5869, + "step": 4927 + }, + { + "epoch": 9.125925925925927, + "grad_norm": 1.5412123091510965, + "learning_rate": 1.8010474223595868e-07, + "loss": 0.6555, + "step": 4928 + }, + { + "epoch": 9.127777777777778, + "grad_norm": 1.3476337243530934, + "learning_rate": 1.8009576577153758e-07, + "loss": 0.6275, + "step": 4929 + }, + { + "epoch": 9.12962962962963, + "grad_norm": 1.840453833449504, + "learning_rate": 1.8008678750634614e-07, + "loss": 0.6202, + "step": 4930 + }, + { + "epoch": 9.131481481481481, + "grad_norm": 1.5198333023871702, + "learning_rate": 1.8007780744058617e-07, + "loss": 0.6419, + "step": 4931 + }, + { + "epoch": 9.133333333333333, + "grad_norm": 1.5909486322479607, + "learning_rate": 1.8006882557445962e-07, + "loss": 0.6524, + "step": 4932 + }, + { + "epoch": 9.135185185185184, + "grad_norm": 1.7759503810532487, + "learning_rate": 1.800598419081684e-07, + "loss": 0.6276, + "step": 4933 + }, + { + "epoch": 9.137037037037038, + "grad_norm": 1.6247030790685282, + "learning_rate": 1.800508564419145e-07, + "loss": 0.6128, + "step": 4934 + }, + { + "epoch": 9.13888888888889, + "grad_norm": 1.5383025035594717, + "learning_rate": 1.800418691758999e-07, + "loss": 0.6687, + "step": 4935 + }, + { + "epoch": 9.14074074074074, + "grad_norm": 1.8208035043933768, + "learning_rate": 1.8003288011032675e-07, + "loss": 0.645, + "step": 4936 + }, + { + "epoch": 9.142592592592592, + "grad_norm": 1.7065749632157245, + "learning_rate": 1.8002388924539703e-07, + "loss": 0.6567, + "step": 4937 + }, + { + "epoch": 9.144444444444444, + "grad_norm": 1.357175243788134, + "learning_rate": 1.80014896581313e-07, + "loss": 0.5964, + "step": 4938 + }, + { + "epoch": 9.146296296296295, + "grad_norm": 1.4591828467543584, + "learning_rate": 1.8000590211827674e-07, + "loss": 0.6301, + "step": 4939 + }, + { + "epoch": 9.148148148148149, + "grad_norm": 1.4515861833320292, + "learning_rate": 1.799969058564905e-07, + "loss": 0.6093, + "step": 4940 + }, + { + "epoch": 9.15, + "grad_norm": 1.401304633813347, + "learning_rate": 1.799879077961566e-07, + "loss": 0.5998, + "step": 4941 + }, + { + "epoch": 9.151851851851852, + "grad_norm": 1.631327827543034, + "learning_rate": 1.799789079374772e-07, + "loss": 0.646, + "step": 4942 + }, + { + "epoch": 9.153703703703703, + "grad_norm": 1.3837585331254347, + "learning_rate": 1.7996990628065478e-07, + "loss": 0.6236, + "step": 4943 + }, + { + "epoch": 9.155555555555555, + "grad_norm": 1.4647990563982192, + "learning_rate": 1.7996090282589166e-07, + "loss": 0.6437, + "step": 4944 + }, + { + "epoch": 9.157407407407407, + "grad_norm": 1.8412269214564083, + "learning_rate": 1.7995189757339032e-07, + "loss": 0.6281, + "step": 4945 + }, + { + "epoch": 9.15925925925926, + "grad_norm": 1.4354687122024636, + "learning_rate": 1.799428905233531e-07, + "loss": 0.6727, + "step": 4946 + }, + { + "epoch": 9.161111111111111, + "grad_norm": 2.645241182905206, + "learning_rate": 1.7993388167598266e-07, + "loss": 0.6025, + "step": 4947 + }, + { + "epoch": 9.162962962962963, + "grad_norm": 1.5911301601033796, + "learning_rate": 1.799248710314814e-07, + "loss": 0.6279, + "step": 4948 + }, + { + "epoch": 9.164814814814815, + "grad_norm": 1.4966637795374393, + "learning_rate": 1.7991585859005201e-07, + "loss": 0.6311, + "step": 4949 + }, + { + "epoch": 9.166666666666666, + "grad_norm": 1.5328918698937555, + "learning_rate": 1.7990684435189705e-07, + "loss": 0.6354, + "step": 4950 + }, + { + "epoch": 9.168518518518518, + "grad_norm": 1.5665955349191336, + "learning_rate": 1.798978283172192e-07, + "loss": 0.6376, + "step": 4951 + }, + { + "epoch": 9.170370370370371, + "grad_norm": 1.5761449533059166, + "learning_rate": 1.798888104862212e-07, + "loss": 0.6243, + "step": 4952 + }, + { + "epoch": 9.172222222222222, + "grad_norm": 1.6093057563831608, + "learning_rate": 1.798797908591057e-07, + "loss": 0.6309, + "step": 4953 + }, + { + "epoch": 9.174074074074074, + "grad_norm": 1.668206834831381, + "learning_rate": 1.798707694360756e-07, + "loss": 0.6858, + "step": 4954 + }, + { + "epoch": 9.175925925925926, + "grad_norm": 1.6301226639602802, + "learning_rate": 1.798617462173337e-07, + "loss": 0.632, + "step": 4955 + }, + { + "epoch": 9.177777777777777, + "grad_norm": 1.35186316364307, + "learning_rate": 1.798527212030828e-07, + "loss": 0.6525, + "step": 4956 + }, + { + "epoch": 9.17962962962963, + "grad_norm": 1.8015821596486283, + "learning_rate": 1.7984369439352585e-07, + "loss": 0.6318, + "step": 4957 + }, + { + "epoch": 9.181481481481482, + "grad_norm": 1.6094478716523994, + "learning_rate": 1.7983466578886582e-07, + "loss": 0.6479, + "step": 4958 + }, + { + "epoch": 9.183333333333334, + "grad_norm": 1.5646830130068923, + "learning_rate": 1.798256353893057e-07, + "loss": 0.6496, + "step": 4959 + }, + { + "epoch": 9.185185185185185, + "grad_norm": 1.5096535566202443, + "learning_rate": 1.7981660319504844e-07, + "loss": 0.63, + "step": 4960 + }, + { + "epoch": 9.187037037037037, + "grad_norm": 1.484454742990276, + "learning_rate": 1.7980756920629717e-07, + "loss": 0.6213, + "step": 4961 + }, + { + "epoch": 9.188888888888888, + "grad_norm": 1.9892772178648397, + "learning_rate": 1.79798533423255e-07, + "loss": 0.6571, + "step": 4962 + }, + { + "epoch": 9.190740740740742, + "grad_norm": 1.7467686499634199, + "learning_rate": 1.797894958461251e-07, + "loss": 0.6355, + "step": 4963 + }, + { + "epoch": 9.192592592592593, + "grad_norm": 1.7687282721631068, + "learning_rate": 1.797804564751106e-07, + "loss": 0.635, + "step": 4964 + }, + { + "epoch": 9.194444444444445, + "grad_norm": 1.4851070830982465, + "learning_rate": 1.7977141531041476e-07, + "loss": 0.6618, + "step": 4965 + }, + { + "epoch": 9.196296296296296, + "grad_norm": 1.615517358584911, + "learning_rate": 1.7976237235224084e-07, + "loss": 0.6363, + "step": 4966 + }, + { + "epoch": 9.198148148148148, + "grad_norm": 2.4971452816804574, + "learning_rate": 1.7975332760079216e-07, + "loss": 0.6355, + "step": 4967 + }, + { + "epoch": 9.2, + "grad_norm": 1.4223137090744706, + "learning_rate": 1.7974428105627206e-07, + "loss": 0.6648, + "step": 4968 + }, + { + "epoch": 9.201851851851853, + "grad_norm": 1.8314652308840857, + "learning_rate": 1.7973523271888397e-07, + "loss": 0.6548, + "step": 4969 + }, + { + "epoch": 9.203703703703704, + "grad_norm": 1.4900631781858213, + "learning_rate": 1.797261825888313e-07, + "loss": 0.6516, + "step": 4970 + }, + { + "epoch": 9.205555555555556, + "grad_norm": 1.699847563799028, + "learning_rate": 1.7971713066631745e-07, + "loss": 0.6544, + "step": 4971 + }, + { + "epoch": 9.207407407407407, + "grad_norm": 1.4835502298931005, + "learning_rate": 1.7970807695154603e-07, + "loss": 0.5828, + "step": 4972 + }, + { + "epoch": 9.209259259259259, + "grad_norm": 1.5407325633124669, + "learning_rate": 1.796990214447206e-07, + "loss": 0.6799, + "step": 4973 + }, + { + "epoch": 9.21111111111111, + "grad_norm": 1.4335740288818684, + "learning_rate": 1.7968996414604462e-07, + "loss": 0.6664, + "step": 4974 + }, + { + "epoch": 9.212962962962964, + "grad_norm": 2.8202926237552584, + "learning_rate": 1.7968090505572188e-07, + "loss": 0.631, + "step": 4975 + }, + { + "epoch": 9.214814814814815, + "grad_norm": 2.2964759395350054, + "learning_rate": 1.7967184417395596e-07, + "loss": 0.6752, + "step": 4976 + }, + { + "epoch": 9.216666666666667, + "grad_norm": 1.4400858026515644, + "learning_rate": 1.796627815009506e-07, + "loss": 0.6465, + "step": 4977 + }, + { + "epoch": 9.218518518518518, + "grad_norm": 1.4722258708980698, + "learning_rate": 1.7965371703690955e-07, + "loss": 0.6335, + "step": 4978 + }, + { + "epoch": 9.22037037037037, + "grad_norm": 1.657948224794598, + "learning_rate": 1.796446507820366e-07, + "loss": 0.6369, + "step": 4979 + }, + { + "epoch": 9.222222222222221, + "grad_norm": 1.3757415842306018, + "learning_rate": 1.796355827365356e-07, + "loss": 0.6434, + "step": 4980 + }, + { + "epoch": 9.224074074074075, + "grad_norm": 1.3779079858147176, + "learning_rate": 1.796265129006104e-07, + "loss": 0.6718, + "step": 4981 + }, + { + "epoch": 9.225925925925926, + "grad_norm": 1.67085509614113, + "learning_rate": 1.7961744127446493e-07, + "loss": 0.6541, + "step": 4982 + }, + { + "epoch": 9.227777777777778, + "grad_norm": 1.9355580127199132, + "learning_rate": 1.7960836785830315e-07, + "loss": 0.6151, + "step": 4983 + }, + { + "epoch": 9.22962962962963, + "grad_norm": 1.4446134482573796, + "learning_rate": 1.7959929265232905e-07, + "loss": 0.597, + "step": 4984 + }, + { + "epoch": 9.231481481481481, + "grad_norm": 2.0160220046845727, + "learning_rate": 1.7959021565674665e-07, + "loss": 0.66, + "step": 4985 + }, + { + "epoch": 9.233333333333333, + "grad_norm": 1.715321013057844, + "learning_rate": 1.7958113687176005e-07, + "loss": 0.6764, + "step": 4986 + }, + { + "epoch": 9.235185185185186, + "grad_norm": 1.8339720848478032, + "learning_rate": 1.7957205629757335e-07, + "loss": 0.6582, + "step": 4987 + }, + { + "epoch": 9.237037037037037, + "grad_norm": 1.5356065454225052, + "learning_rate": 1.795629739343907e-07, + "loss": 0.5962, + "step": 4988 + }, + { + "epoch": 9.238888888888889, + "grad_norm": 1.639472161260969, + "learning_rate": 1.795538897824163e-07, + "loss": 0.643, + "step": 4989 + }, + { + "epoch": 9.24074074074074, + "grad_norm": 1.4742147555421299, + "learning_rate": 1.7954480384185444e-07, + "loss": 0.6185, + "step": 4990 + }, + { + "epoch": 9.242592592592592, + "grad_norm": 1.6465143612137927, + "learning_rate": 1.795357161129093e-07, + "loss": 0.6757, + "step": 4991 + }, + { + "epoch": 9.244444444444444, + "grad_norm": 1.7632031636729146, + "learning_rate": 1.7952662659578528e-07, + "loss": 0.6389, + "step": 4992 + }, + { + "epoch": 9.246296296296297, + "grad_norm": 1.7586768861538635, + "learning_rate": 1.7951753529068666e-07, + "loss": 0.6253, + "step": 4993 + }, + { + "epoch": 9.248148148148148, + "grad_norm": 1.550385184948432, + "learning_rate": 1.7950844219781794e-07, + "loss": 0.6437, + "step": 4994 + }, + { + "epoch": 9.25, + "grad_norm": 1.5605681744560167, + "learning_rate": 1.7949934731738346e-07, + "loss": 0.6486, + "step": 4995 + }, + { + "epoch": 9.251851851851852, + "grad_norm": 1.7945936727718752, + "learning_rate": 1.7949025064958775e-07, + "loss": 0.6621, + "step": 4996 + }, + { + "epoch": 9.253703703703703, + "grad_norm": 1.684355890701175, + "learning_rate": 1.794811521946353e-07, + "loss": 0.6353, + "step": 4997 + }, + { + "epoch": 9.255555555555556, + "grad_norm": 1.6486710606711317, + "learning_rate": 1.7947205195273068e-07, + "loss": 0.6291, + "step": 4998 + }, + { + "epoch": 9.257407407407408, + "grad_norm": 4.526348700046569, + "learning_rate": 1.7946294992407848e-07, + "loss": 0.6539, + "step": 4999 + }, + { + "epoch": 9.25925925925926, + "grad_norm": 1.5726024947315718, + "learning_rate": 1.794538461088834e-07, + "loss": 0.6852, + "step": 5000 + }, + { + "epoch": 9.261111111111111, + "grad_norm": 1.4670976646891696, + "learning_rate": 1.7944474050734998e-07, + "loss": 0.6802, + "step": 5001 + }, + { + "epoch": 9.262962962962963, + "grad_norm": 1.4798819637549998, + "learning_rate": 1.794356331196831e-07, + "loss": 0.6725, + "step": 5002 + }, + { + "epoch": 9.264814814814814, + "grad_norm": 1.4375463696445736, + "learning_rate": 1.794265239460874e-07, + "loss": 0.6283, + "step": 5003 + }, + { + "epoch": 9.266666666666667, + "grad_norm": 1.7066915421212974, + "learning_rate": 1.7941741298676774e-07, + "loss": 0.5981, + "step": 5004 + }, + { + "epoch": 9.268518518518519, + "grad_norm": 1.5268251717681938, + "learning_rate": 1.7940830024192896e-07, + "loss": 0.6221, + "step": 5005 + }, + { + "epoch": 9.27037037037037, + "grad_norm": 1.3769626220092266, + "learning_rate": 1.793991857117759e-07, + "loss": 0.6737, + "step": 5006 + }, + { + "epoch": 9.272222222222222, + "grad_norm": 2.23314283440295, + "learning_rate": 1.793900693965135e-07, + "loss": 0.6003, + "step": 5007 + }, + { + "epoch": 9.274074074074074, + "grad_norm": 1.4265979488281466, + "learning_rate": 1.7938095129634672e-07, + "loss": 0.6762, + "step": 5008 + }, + { + "epoch": 9.275925925925925, + "grad_norm": 1.8198743096616483, + "learning_rate": 1.7937183141148055e-07, + "loss": 0.6462, + "step": 5009 + }, + { + "epoch": 9.277777777777779, + "grad_norm": 1.5188661639899825, + "learning_rate": 1.7936270974212007e-07, + "loss": 0.668, + "step": 5010 + }, + { + "epoch": 9.27962962962963, + "grad_norm": 1.5833737296242645, + "learning_rate": 1.793535862884703e-07, + "loss": 0.622, + "step": 5011 + }, + { + "epoch": 9.281481481481482, + "grad_norm": 1.5193944196457194, + "learning_rate": 1.793444610507364e-07, + "loss": 0.631, + "step": 5012 + }, + { + "epoch": 9.283333333333333, + "grad_norm": 1.672919488798573, + "learning_rate": 1.793353340291235e-07, + "loss": 0.6653, + "step": 5013 + }, + { + "epoch": 9.285185185185185, + "grad_norm": 1.7101912463159068, + "learning_rate": 1.7932620522383685e-07, + "loss": 0.6326, + "step": 5014 + }, + { + "epoch": 9.287037037037036, + "grad_norm": 1.533147391794233, + "learning_rate": 1.7931707463508166e-07, + "loss": 0.6221, + "step": 5015 + }, + { + "epoch": 9.28888888888889, + "grad_norm": 1.455199130564899, + "learning_rate": 1.793079422630632e-07, + "loss": 0.6618, + "step": 5016 + }, + { + "epoch": 9.290740740740741, + "grad_norm": 1.635680136387207, + "learning_rate": 1.792988081079868e-07, + "loss": 0.5942, + "step": 5017 + }, + { + "epoch": 9.292592592592593, + "grad_norm": 1.455392630163267, + "learning_rate": 1.7928967217005779e-07, + "loss": 0.6452, + "step": 5018 + }, + { + "epoch": 9.294444444444444, + "grad_norm": 1.584390106076646, + "learning_rate": 1.7928053444948167e-07, + "loss": 0.6431, + "step": 5019 + }, + { + "epoch": 9.296296296296296, + "grad_norm": 1.4646706362416826, + "learning_rate": 1.7927139494646375e-07, + "loss": 0.6261, + "step": 5020 + }, + { + "epoch": 9.298148148148147, + "grad_norm": 1.6056776596207563, + "learning_rate": 1.792622536612096e-07, + "loss": 0.6545, + "step": 5021 + }, + { + "epoch": 9.3, + "grad_norm": 1.6517396898267862, + "learning_rate": 1.792531105939247e-07, + "loss": 0.6213, + "step": 5022 + }, + { + "epoch": 9.301851851851852, + "grad_norm": 1.624852568444863, + "learning_rate": 1.7924396574481464e-07, + "loss": 0.6624, + "step": 5023 + }, + { + "epoch": 9.303703703703704, + "grad_norm": 1.6881408896966401, + "learning_rate": 1.7923481911408502e-07, + "loss": 0.6024, + "step": 5024 + }, + { + "epoch": 9.305555555555555, + "grad_norm": 1.545826666474559, + "learning_rate": 1.7922567070194146e-07, + "loss": 0.6391, + "step": 5025 + }, + { + "epoch": 9.307407407407407, + "grad_norm": 1.8592296379740392, + "learning_rate": 1.7921652050858965e-07, + "loss": 0.6494, + "step": 5026 + }, + { + "epoch": 9.309259259259258, + "grad_norm": 1.5960725446946296, + "learning_rate": 1.7920736853423528e-07, + "loss": 0.6654, + "step": 5027 + }, + { + "epoch": 9.311111111111112, + "grad_norm": 1.4753447139446967, + "learning_rate": 1.7919821477908417e-07, + "loss": 0.6531, + "step": 5028 + }, + { + "epoch": 9.312962962962963, + "grad_norm": 1.6716827366013367, + "learning_rate": 1.791890592433421e-07, + "loss": 0.6645, + "step": 5029 + }, + { + "epoch": 9.314814814814815, + "grad_norm": 1.984200267143179, + "learning_rate": 1.791799019272149e-07, + "loss": 0.5852, + "step": 5030 + }, + { + "epoch": 9.316666666666666, + "grad_norm": 1.5056632470097804, + "learning_rate": 1.7917074283090843e-07, + "loss": 0.6549, + "step": 5031 + }, + { + "epoch": 9.318518518518518, + "grad_norm": 1.4430046489844697, + "learning_rate": 1.7916158195462865e-07, + "loss": 0.6323, + "step": 5032 + }, + { + "epoch": 9.32037037037037, + "grad_norm": 1.7069982872656422, + "learning_rate": 1.7915241929858152e-07, + "loss": 0.593, + "step": 5033 + }, + { + "epoch": 9.322222222222223, + "grad_norm": 1.5312560951397207, + "learning_rate": 1.79143254862973e-07, + "loss": 0.6298, + "step": 5034 + }, + { + "epoch": 9.324074074074074, + "grad_norm": 1.414994024703364, + "learning_rate": 1.7913408864800919e-07, + "loss": 0.6053, + "step": 5035 + }, + { + "epoch": 9.325925925925926, + "grad_norm": 1.5320441508491534, + "learning_rate": 1.7912492065389613e-07, + "loss": 0.6364, + "step": 5036 + }, + { + "epoch": 9.327777777777778, + "grad_norm": 1.6043268445240035, + "learning_rate": 1.7911575088083993e-07, + "loss": 0.6392, + "step": 5037 + }, + { + "epoch": 9.329629629629629, + "grad_norm": 1.6424564574926412, + "learning_rate": 1.7910657932904683e-07, + "loss": 0.6371, + "step": 5038 + }, + { + "epoch": 9.331481481481482, + "grad_norm": 1.6082424307416978, + "learning_rate": 1.7909740599872296e-07, + "loss": 0.595, + "step": 5039 + }, + { + "epoch": 9.333333333333334, + "grad_norm": 1.6253009453461835, + "learning_rate": 1.7908823089007456e-07, + "loss": 0.634, + "step": 5040 + }, + { + "epoch": 9.335185185185185, + "grad_norm": 1.6991740815851821, + "learning_rate": 1.7907905400330795e-07, + "loss": 0.6424, + "step": 5041 + }, + { + "epoch": 9.337037037037037, + "grad_norm": 1.5019910570051236, + "learning_rate": 1.790698753386294e-07, + "loss": 0.6376, + "step": 5042 + }, + { + "epoch": 9.338888888888889, + "grad_norm": 1.56397856475829, + "learning_rate": 1.7906069489624532e-07, + "loss": 0.6495, + "step": 5043 + }, + { + "epoch": 9.34074074074074, + "grad_norm": 2.7932846153355073, + "learning_rate": 1.790515126763621e-07, + "loss": 0.6404, + "step": 5044 + }, + { + "epoch": 9.342592592592593, + "grad_norm": 1.583012494386562, + "learning_rate": 1.7904232867918614e-07, + "loss": 0.5802, + "step": 5045 + }, + { + "epoch": 9.344444444444445, + "grad_norm": 1.333726286358759, + "learning_rate": 1.79033142904924e-07, + "loss": 0.6004, + "step": 5046 + }, + { + "epoch": 9.346296296296297, + "grad_norm": 1.459715078531212, + "learning_rate": 1.7902395535378213e-07, + "loss": 0.6563, + "step": 5047 + }, + { + "epoch": 9.348148148148148, + "grad_norm": 1.615068153046038, + "learning_rate": 1.7901476602596714e-07, + "loss": 0.6644, + "step": 5048 + }, + { + "epoch": 9.35, + "grad_norm": 1.6605966537204446, + "learning_rate": 1.7900557492168559e-07, + "loss": 0.6561, + "step": 5049 + }, + { + "epoch": 9.351851851851851, + "grad_norm": 1.4285911454793239, + "learning_rate": 1.7899638204114413e-07, + "loss": 0.6379, + "step": 5050 + }, + { + "epoch": 9.353703703703705, + "grad_norm": 1.8774116831973644, + "learning_rate": 1.7898718738454947e-07, + "loss": 0.6315, + "step": 5051 + }, + { + "epoch": 9.355555555555556, + "grad_norm": 1.57376946696217, + "learning_rate": 1.789779909521083e-07, + "loss": 0.6354, + "step": 5052 + }, + { + "epoch": 9.357407407407408, + "grad_norm": 1.6494245917492636, + "learning_rate": 1.7896879274402742e-07, + "loss": 0.6921, + "step": 5053 + }, + { + "epoch": 9.35925925925926, + "grad_norm": 2.6664923061154884, + "learning_rate": 1.7895959276051355e-07, + "loss": 0.6535, + "step": 5054 + }, + { + "epoch": 9.36111111111111, + "grad_norm": 1.773639458781967, + "learning_rate": 1.7895039100177365e-07, + "loss": 0.6038, + "step": 5055 + }, + { + "epoch": 9.362962962962962, + "grad_norm": 1.736902489949008, + "learning_rate": 1.789411874680145e-07, + "loss": 0.6338, + "step": 5056 + }, + { + "epoch": 9.364814814814816, + "grad_norm": 1.6014590616897588, + "learning_rate": 1.7893198215944304e-07, + "loss": 0.6465, + "step": 5057 + }, + { + "epoch": 9.366666666666667, + "grad_norm": 1.5141258118210914, + "learning_rate": 1.7892277507626626e-07, + "loss": 0.6235, + "step": 5058 + }, + { + "epoch": 9.368518518518519, + "grad_norm": 1.76410411113449, + "learning_rate": 1.7891356621869115e-07, + "loss": 0.6613, + "step": 5059 + }, + { + "epoch": 9.37037037037037, + "grad_norm": 1.7630720752523528, + "learning_rate": 1.7890435558692474e-07, + "loss": 0.6303, + "step": 5060 + }, + { + "epoch": 9.372222222222222, + "grad_norm": 3.0677187889485094, + "learning_rate": 1.788951431811741e-07, + "loss": 0.629, + "step": 5061 + }, + { + "epoch": 9.374074074074073, + "grad_norm": 1.3081737218519582, + "learning_rate": 1.788859290016464e-07, + "loss": 0.6191, + "step": 5062 + }, + { + "epoch": 9.375925925925927, + "grad_norm": 1.3643717661850152, + "learning_rate": 1.7887671304854874e-07, + "loss": 0.6237, + "step": 5063 + }, + { + "epoch": 9.377777777777778, + "grad_norm": 2.0807765331599275, + "learning_rate": 1.7886749532208836e-07, + "loss": 0.68, + "step": 5064 + }, + { + "epoch": 9.37962962962963, + "grad_norm": 1.4550154309273073, + "learning_rate": 1.7885827582247245e-07, + "loss": 0.6136, + "step": 5065 + }, + { + "epoch": 9.381481481481481, + "grad_norm": 1.7066524589216818, + "learning_rate": 1.7884905454990837e-07, + "loss": 0.6129, + "step": 5066 + }, + { + "epoch": 9.383333333333333, + "grad_norm": 1.4183323629222169, + "learning_rate": 1.7883983150460336e-07, + "loss": 0.6747, + "step": 5067 + }, + { + "epoch": 9.385185185185184, + "grad_norm": 1.7227217516296243, + "learning_rate": 1.7883060668676483e-07, + "loss": 0.6271, + "step": 5068 + }, + { + "epoch": 9.387037037037038, + "grad_norm": 1.520639856957348, + "learning_rate": 1.7882138009660018e-07, + "loss": 0.5745, + "step": 5069 + }, + { + "epoch": 9.38888888888889, + "grad_norm": 1.5521383855107516, + "learning_rate": 1.7881215173431678e-07, + "loss": 0.6217, + "step": 5070 + }, + { + "epoch": 9.39074074074074, + "grad_norm": 1.364430124886182, + "learning_rate": 1.788029216001222e-07, + "loss": 0.6758, + "step": 5071 + }, + { + "epoch": 9.392592592592592, + "grad_norm": 1.824452990571026, + "learning_rate": 1.787936896942239e-07, + "loss": 0.58, + "step": 5072 + }, + { + "epoch": 9.394444444444444, + "grad_norm": 1.4294758168433916, + "learning_rate": 1.7878445601682947e-07, + "loss": 0.6548, + "step": 5073 + }, + { + "epoch": 9.396296296296295, + "grad_norm": 1.6753701967047336, + "learning_rate": 1.7877522056814644e-07, + "loss": 0.6287, + "step": 5074 + }, + { + "epoch": 9.398148148148149, + "grad_norm": 1.3616281473806655, + "learning_rate": 1.7876598334838255e-07, + "loss": 0.643, + "step": 5075 + }, + { + "epoch": 9.4, + "grad_norm": 1.6751669454515343, + "learning_rate": 1.7875674435774544e-07, + "loss": 0.6656, + "step": 5076 + }, + { + "epoch": 9.401851851851852, + "grad_norm": 1.6106451220892728, + "learning_rate": 1.787475035964428e-07, + "loss": 0.5968, + "step": 5077 + }, + { + "epoch": 9.403703703703703, + "grad_norm": 1.4867019568871895, + "learning_rate": 1.787382610646824e-07, + "loss": 0.6628, + "step": 5078 + }, + { + "epoch": 9.405555555555555, + "grad_norm": 1.9228905910102003, + "learning_rate": 1.7872901676267206e-07, + "loss": 0.6343, + "step": 5079 + }, + { + "epoch": 9.407407407407407, + "grad_norm": 1.6840921115711178, + "learning_rate": 1.7871977069061957e-07, + "loss": 0.6106, + "step": 5080 + }, + { + "epoch": 9.40925925925926, + "grad_norm": 1.7853776228622724, + "learning_rate": 1.7871052284873284e-07, + "loss": 0.6283, + "step": 5081 + }, + { + "epoch": 9.411111111111111, + "grad_norm": 2.6738343791571584, + "learning_rate": 1.7870127323721977e-07, + "loss": 0.6084, + "step": 5082 + }, + { + "epoch": 9.412962962962963, + "grad_norm": 1.5013127229298282, + "learning_rate": 1.7869202185628836e-07, + "loss": 0.5787, + "step": 5083 + }, + { + "epoch": 9.414814814814815, + "grad_norm": 1.523815238755022, + "learning_rate": 1.7868276870614656e-07, + "loss": 0.601, + "step": 5084 + }, + { + "epoch": 9.416666666666666, + "grad_norm": 1.8533380354214934, + "learning_rate": 1.786735137870024e-07, + "loss": 0.6182, + "step": 5085 + }, + { + "epoch": 9.418518518518518, + "grad_norm": 1.580346653812946, + "learning_rate": 1.7866425709906402e-07, + "loss": 0.6084, + "step": 5086 + }, + { + "epoch": 9.420370370370371, + "grad_norm": 1.6970851324684704, + "learning_rate": 1.7865499864253945e-07, + "loss": 0.5958, + "step": 5087 + }, + { + "epoch": 9.422222222222222, + "grad_norm": 1.719140709402049, + "learning_rate": 1.7864573841763693e-07, + "loss": 0.6361, + "step": 5088 + }, + { + "epoch": 9.424074074074074, + "grad_norm": 2.84831840004522, + "learning_rate": 1.7863647642456456e-07, + "loss": 0.654, + "step": 5089 + }, + { + "epoch": 9.425925925925926, + "grad_norm": 7.152127254163449, + "learning_rate": 1.7862721266353067e-07, + "loss": 0.6044, + "step": 5090 + }, + { + "epoch": 9.427777777777777, + "grad_norm": 1.4784232893295477, + "learning_rate": 1.7861794713474347e-07, + "loss": 0.6742, + "step": 5091 + }, + { + "epoch": 9.42962962962963, + "grad_norm": 1.509768925031899, + "learning_rate": 1.786086798384113e-07, + "loss": 0.618, + "step": 5092 + }, + { + "epoch": 9.431481481481482, + "grad_norm": 1.3703300928592905, + "learning_rate": 1.7859941077474253e-07, + "loss": 0.6689, + "step": 5093 + }, + { + "epoch": 9.433333333333334, + "grad_norm": 1.6585919113697565, + "learning_rate": 1.785901399439455e-07, + "loss": 0.6496, + "step": 5094 + }, + { + "epoch": 9.435185185185185, + "grad_norm": 1.4089886894937549, + "learning_rate": 1.7858086734622868e-07, + "loss": 0.6348, + "step": 5095 + }, + { + "epoch": 9.437037037037037, + "grad_norm": 1.5095432816949876, + "learning_rate": 1.7857159298180054e-07, + "loss": 0.6612, + "step": 5096 + }, + { + "epoch": 9.438888888888888, + "grad_norm": 1.7114269335120136, + "learning_rate": 1.785623168508696e-07, + "loss": 0.6198, + "step": 5097 + }, + { + "epoch": 9.440740740740742, + "grad_norm": 1.7210746541647095, + "learning_rate": 1.785530389536444e-07, + "loss": 0.6665, + "step": 5098 + }, + { + "epoch": 9.442592592592593, + "grad_norm": 1.7576859124920148, + "learning_rate": 1.7854375929033355e-07, + "loss": 0.6513, + "step": 5099 + }, + { + "epoch": 9.444444444444445, + "grad_norm": 1.5658834118143432, + "learning_rate": 1.785344778611457e-07, + "loss": 0.5689, + "step": 5100 + }, + { + "epoch": 9.446296296296296, + "grad_norm": 1.540458788306279, + "learning_rate": 1.7852519466628944e-07, + "loss": 0.5959, + "step": 5101 + }, + { + "epoch": 9.448148148148148, + "grad_norm": 1.6910202056877777, + "learning_rate": 1.7851590970597356e-07, + "loss": 0.625, + "step": 5102 + }, + { + "epoch": 9.45, + "grad_norm": 1.5519802611360196, + "learning_rate": 1.7850662298040676e-07, + "loss": 0.6103, + "step": 5103 + }, + { + "epoch": 9.451851851851853, + "grad_norm": 1.6778469011059622, + "learning_rate": 1.7849733448979784e-07, + "loss": 0.6501, + "step": 5104 + }, + { + "epoch": 9.453703703703704, + "grad_norm": 1.5762673708222676, + "learning_rate": 1.7848804423435568e-07, + "loss": 0.6202, + "step": 5105 + }, + { + "epoch": 9.455555555555556, + "grad_norm": 1.879610652802891, + "learning_rate": 1.784787522142891e-07, + "loss": 0.6902, + "step": 5106 + }, + { + "epoch": 9.457407407407407, + "grad_norm": 1.5301889028355498, + "learning_rate": 1.7846945842980702e-07, + "loss": 0.6206, + "step": 5107 + }, + { + "epoch": 9.459259259259259, + "grad_norm": 2.1021513100141274, + "learning_rate": 1.7846016288111843e-07, + "loss": 0.6385, + "step": 5108 + }, + { + "epoch": 9.46111111111111, + "grad_norm": 1.6625489146824344, + "learning_rate": 1.7845086556843224e-07, + "loss": 0.6725, + "step": 5109 + }, + { + "epoch": 9.462962962962964, + "grad_norm": 2.062150417433819, + "learning_rate": 1.7844156649195757e-07, + "loss": 0.5987, + "step": 5110 + }, + { + "epoch": 9.464814814814815, + "grad_norm": 1.6708835458354963, + "learning_rate": 1.7843226565190342e-07, + "loss": 0.6644, + "step": 5111 + }, + { + "epoch": 9.466666666666667, + "grad_norm": 1.7331393540503923, + "learning_rate": 1.784229630484789e-07, + "loss": 0.6783, + "step": 5112 + }, + { + "epoch": 9.468518518518518, + "grad_norm": 1.70479119710824, + "learning_rate": 1.784136586818932e-07, + "loss": 0.6195, + "step": 5113 + }, + { + "epoch": 9.47037037037037, + "grad_norm": 1.434663607746278, + "learning_rate": 1.7840435255235547e-07, + "loss": 0.6209, + "step": 5114 + }, + { + "epoch": 9.472222222222221, + "grad_norm": 2.277514589162537, + "learning_rate": 1.7839504466007497e-07, + "loss": 0.6388, + "step": 5115 + }, + { + "epoch": 9.474074074074075, + "grad_norm": 1.4296825968964497, + "learning_rate": 1.7838573500526093e-07, + "loss": 0.6048, + "step": 5116 + }, + { + "epoch": 9.475925925925926, + "grad_norm": 1.8716381595229283, + "learning_rate": 1.7837642358812267e-07, + "loss": 0.6581, + "step": 5117 + }, + { + "epoch": 9.477777777777778, + "grad_norm": 1.5201973085330085, + "learning_rate": 1.7836711040886953e-07, + "loss": 0.5798, + "step": 5118 + }, + { + "epoch": 9.47962962962963, + "grad_norm": 1.6050180766263866, + "learning_rate": 1.7835779546771096e-07, + "loss": 0.6443, + "step": 5119 + }, + { + "epoch": 9.481481481481481, + "grad_norm": 1.5172610186244304, + "learning_rate": 1.783484787648563e-07, + "loss": 0.6428, + "step": 5120 + }, + { + "epoch": 9.483333333333333, + "grad_norm": 1.618322725372126, + "learning_rate": 1.7833916030051503e-07, + "loss": 0.6407, + "step": 5121 + }, + { + "epoch": 9.485185185185186, + "grad_norm": 1.8773323526928989, + "learning_rate": 1.7832984007489668e-07, + "loss": 0.6215, + "step": 5122 + }, + { + "epoch": 9.487037037037037, + "grad_norm": 2.108726855722813, + "learning_rate": 1.783205180882108e-07, + "loss": 0.6491, + "step": 5123 + }, + { + "epoch": 9.488888888888889, + "grad_norm": 1.639773408382761, + "learning_rate": 1.7831119434066694e-07, + "loss": 0.6489, + "step": 5124 + }, + { + "epoch": 9.49074074074074, + "grad_norm": 2.0722175704082835, + "learning_rate": 1.7830186883247476e-07, + "loss": 0.661, + "step": 5125 + }, + { + "epoch": 9.492592592592592, + "grad_norm": 2.067283352480185, + "learning_rate": 1.782925415638439e-07, + "loss": 0.6838, + "step": 5126 + }, + { + "epoch": 9.494444444444444, + "grad_norm": 1.3923213384515385, + "learning_rate": 1.7828321253498403e-07, + "loss": 0.5992, + "step": 5127 + }, + { + "epoch": 9.496296296296297, + "grad_norm": 1.5739686797871817, + "learning_rate": 1.7827388174610497e-07, + "loss": 0.6504, + "step": 5128 + }, + { + "epoch": 9.498148148148148, + "grad_norm": 1.6540811415318923, + "learning_rate": 1.7826454919741645e-07, + "loss": 0.6271, + "step": 5129 + }, + { + "epoch": 9.5, + "grad_norm": 3.8587257914372284, + "learning_rate": 1.782552148891283e-07, + "loss": 0.6497, + "step": 5130 + }, + { + "epoch": 9.501851851851852, + "grad_norm": 1.6799831928268083, + "learning_rate": 1.7824587882145036e-07, + "loss": 0.5996, + "step": 5131 + }, + { + "epoch": 9.503703703703703, + "grad_norm": 1.5049728621744716, + "learning_rate": 1.7823654099459258e-07, + "loss": 0.6363, + "step": 5132 + }, + { + "epoch": 9.505555555555556, + "grad_norm": 3.021252466015714, + "learning_rate": 1.7822720140876485e-07, + "loss": 0.6546, + "step": 5133 + }, + { + "epoch": 9.507407407407408, + "grad_norm": 3.196935502182323, + "learning_rate": 1.7821786006417718e-07, + "loss": 0.5997, + "step": 5134 + }, + { + "epoch": 9.50925925925926, + "grad_norm": 1.5023900063816933, + "learning_rate": 1.7820851696103958e-07, + "loss": 0.6421, + "step": 5135 + }, + { + "epoch": 9.511111111111111, + "grad_norm": 1.6237543035315154, + "learning_rate": 1.7819917209956211e-07, + "loss": 0.6999, + "step": 5136 + }, + { + "epoch": 9.512962962962963, + "grad_norm": 1.5790958928468297, + "learning_rate": 1.781898254799549e-07, + "loss": 0.6622, + "step": 5137 + }, + { + "epoch": 9.514814814814814, + "grad_norm": 1.7173060074627764, + "learning_rate": 1.78180477102428e-07, + "loss": 0.6191, + "step": 5138 + }, + { + "epoch": 9.516666666666667, + "grad_norm": 1.6711628774117762, + "learning_rate": 1.7817112696719166e-07, + "loss": 0.607, + "step": 5139 + }, + { + "epoch": 9.518518518518519, + "grad_norm": 1.6399646472875993, + "learning_rate": 1.7816177507445606e-07, + "loss": 0.6494, + "step": 5140 + }, + { + "epoch": 9.52037037037037, + "grad_norm": 1.695678394425491, + "learning_rate": 1.7815242142443147e-07, + "loss": 0.6201, + "step": 5141 + }, + { + "epoch": 9.522222222222222, + "grad_norm": 1.6957190773111068, + "learning_rate": 1.781430660173282e-07, + "loss": 0.6229, + "step": 5142 + }, + { + "epoch": 9.524074074074074, + "grad_norm": 1.497449595135095, + "learning_rate": 1.781337088533566e-07, + "loss": 0.6557, + "step": 5143 + }, + { + "epoch": 9.525925925925925, + "grad_norm": 1.6345201826860112, + "learning_rate": 1.7812434993272702e-07, + "loss": 0.6589, + "step": 5144 + }, + { + "epoch": 9.527777777777779, + "grad_norm": 1.5690363683069524, + "learning_rate": 1.781149892556498e-07, + "loss": 0.5877, + "step": 5145 + }, + { + "epoch": 9.52962962962963, + "grad_norm": 2.1255811656063455, + "learning_rate": 1.7810562682233552e-07, + "loss": 0.6219, + "step": 5146 + }, + { + "epoch": 9.531481481481482, + "grad_norm": 2.534527525318062, + "learning_rate": 1.780962626329946e-07, + "loss": 0.6553, + "step": 5147 + }, + { + "epoch": 9.533333333333333, + "grad_norm": 1.64193888496852, + "learning_rate": 1.780868966878376e-07, + "loss": 0.6376, + "step": 5148 + }, + { + "epoch": 9.535185185185185, + "grad_norm": 1.725739169929844, + "learning_rate": 1.7807752898707507e-07, + "loss": 0.6249, + "step": 5149 + }, + { + "epoch": 9.537037037037036, + "grad_norm": 1.4721775872082243, + "learning_rate": 1.7806815953091766e-07, + "loss": 0.6355, + "step": 5150 + }, + { + "epoch": 9.53888888888889, + "grad_norm": 1.8340492004921911, + "learning_rate": 1.7805878831957597e-07, + "loss": 0.6165, + "step": 5151 + }, + { + "epoch": 9.540740740740741, + "grad_norm": 1.5088477590182456, + "learning_rate": 1.780494153532607e-07, + "loss": 0.6542, + "step": 5152 + }, + { + "epoch": 9.542592592592593, + "grad_norm": 1.6109125651047485, + "learning_rate": 1.780400406321826e-07, + "loss": 0.6269, + "step": 5153 + }, + { + "epoch": 9.544444444444444, + "grad_norm": 1.5492598227531769, + "learning_rate": 1.7803066415655245e-07, + "loss": 0.6053, + "step": 5154 + }, + { + "epoch": 9.546296296296296, + "grad_norm": 1.671871532019184, + "learning_rate": 1.7802128592658105e-07, + "loss": 0.6512, + "step": 5155 + }, + { + "epoch": 9.548148148148147, + "grad_norm": 1.6899592914073456, + "learning_rate": 1.7801190594247923e-07, + "loss": 0.6465, + "step": 5156 + }, + { + "epoch": 9.55, + "grad_norm": 1.456399761086655, + "learning_rate": 1.7800252420445787e-07, + "loss": 0.6107, + "step": 5157 + }, + { + "epoch": 9.551851851851852, + "grad_norm": 1.7004387622656227, + "learning_rate": 1.7799314071272794e-07, + "loss": 0.6229, + "step": 5158 + }, + { + "epoch": 9.553703703703704, + "grad_norm": 1.6615469148594542, + "learning_rate": 1.7798375546750038e-07, + "loss": 0.609, + "step": 5159 + }, + { + "epoch": 9.555555555555555, + "grad_norm": 1.4667138801761859, + "learning_rate": 1.7797436846898617e-07, + "loss": 0.5894, + "step": 5160 + }, + { + "epoch": 9.557407407407407, + "grad_norm": 1.9053834502967453, + "learning_rate": 1.779649797173964e-07, + "loss": 0.6887, + "step": 5161 + }, + { + "epoch": 9.559259259259258, + "grad_norm": 1.4768371089592753, + "learning_rate": 1.7795558921294212e-07, + "loss": 0.622, + "step": 5162 + }, + { + "epoch": 9.561111111111112, + "grad_norm": 2.6354511803852234, + "learning_rate": 1.779461969558345e-07, + "loss": 0.6399, + "step": 5163 + }, + { + "epoch": 9.562962962962963, + "grad_norm": 1.4310463367017272, + "learning_rate": 1.7793680294628465e-07, + "loss": 0.6697, + "step": 5164 + }, + { + "epoch": 9.564814814814815, + "grad_norm": 2.484626730847231, + "learning_rate": 1.779274071845038e-07, + "loss": 0.6155, + "step": 5165 + }, + { + "epoch": 9.566666666666666, + "grad_norm": 1.4971905872094466, + "learning_rate": 1.779180096707032e-07, + "loss": 0.6225, + "step": 5166 + }, + { + "epoch": 9.568518518518518, + "grad_norm": 1.5802732312685048, + "learning_rate": 1.779086104050941e-07, + "loss": 0.6346, + "step": 5167 + }, + { + "epoch": 9.57037037037037, + "grad_norm": 1.394884073892824, + "learning_rate": 1.7789920938788787e-07, + "loss": 0.6268, + "step": 5168 + }, + { + "epoch": 9.572222222222223, + "grad_norm": 1.5276213110667942, + "learning_rate": 1.7788980661929583e-07, + "loss": 0.629, + "step": 5169 + }, + { + "epoch": 9.574074074074074, + "grad_norm": 1.546255824459983, + "learning_rate": 1.778804020995294e-07, + "loss": 0.6479, + "step": 5170 + }, + { + "epoch": 9.575925925925926, + "grad_norm": 1.6529362687982083, + "learning_rate": 1.7787099582879998e-07, + "loss": 0.6371, + "step": 5171 + }, + { + "epoch": 9.577777777777778, + "grad_norm": 1.4738197643713067, + "learning_rate": 1.778615878073191e-07, + "loss": 0.6295, + "step": 5172 + }, + { + "epoch": 9.579629629629629, + "grad_norm": 1.6154524152006462, + "learning_rate": 1.7785217803529825e-07, + "loss": 0.5942, + "step": 5173 + }, + { + "epoch": 9.581481481481482, + "grad_norm": 1.6714505474308243, + "learning_rate": 1.77842766512949e-07, + "loss": 0.6211, + "step": 5174 + }, + { + "epoch": 9.583333333333334, + "grad_norm": 1.4056140393538294, + "learning_rate": 1.7783335324048294e-07, + "loss": 0.6406, + "step": 5175 + }, + { + "epoch": 9.585185185185185, + "grad_norm": 1.7289370028218976, + "learning_rate": 1.778239382181117e-07, + "loss": 0.6361, + "step": 5176 + }, + { + "epoch": 9.587037037037037, + "grad_norm": 1.4221791421804786, + "learning_rate": 1.77814521446047e-07, + "loss": 0.6235, + "step": 5177 + }, + { + "epoch": 9.588888888888889, + "grad_norm": 1.6712950723596691, + "learning_rate": 1.7780510292450046e-07, + "loss": 0.6586, + "step": 5178 + }, + { + "epoch": 9.59074074074074, + "grad_norm": 1.643774154377951, + "learning_rate": 1.7779568265368394e-07, + "loss": 0.6017, + "step": 5179 + }, + { + "epoch": 9.592592592592592, + "grad_norm": 1.4603321904356052, + "learning_rate": 1.7778626063380915e-07, + "loss": 0.6208, + "step": 5180 + }, + { + "epoch": 9.594444444444445, + "grad_norm": 1.3769500135065982, + "learning_rate": 1.77776836865088e-07, + "loss": 0.6763, + "step": 5181 + }, + { + "epoch": 9.596296296296297, + "grad_norm": 1.5225688446939925, + "learning_rate": 1.7776741134773227e-07, + "loss": 0.6339, + "step": 5182 + }, + { + "epoch": 9.598148148148148, + "grad_norm": 1.41329568822239, + "learning_rate": 1.7775798408195392e-07, + "loss": 0.6313, + "step": 5183 + }, + { + "epoch": 9.6, + "grad_norm": 2.763628993492429, + "learning_rate": 1.7774855506796493e-07, + "loss": 0.6063, + "step": 5184 + }, + { + "epoch": 9.601851851851851, + "grad_norm": 2.099629127338547, + "learning_rate": 1.7773912430597723e-07, + "loss": 0.648, + "step": 5185 + }, + { + "epoch": 9.603703703703705, + "grad_norm": 1.6274385462837548, + "learning_rate": 1.7772969179620293e-07, + "loss": 0.6653, + "step": 5186 + }, + { + "epoch": 9.605555555555556, + "grad_norm": 1.4847810567716533, + "learning_rate": 1.7772025753885397e-07, + "loss": 0.6121, + "step": 5187 + }, + { + "epoch": 9.607407407407408, + "grad_norm": 1.552257872441247, + "learning_rate": 1.777108215341426e-07, + "loss": 0.6946, + "step": 5188 + }, + { + "epoch": 9.60925925925926, + "grad_norm": 2.602667836660394, + "learning_rate": 1.777013837822809e-07, + "loss": 0.638, + "step": 5189 + }, + { + "epoch": 9.61111111111111, + "grad_norm": 1.4428224745624634, + "learning_rate": 1.7769194428348105e-07, + "loss": 0.6261, + "step": 5190 + }, + { + "epoch": 9.612962962962962, + "grad_norm": 1.385332057250731, + "learning_rate": 1.7768250303795527e-07, + "loss": 0.6089, + "step": 5191 + }, + { + "epoch": 9.614814814814816, + "grad_norm": 1.6071243225937228, + "learning_rate": 1.7767306004591584e-07, + "loss": 0.6226, + "step": 5192 + }, + { + "epoch": 9.616666666666667, + "grad_norm": 1.714399645500842, + "learning_rate": 1.7766361530757506e-07, + "loss": 0.6649, + "step": 5193 + }, + { + "epoch": 9.618518518518519, + "grad_norm": 1.8431869429813752, + "learning_rate": 1.776541688231453e-07, + "loss": 0.62, + "step": 5194 + }, + { + "epoch": 9.62037037037037, + "grad_norm": 1.5711684801866628, + "learning_rate": 1.776447205928389e-07, + "loss": 0.6078, + "step": 5195 + }, + { + "epoch": 9.622222222222222, + "grad_norm": 1.5294516711240607, + "learning_rate": 1.776352706168683e-07, + "loss": 0.6681, + "step": 5196 + }, + { + "epoch": 9.624074074074073, + "grad_norm": 1.7868353780968012, + "learning_rate": 1.7762581889544597e-07, + "loss": 0.6329, + "step": 5197 + }, + { + "epoch": 9.625925925925927, + "grad_norm": 1.3568398851367105, + "learning_rate": 1.776163654287844e-07, + "loss": 0.6291, + "step": 5198 + }, + { + "epoch": 9.627777777777778, + "grad_norm": 1.4893094509209255, + "learning_rate": 1.7760691021709613e-07, + "loss": 0.6506, + "step": 5199 + }, + { + "epoch": 9.62962962962963, + "grad_norm": 1.874732365358208, + "learning_rate": 1.7759745326059376e-07, + "loss": 0.6492, + "step": 5200 + }, + { + "epoch": 9.631481481481481, + "grad_norm": 1.9202279984534458, + "learning_rate": 1.7758799455948985e-07, + "loss": 0.642, + "step": 5201 + }, + { + "epoch": 9.633333333333333, + "grad_norm": 2.3227677664316544, + "learning_rate": 1.7757853411399712e-07, + "loss": 0.6467, + "step": 5202 + }, + { + "epoch": 9.635185185185184, + "grad_norm": 1.613038984739922, + "learning_rate": 1.7756907192432826e-07, + "loss": 0.6274, + "step": 5203 + }, + { + "epoch": 9.637037037037038, + "grad_norm": 1.4347573558348037, + "learning_rate": 1.7755960799069597e-07, + "loss": 0.6263, + "step": 5204 + }, + { + "epoch": 9.63888888888889, + "grad_norm": 1.7466899842732566, + "learning_rate": 1.7755014231331305e-07, + "loss": 0.638, + "step": 5205 + }, + { + "epoch": 9.64074074074074, + "grad_norm": 1.5846791364161328, + "learning_rate": 1.7754067489239227e-07, + "loss": 0.6778, + "step": 5206 + }, + { + "epoch": 9.642592592592592, + "grad_norm": 3.6658872901621606, + "learning_rate": 1.7753120572814656e-07, + "loss": 0.6194, + "step": 5207 + }, + { + "epoch": 9.644444444444444, + "grad_norm": 1.5164463332868299, + "learning_rate": 1.7752173482078878e-07, + "loss": 0.6609, + "step": 5208 + }, + { + "epoch": 9.646296296296295, + "grad_norm": 1.7075193381584803, + "learning_rate": 1.7751226217053183e-07, + "loss": 0.6115, + "step": 5209 + }, + { + "epoch": 9.648148148148149, + "grad_norm": 1.7424322027265615, + "learning_rate": 1.7750278777758873e-07, + "loss": 0.6028, + "step": 5210 + }, + { + "epoch": 9.65, + "grad_norm": 1.5986285043153505, + "learning_rate": 1.7749331164217247e-07, + "loss": 0.6683, + "step": 5211 + }, + { + "epoch": 9.651851851851852, + "grad_norm": 1.747458245049754, + "learning_rate": 1.7748383376449608e-07, + "loss": 0.639, + "step": 5212 + }, + { + "epoch": 9.653703703703703, + "grad_norm": 1.815650062970967, + "learning_rate": 1.7747435414477265e-07, + "loss": 0.6187, + "step": 5213 + }, + { + "epoch": 9.655555555555555, + "grad_norm": 1.4163869171878571, + "learning_rate": 1.7746487278321536e-07, + "loss": 0.6136, + "step": 5214 + }, + { + "epoch": 9.657407407407408, + "grad_norm": 1.37657137540675, + "learning_rate": 1.774553896800373e-07, + "loss": 0.6349, + "step": 5215 + }, + { + "epoch": 9.65925925925926, + "grad_norm": 1.682017507004502, + "learning_rate": 1.7744590483545174e-07, + "loss": 0.6413, + "step": 5216 + }, + { + "epoch": 9.661111111111111, + "grad_norm": 1.7080671854608027, + "learning_rate": 1.774364182496719e-07, + "loss": 0.6344, + "step": 5217 + }, + { + "epoch": 9.662962962962963, + "grad_norm": 1.74956277306685, + "learning_rate": 1.7742692992291103e-07, + "loss": 0.6696, + "step": 5218 + }, + { + "epoch": 9.664814814814815, + "grad_norm": 1.8695292464640438, + "learning_rate": 1.7741743985538253e-07, + "loss": 0.6596, + "step": 5219 + }, + { + "epoch": 9.666666666666666, + "grad_norm": 1.6476566264566908, + "learning_rate": 1.774079480472997e-07, + "loss": 0.6604, + "step": 5220 + }, + { + "epoch": 9.668518518518518, + "grad_norm": 1.5783736071658105, + "learning_rate": 1.7739845449887593e-07, + "loss": 0.6457, + "step": 5221 + }, + { + "epoch": 9.670370370370371, + "grad_norm": 1.9794647868971647, + "learning_rate": 1.7738895921032475e-07, + "loss": 0.655, + "step": 5222 + }, + { + "epoch": 9.672222222222222, + "grad_norm": 2.006583263700009, + "learning_rate": 1.7737946218185956e-07, + "loss": 0.6569, + "step": 5223 + }, + { + "epoch": 9.674074074074074, + "grad_norm": 2.2476789975727973, + "learning_rate": 1.773699634136939e-07, + "loss": 0.6326, + "step": 5224 + }, + { + "epoch": 9.675925925925926, + "grad_norm": 1.5229373863324145, + "learning_rate": 1.7736046290604132e-07, + "loss": 0.6416, + "step": 5225 + }, + { + "epoch": 9.677777777777777, + "grad_norm": 1.524412022051299, + "learning_rate": 1.7735096065911543e-07, + "loss": 0.6532, + "step": 5226 + }, + { + "epoch": 9.67962962962963, + "grad_norm": 1.7604936797410624, + "learning_rate": 1.7734145667312984e-07, + "loss": 0.6304, + "step": 5227 + }, + { + "epoch": 9.681481481481482, + "grad_norm": 1.5751223914392847, + "learning_rate": 1.773319509482983e-07, + "loss": 0.6129, + "step": 5228 + }, + { + "epoch": 9.683333333333334, + "grad_norm": 1.5848999306656517, + "learning_rate": 1.7732244348483443e-07, + "loss": 0.6085, + "step": 5229 + }, + { + "epoch": 9.685185185185185, + "grad_norm": 1.5888462830377457, + "learning_rate": 1.7731293428295204e-07, + "loss": 0.5864, + "step": 5230 + }, + { + "epoch": 9.687037037037037, + "grad_norm": 1.6421499962017303, + "learning_rate": 1.7730342334286492e-07, + "loss": 0.6146, + "step": 5231 + }, + { + "epoch": 9.688888888888888, + "grad_norm": 1.5107469942687584, + "learning_rate": 1.7729391066478686e-07, + "loss": 0.6208, + "step": 5232 + }, + { + "epoch": 9.690740740740742, + "grad_norm": 1.747309535084251, + "learning_rate": 1.7728439624893178e-07, + "loss": 0.6573, + "step": 5233 + }, + { + "epoch": 9.692592592592593, + "grad_norm": 1.9454191852625105, + "learning_rate": 1.772748800955136e-07, + "loss": 0.629, + "step": 5234 + }, + { + "epoch": 9.694444444444445, + "grad_norm": 1.6448648014087421, + "learning_rate": 1.772653622047462e-07, + "loss": 0.6271, + "step": 5235 + }, + { + "epoch": 9.696296296296296, + "grad_norm": 1.4772561174760583, + "learning_rate": 1.7725584257684362e-07, + "loss": 0.6263, + "step": 5236 + }, + { + "epoch": 9.698148148148148, + "grad_norm": 1.650343331512508, + "learning_rate": 1.7724632121201986e-07, + "loss": 0.6498, + "step": 5237 + }, + { + "epoch": 9.7, + "grad_norm": 1.8749089689357918, + "learning_rate": 1.77236798110489e-07, + "loss": 0.6204, + "step": 5238 + }, + { + "epoch": 9.701851851851853, + "grad_norm": 1.4546396285868493, + "learning_rate": 1.7722727327246515e-07, + "loss": 0.642, + "step": 5239 + }, + { + "epoch": 9.703703703703704, + "grad_norm": 1.4572608165712329, + "learning_rate": 1.7721774669816248e-07, + "loss": 0.6653, + "step": 5240 + }, + { + "epoch": 9.705555555555556, + "grad_norm": 1.5192498582193945, + "learning_rate": 1.7720821838779511e-07, + "loss": 0.67, + "step": 5241 + }, + { + "epoch": 9.707407407407407, + "grad_norm": 1.494429690589304, + "learning_rate": 1.7719868834157734e-07, + "loss": 0.6647, + "step": 5242 + }, + { + "epoch": 9.709259259259259, + "grad_norm": 1.4287165905348957, + "learning_rate": 1.7718915655972336e-07, + "loss": 0.6596, + "step": 5243 + }, + { + "epoch": 9.71111111111111, + "grad_norm": 1.5115700700826356, + "learning_rate": 1.7717962304244748e-07, + "loss": 0.6411, + "step": 5244 + }, + { + "epoch": 9.712962962962964, + "grad_norm": 1.5400094203294483, + "learning_rate": 1.771700877899641e-07, + "loss": 0.615, + "step": 5245 + }, + { + "epoch": 9.714814814814815, + "grad_norm": 1.775503954826172, + "learning_rate": 1.771605508024875e-07, + "loss": 0.6644, + "step": 5246 + }, + { + "epoch": 9.716666666666667, + "grad_norm": 1.6123763552366148, + "learning_rate": 1.771510120802322e-07, + "loss": 0.6777, + "step": 5247 + }, + { + "epoch": 9.718518518518518, + "grad_norm": 1.5934074967841956, + "learning_rate": 1.771414716234126e-07, + "loss": 0.66, + "step": 5248 + }, + { + "epoch": 9.72037037037037, + "grad_norm": 3.6390642150073647, + "learning_rate": 1.771319294322432e-07, + "loss": 0.6279, + "step": 5249 + }, + { + "epoch": 9.722222222222221, + "grad_norm": 1.4559192247007011, + "learning_rate": 1.7712238550693853e-07, + "loss": 0.6046, + "step": 5250 + }, + { + "epoch": 9.724074074074075, + "grad_norm": 1.5205214395705757, + "learning_rate": 1.7711283984771317e-07, + "loss": 0.6399, + "step": 5251 + }, + { + "epoch": 9.725925925925926, + "grad_norm": 1.3154845564893412, + "learning_rate": 1.7710329245478175e-07, + "loss": 0.6508, + "step": 5252 + }, + { + "epoch": 9.727777777777778, + "grad_norm": 1.7129914179371908, + "learning_rate": 1.7709374332835892e-07, + "loss": 0.6251, + "step": 5253 + }, + { + "epoch": 9.72962962962963, + "grad_norm": 1.4944424389958704, + "learning_rate": 1.7708419246865934e-07, + "loss": 0.6002, + "step": 5254 + }, + { + "epoch": 9.731481481481481, + "grad_norm": 1.5077483905996683, + "learning_rate": 1.7707463987589772e-07, + "loss": 0.6613, + "step": 5255 + }, + { + "epoch": 9.733333333333333, + "grad_norm": 1.630186030937466, + "learning_rate": 1.7706508555028893e-07, + "loss": 0.652, + "step": 5256 + }, + { + "epoch": 9.735185185185186, + "grad_norm": 1.5880493840753005, + "learning_rate": 1.7705552949204767e-07, + "loss": 0.6419, + "step": 5257 + }, + { + "epoch": 9.737037037037037, + "grad_norm": 1.5770507669124723, + "learning_rate": 1.7704597170138884e-07, + "loss": 0.6157, + "step": 5258 + }, + { + "epoch": 9.738888888888889, + "grad_norm": 1.4930190540130588, + "learning_rate": 1.770364121785273e-07, + "loss": 0.6493, + "step": 5259 + }, + { + "epoch": 9.74074074074074, + "grad_norm": 1.646377011549875, + "learning_rate": 1.77026850923678e-07, + "loss": 0.6613, + "step": 5260 + }, + { + "epoch": 9.742592592592592, + "grad_norm": 1.5063095099695083, + "learning_rate": 1.7701728793705589e-07, + "loss": 0.6197, + "step": 5261 + }, + { + "epoch": 9.744444444444444, + "grad_norm": 1.6553336471964604, + "learning_rate": 1.7700772321887595e-07, + "loss": 0.6186, + "step": 5262 + }, + { + "epoch": 9.746296296296297, + "grad_norm": 1.9790675332897267, + "learning_rate": 1.7699815676935327e-07, + "loss": 0.6633, + "step": 5263 + }, + { + "epoch": 9.748148148148148, + "grad_norm": 1.363777732735928, + "learning_rate": 1.769885885887029e-07, + "loss": 0.6359, + "step": 5264 + }, + { + "epoch": 9.75, + "grad_norm": 1.498978146140254, + "learning_rate": 1.7697901867713995e-07, + "loss": 0.6541, + "step": 5265 + }, + { + "epoch": 9.751851851851852, + "grad_norm": 1.6838159098378294, + "learning_rate": 1.7696944703487958e-07, + "loss": 0.6837, + "step": 5266 + }, + { + "epoch": 9.753703703703703, + "grad_norm": 1.514738306471442, + "learning_rate": 1.7695987366213702e-07, + "loss": 0.6426, + "step": 5267 + }, + { + "epoch": 9.755555555555556, + "grad_norm": 1.6546117840693506, + "learning_rate": 1.7695029855912743e-07, + "loss": 0.6833, + "step": 5268 + }, + { + "epoch": 9.757407407407408, + "grad_norm": 1.4651313483032045, + "learning_rate": 1.7694072172606618e-07, + "loss": 0.6115, + "step": 5269 + }, + { + "epoch": 9.75925925925926, + "grad_norm": 2.0002075217696738, + "learning_rate": 1.7693114316316854e-07, + "loss": 0.6583, + "step": 5270 + }, + { + "epoch": 9.761111111111111, + "grad_norm": 1.7746366776187028, + "learning_rate": 1.7692156287064984e-07, + "loss": 0.6356, + "step": 5271 + }, + { + "epoch": 9.762962962962963, + "grad_norm": 1.786737727680658, + "learning_rate": 1.769119808487255e-07, + "loss": 0.6106, + "step": 5272 + }, + { + "epoch": 9.764814814814814, + "grad_norm": 1.464309271783353, + "learning_rate": 1.7690239709761095e-07, + "loss": 0.5845, + "step": 5273 + }, + { + "epoch": 9.766666666666667, + "grad_norm": 1.5374527532805757, + "learning_rate": 1.7689281161752164e-07, + "loss": 0.5869, + "step": 5274 + }, + { + "epoch": 9.768518518518519, + "grad_norm": 1.7620044736693015, + "learning_rate": 1.7688322440867307e-07, + "loss": 0.6231, + "step": 5275 + }, + { + "epoch": 9.77037037037037, + "grad_norm": 1.4999844246670662, + "learning_rate": 1.768736354712808e-07, + "loss": 0.6649, + "step": 5276 + }, + { + "epoch": 9.772222222222222, + "grad_norm": 1.5764728308769347, + "learning_rate": 1.7686404480556046e-07, + "loss": 0.66, + "step": 5277 + }, + { + "epoch": 9.774074074074074, + "grad_norm": 1.886952361432808, + "learning_rate": 1.768544524117276e-07, + "loss": 0.6631, + "step": 5278 + }, + { + "epoch": 9.775925925925925, + "grad_norm": 1.5460929826942942, + "learning_rate": 1.7684485828999794e-07, + "loss": 0.6338, + "step": 5279 + }, + { + "epoch": 9.777777777777779, + "grad_norm": 2.708974725338093, + "learning_rate": 1.7683526244058716e-07, + "loss": 0.6425, + "step": 5280 + }, + { + "epoch": 9.77962962962963, + "grad_norm": 1.7331154549128358, + "learning_rate": 1.7682566486371094e-07, + "loss": 0.6274, + "step": 5281 + }, + { + "epoch": 9.781481481481482, + "grad_norm": 1.6884322541823826, + "learning_rate": 1.7681606555958518e-07, + "loss": 0.6599, + "step": 5282 + }, + { + "epoch": 9.783333333333333, + "grad_norm": 1.7651416967728226, + "learning_rate": 1.7680646452842562e-07, + "loss": 0.6081, + "step": 5283 + }, + { + "epoch": 9.785185185185185, + "grad_norm": 1.740057331475001, + "learning_rate": 1.7679686177044814e-07, + "loss": 0.627, + "step": 5284 + }, + { + "epoch": 9.787037037037036, + "grad_norm": 1.5136167931913198, + "learning_rate": 1.7678725728586863e-07, + "loss": 0.6463, + "step": 5285 + }, + { + "epoch": 9.78888888888889, + "grad_norm": 1.5109049800713026, + "learning_rate": 1.7677765107490302e-07, + "loss": 0.652, + "step": 5286 + }, + { + "epoch": 9.790740740740741, + "grad_norm": 1.6276218339291435, + "learning_rate": 1.767680431377673e-07, + "loss": 0.6248, + "step": 5287 + }, + { + "epoch": 9.792592592592593, + "grad_norm": 2.7170698614671567, + "learning_rate": 1.7675843347467746e-07, + "loss": 0.6184, + "step": 5288 + }, + { + "epoch": 9.794444444444444, + "grad_norm": 1.4623660631081234, + "learning_rate": 1.7674882208584955e-07, + "loss": 0.6526, + "step": 5289 + }, + { + "epoch": 9.796296296296296, + "grad_norm": 1.521351286985686, + "learning_rate": 1.7673920897149967e-07, + "loss": 0.6159, + "step": 5290 + }, + { + "epoch": 9.798148148148147, + "grad_norm": 1.4630503565355908, + "learning_rate": 1.76729594131844e-07, + "loss": 0.6392, + "step": 5291 + }, + { + "epoch": 9.8, + "grad_norm": 1.7630952834722897, + "learning_rate": 1.767199775670986e-07, + "loss": 0.6093, + "step": 5292 + }, + { + "epoch": 9.801851851851852, + "grad_norm": 1.4478057493671301, + "learning_rate": 1.7671035927747977e-07, + "loss": 0.6181, + "step": 5293 + }, + { + "epoch": 9.803703703703704, + "grad_norm": 1.5197541075930006, + "learning_rate": 1.767007392632037e-07, + "loss": 0.622, + "step": 5294 + }, + { + "epoch": 9.805555555555555, + "grad_norm": 1.8091658952086973, + "learning_rate": 1.7669111752448672e-07, + "loss": 0.6152, + "step": 5295 + }, + { + "epoch": 9.807407407407407, + "grad_norm": 1.7136907115795952, + "learning_rate": 1.766814940615451e-07, + "loss": 0.6496, + "step": 5296 + }, + { + "epoch": 9.809259259259258, + "grad_norm": 1.6788840747635472, + "learning_rate": 1.7667186887459527e-07, + "loss": 0.6679, + "step": 5297 + }, + { + "epoch": 9.811111111111112, + "grad_norm": 1.7258284364674372, + "learning_rate": 1.7666224196385359e-07, + "loss": 0.6687, + "step": 5298 + }, + { + "epoch": 9.812962962962963, + "grad_norm": 1.697097113947268, + "learning_rate": 1.7665261332953646e-07, + "loss": 0.5883, + "step": 5299 + }, + { + "epoch": 9.814814814814815, + "grad_norm": 1.4382860883142714, + "learning_rate": 1.766429829718604e-07, + "loss": 0.6054, + "step": 5300 + }, + { + "epoch": 9.816666666666666, + "grad_norm": 1.7931926426564386, + "learning_rate": 1.7663335089104194e-07, + "loss": 0.6849, + "step": 5301 + }, + { + "epoch": 9.818518518518518, + "grad_norm": 1.4821293743274302, + "learning_rate": 1.766237170872976e-07, + "loss": 0.6523, + "step": 5302 + }, + { + "epoch": 9.82037037037037, + "grad_norm": 1.7843876403444132, + "learning_rate": 1.7661408156084402e-07, + "loss": 0.5702, + "step": 5303 + }, + { + "epoch": 9.822222222222223, + "grad_norm": 1.4007354210390233, + "learning_rate": 1.766044443118978e-07, + "loss": 0.6183, + "step": 5304 + }, + { + "epoch": 9.824074074074074, + "grad_norm": 1.4357521960087263, + "learning_rate": 1.765948053406756e-07, + "loss": 0.5976, + "step": 5305 + }, + { + "epoch": 9.825925925925926, + "grad_norm": 1.5725788271690366, + "learning_rate": 1.7658516464739417e-07, + "loss": 0.5929, + "step": 5306 + }, + { + "epoch": 9.827777777777778, + "grad_norm": 4.937413932776677, + "learning_rate": 1.7657552223227024e-07, + "loss": 0.634, + "step": 5307 + }, + { + "epoch": 9.829629629629629, + "grad_norm": 1.5047120246494943, + "learning_rate": 1.7656587809552054e-07, + "loss": 0.6355, + "step": 5308 + }, + { + "epoch": 9.831481481481482, + "grad_norm": 1.5846261773078147, + "learning_rate": 1.7655623223736204e-07, + "loss": 0.6933, + "step": 5309 + }, + { + "epoch": 9.833333333333334, + "grad_norm": 1.6748626802572621, + "learning_rate": 1.7654658465801146e-07, + "loss": 0.5682, + "step": 5310 + }, + { + "epoch": 9.835185185185185, + "grad_norm": 1.5915870294780337, + "learning_rate": 1.7653693535768578e-07, + "loss": 0.5975, + "step": 5311 + }, + { + "epoch": 9.837037037037037, + "grad_norm": 1.5562064018839896, + "learning_rate": 1.7652728433660187e-07, + "loss": 0.6757, + "step": 5312 + }, + { + "epoch": 9.838888888888889, + "grad_norm": 1.6453119304021864, + "learning_rate": 1.765176315949768e-07, + "loss": 0.608, + "step": 5313 + }, + { + "epoch": 9.84074074074074, + "grad_norm": 1.4980455477625951, + "learning_rate": 1.7650797713302756e-07, + "loss": 0.6574, + "step": 5314 + }, + { + "epoch": 9.842592592592592, + "grad_norm": 1.8328229344616933, + "learning_rate": 1.7649832095097121e-07, + "loss": 0.6033, + "step": 5315 + }, + { + "epoch": 9.844444444444445, + "grad_norm": 1.3542069895684337, + "learning_rate": 1.7648866304902482e-07, + "loss": 0.6447, + "step": 5316 + }, + { + "epoch": 9.846296296296297, + "grad_norm": 1.707694633535408, + "learning_rate": 1.7647900342740556e-07, + "loss": 0.6693, + "step": 5317 + }, + { + "epoch": 9.848148148148148, + "grad_norm": 2.0291500123416286, + "learning_rate": 1.7646934208633056e-07, + "loss": 0.6256, + "step": 5318 + }, + { + "epoch": 9.85, + "grad_norm": 1.7503395391193268, + "learning_rate": 1.7645967902601707e-07, + "loss": 0.6492, + "step": 5319 + }, + { + "epoch": 9.851851851851851, + "grad_norm": 5.864940814403709, + "learning_rate": 1.7645001424668235e-07, + "loss": 0.6285, + "step": 5320 + }, + { + "epoch": 9.853703703703705, + "grad_norm": 2.0773261135265457, + "learning_rate": 1.764403477485437e-07, + "loss": 0.6434, + "step": 5321 + }, + { + "epoch": 9.855555555555556, + "grad_norm": 1.699094600802781, + "learning_rate": 1.7643067953181837e-07, + "loss": 0.6079, + "step": 5322 + }, + { + "epoch": 9.857407407407408, + "grad_norm": 1.4861082914699397, + "learning_rate": 1.764210095967238e-07, + "loss": 0.6191, + "step": 5323 + }, + { + "epoch": 9.85925925925926, + "grad_norm": 1.577006047602114, + "learning_rate": 1.764113379434774e-07, + "loss": 0.6471, + "step": 5324 + }, + { + "epoch": 9.86111111111111, + "grad_norm": 1.9236590728273362, + "learning_rate": 1.7640166457229656e-07, + "loss": 0.6571, + "step": 5325 + }, + { + "epoch": 9.862962962962962, + "grad_norm": 1.5246499120812889, + "learning_rate": 1.763919894833988e-07, + "loss": 0.6233, + "step": 5326 + }, + { + "epoch": 9.864814814814816, + "grad_norm": 1.8307438622079537, + "learning_rate": 1.763823126770017e-07, + "loss": 0.6496, + "step": 5327 + }, + { + "epoch": 9.866666666666667, + "grad_norm": 1.7739520451367077, + "learning_rate": 1.763726341533227e-07, + "loss": 0.6594, + "step": 5328 + }, + { + "epoch": 9.868518518518519, + "grad_norm": 1.3932948409685997, + "learning_rate": 1.7636295391257944e-07, + "loss": 0.6342, + "step": 5329 + }, + { + "epoch": 9.87037037037037, + "grad_norm": 1.497762520048915, + "learning_rate": 1.7635327195498962e-07, + "loss": 0.6341, + "step": 5330 + }, + { + "epoch": 9.872222222222222, + "grad_norm": 1.4184599020485498, + "learning_rate": 1.7634358828077088e-07, + "loss": 0.6494, + "step": 5331 + }, + { + "epoch": 9.874074074074073, + "grad_norm": 1.663299106333982, + "learning_rate": 1.763339028901409e-07, + "loss": 0.6449, + "step": 5332 + }, + { + "epoch": 9.875925925925927, + "grad_norm": 1.7195664615453008, + "learning_rate": 1.7632421578331748e-07, + "loss": 0.6277, + "step": 5333 + }, + { + "epoch": 9.877777777777778, + "grad_norm": 1.6271712449034368, + "learning_rate": 1.763145269605184e-07, + "loss": 0.6341, + "step": 5334 + }, + { + "epoch": 9.87962962962963, + "grad_norm": 1.4758485475156362, + "learning_rate": 1.7630483642196148e-07, + "loss": 0.6476, + "step": 5335 + }, + { + "epoch": 9.881481481481481, + "grad_norm": 1.6705914011332588, + "learning_rate": 1.7629514416786456e-07, + "loss": 0.7132, + "step": 5336 + }, + { + "epoch": 9.883333333333333, + "grad_norm": 1.6184902163359418, + "learning_rate": 1.7628545019844564e-07, + "loss": 0.6398, + "step": 5337 + }, + { + "epoch": 9.885185185185184, + "grad_norm": 1.4582106948595206, + "learning_rate": 1.762757545139226e-07, + "loss": 0.6158, + "step": 5338 + }, + { + "epoch": 9.887037037037038, + "grad_norm": 1.6877987475704237, + "learning_rate": 1.7626605711451342e-07, + "loss": 0.6593, + "step": 5339 + }, + { + "epoch": 9.88888888888889, + "grad_norm": 1.8651588210336025, + "learning_rate": 1.7625635800043616e-07, + "loss": 0.6497, + "step": 5340 + }, + { + "epoch": 9.89074074074074, + "grad_norm": 1.6340075736002309, + "learning_rate": 1.7624665717190883e-07, + "loss": 0.6583, + "step": 5341 + }, + { + "epoch": 9.892592592592592, + "grad_norm": 4.53238246778444, + "learning_rate": 1.762369546291496e-07, + "loss": 0.621, + "step": 5342 + }, + { + "epoch": 9.894444444444444, + "grad_norm": 1.6127496253864746, + "learning_rate": 1.7622725037237656e-07, + "loss": 0.6421, + "step": 5343 + }, + { + "epoch": 9.896296296296295, + "grad_norm": 1.394107095517648, + "learning_rate": 1.762175444018079e-07, + "loss": 0.5696, + "step": 5344 + }, + { + "epoch": 9.898148148148149, + "grad_norm": 1.5695701952721104, + "learning_rate": 1.7620783671766182e-07, + "loss": 0.665, + "step": 5345 + }, + { + "epoch": 9.9, + "grad_norm": 1.5916270431232358, + "learning_rate": 1.7619812732015663e-07, + "loss": 0.6309, + "step": 5346 + }, + { + "epoch": 9.901851851851852, + "grad_norm": 1.5027975880451432, + "learning_rate": 1.7618841620951054e-07, + "loss": 0.6487, + "step": 5347 + }, + { + "epoch": 9.903703703703703, + "grad_norm": 1.527701969204879, + "learning_rate": 1.7617870338594197e-07, + "loss": 0.6271, + "step": 5348 + }, + { + "epoch": 9.905555555555555, + "grad_norm": 1.4788529705550135, + "learning_rate": 1.761689888496692e-07, + "loss": 0.663, + "step": 5349 + }, + { + "epoch": 9.907407407407408, + "grad_norm": 1.724145661614413, + "learning_rate": 1.7615927260091076e-07, + "loss": 0.6544, + "step": 5350 + }, + { + "epoch": 9.90925925925926, + "grad_norm": 1.634704755671979, + "learning_rate": 1.76149554639885e-07, + "loss": 0.6182, + "step": 5351 + }, + { + "epoch": 9.911111111111111, + "grad_norm": 1.5429822541157592, + "learning_rate": 1.7613983496681041e-07, + "loss": 0.6391, + "step": 5352 + }, + { + "epoch": 9.912962962962963, + "grad_norm": 1.6321498855605314, + "learning_rate": 1.7613011358190554e-07, + "loss": 0.5995, + "step": 5353 + }, + { + "epoch": 9.914814814814815, + "grad_norm": 1.5008649767888658, + "learning_rate": 1.76120390485389e-07, + "loss": 0.6234, + "step": 5354 + }, + { + "epoch": 9.916666666666666, + "grad_norm": 3.3425832365486925, + "learning_rate": 1.761106656774793e-07, + "loss": 0.6615, + "step": 5355 + }, + { + "epoch": 9.918518518518518, + "grad_norm": 1.8997812292990122, + "learning_rate": 1.7610093915839512e-07, + "loss": 0.6791, + "step": 5356 + }, + { + "epoch": 9.920370370370371, + "grad_norm": 1.6848479657868247, + "learning_rate": 1.7609121092835513e-07, + "loss": 0.6052, + "step": 5357 + }, + { + "epoch": 9.922222222222222, + "grad_norm": 1.631792916206395, + "learning_rate": 1.760814809875781e-07, + "loss": 0.6534, + "step": 5358 + }, + { + "epoch": 9.924074074074074, + "grad_norm": 1.6183607352929394, + "learning_rate": 1.760717493362827e-07, + "loss": 0.6203, + "step": 5359 + }, + { + "epoch": 9.925925925925926, + "grad_norm": 1.6124872318896195, + "learning_rate": 1.760620159746878e-07, + "loss": 0.6402, + "step": 5360 + }, + { + "epoch": 9.927777777777777, + "grad_norm": 1.4943692515199258, + "learning_rate": 1.760522809030122e-07, + "loss": 0.6118, + "step": 5361 + }, + { + "epoch": 9.92962962962963, + "grad_norm": 1.777777824796479, + "learning_rate": 1.7604254412147474e-07, + "loss": 0.5952, + "step": 5362 + }, + { + "epoch": 9.931481481481482, + "grad_norm": 1.5167330317720857, + "learning_rate": 1.7603280563029437e-07, + "loss": 0.6375, + "step": 5363 + }, + { + "epoch": 9.933333333333334, + "grad_norm": 1.5923467687297315, + "learning_rate": 1.7602306542969003e-07, + "loss": 0.6558, + "step": 5364 + }, + { + "epoch": 9.935185185185185, + "grad_norm": 1.7759369249187156, + "learning_rate": 1.7601332351988072e-07, + "loss": 0.6209, + "step": 5365 + }, + { + "epoch": 9.937037037037037, + "grad_norm": 1.5618602035154268, + "learning_rate": 1.760035799010854e-07, + "loss": 0.5908, + "step": 5366 + }, + { + "epoch": 9.938888888888888, + "grad_norm": 1.4925859812994535, + "learning_rate": 1.7599383457352324e-07, + "loss": 0.6082, + "step": 5367 + }, + { + "epoch": 9.940740740740742, + "grad_norm": 1.4609494948842388, + "learning_rate": 1.7598408753741326e-07, + "loss": 0.5975, + "step": 5368 + }, + { + "epoch": 9.942592592592593, + "grad_norm": 1.4664494305030327, + "learning_rate": 1.7597433879297464e-07, + "loss": 0.6382, + "step": 5369 + }, + { + "epoch": 9.944444444444445, + "grad_norm": 1.7799782387798435, + "learning_rate": 1.759645883404265e-07, + "loss": 0.6191, + "step": 5370 + }, + { + "epoch": 9.946296296296296, + "grad_norm": 1.8259353449534388, + "learning_rate": 1.7595483617998813e-07, + "loss": 0.6341, + "step": 5371 + }, + { + "epoch": 9.948148148148148, + "grad_norm": 1.6952519825780845, + "learning_rate": 1.7594508231187873e-07, + "loss": 0.6485, + "step": 5372 + }, + { + "epoch": 9.95, + "grad_norm": 2.308087360435426, + "learning_rate": 1.7593532673631763e-07, + "loss": 0.6199, + "step": 5373 + }, + { + "epoch": 9.951851851851853, + "grad_norm": 1.874540168542581, + "learning_rate": 1.7592556945352413e-07, + "loss": 0.6188, + "step": 5374 + }, + { + "epoch": 9.953703703703704, + "grad_norm": 1.5275759330369276, + "learning_rate": 1.7591581046371763e-07, + "loss": 0.616, + "step": 5375 + }, + { + "epoch": 9.955555555555556, + "grad_norm": 1.616211587389306, + "learning_rate": 1.7590604976711754e-07, + "loss": 0.6447, + "step": 5376 + }, + { + "epoch": 9.957407407407407, + "grad_norm": 2.034094775213003, + "learning_rate": 1.7589628736394328e-07, + "loss": 0.5575, + "step": 5377 + }, + { + "epoch": 9.959259259259259, + "grad_norm": 1.6730283451438421, + "learning_rate": 1.7588652325441435e-07, + "loss": 0.6389, + "step": 5378 + }, + { + "epoch": 9.96111111111111, + "grad_norm": 1.6651520189680318, + "learning_rate": 1.7587675743875028e-07, + "loss": 0.6511, + "step": 5379 + }, + { + "epoch": 9.962962962962964, + "grad_norm": 1.6902678548932415, + "learning_rate": 1.7586698991717062e-07, + "loss": 0.6404, + "step": 5380 + }, + { + "epoch": 9.964814814814815, + "grad_norm": 1.9416466745535363, + "learning_rate": 1.7585722068989498e-07, + "loss": 0.6295, + "step": 5381 + }, + { + "epoch": 9.966666666666667, + "grad_norm": 1.995507662113881, + "learning_rate": 1.75847449757143e-07, + "loss": 0.6474, + "step": 5382 + }, + { + "epoch": 9.968518518518518, + "grad_norm": 1.6857697214829845, + "learning_rate": 1.7583767711913432e-07, + "loss": 0.6113, + "step": 5383 + }, + { + "epoch": 9.97037037037037, + "grad_norm": 1.7062100986463136, + "learning_rate": 1.7582790277608872e-07, + "loss": 0.6494, + "step": 5384 + }, + { + "epoch": 9.972222222222221, + "grad_norm": 1.651946582790938, + "learning_rate": 1.7581812672822588e-07, + "loss": 0.6111, + "step": 5385 + }, + { + "epoch": 9.974074074074075, + "grad_norm": 2.1863709450469253, + "learning_rate": 1.7580834897576568e-07, + "loss": 0.6293, + "step": 5386 + }, + { + "epoch": 9.975925925925926, + "grad_norm": 1.6077756891565993, + "learning_rate": 1.7579856951892788e-07, + "loss": 0.6269, + "step": 5387 + }, + { + "epoch": 9.977777777777778, + "grad_norm": 2.0861880204839744, + "learning_rate": 1.7578878835793237e-07, + "loss": 0.5877, + "step": 5388 + }, + { + "epoch": 9.97962962962963, + "grad_norm": 1.5138865055336816, + "learning_rate": 1.7577900549299906e-07, + "loss": 0.616, + "step": 5389 + }, + { + "epoch": 9.981481481481481, + "grad_norm": 1.724550504678632, + "learning_rate": 1.7576922092434789e-07, + "loss": 0.6512, + "step": 5390 + }, + { + "epoch": 9.983333333333333, + "grad_norm": 1.3473731699165248, + "learning_rate": 1.7575943465219884e-07, + "loss": 0.6763, + "step": 5391 + }, + { + "epoch": 9.985185185185186, + "grad_norm": 1.5434756588619276, + "learning_rate": 1.7574964667677197e-07, + "loss": 0.6308, + "step": 5392 + }, + { + "epoch": 9.987037037037037, + "grad_norm": 1.512171513091537, + "learning_rate": 1.757398569982873e-07, + "loss": 0.5866, + "step": 5393 + }, + { + "epoch": 9.988888888888889, + "grad_norm": 1.6069055533257244, + "learning_rate": 1.757300656169649e-07, + "loss": 0.6236, + "step": 5394 + }, + { + "epoch": 9.99074074074074, + "grad_norm": 1.5165261896035422, + "learning_rate": 1.75720272533025e-07, + "loss": 0.6161, + "step": 5395 + }, + { + "epoch": 9.992592592592592, + "grad_norm": 1.4634293951237676, + "learning_rate": 1.757104777466877e-07, + "loss": 0.6208, + "step": 5396 + }, + { + "epoch": 9.994444444444444, + "grad_norm": 1.560741807941215, + "learning_rate": 1.757006812581732e-07, + "loss": 0.6292, + "step": 5397 + }, + { + "epoch": 9.996296296296297, + "grad_norm": 1.7143380232528396, + "learning_rate": 1.7569088306770184e-07, + "loss": 0.6235, + "step": 5398 + }, + { + "epoch": 9.998148148148148, + "grad_norm": 1.5106003236497454, + "learning_rate": 1.7568108317549385e-07, + "loss": 0.6211, + "step": 5399 + }, + { + "epoch": 10.0, + "grad_norm": 1.437281630754344, + "learning_rate": 1.7567128158176952e-07, + "loss": 0.6278, + "step": 5400 + }, + { + "epoch": 10.001851851851852, + "grad_norm": 1.7098573916799773, + "learning_rate": 1.7566147828674928e-07, + "loss": 0.641, + "step": 5401 + }, + { + "epoch": 10.003703703703703, + "grad_norm": 1.762501013213143, + "learning_rate": 1.756516732906535e-07, + "loss": 0.6129, + "step": 5402 + }, + { + "epoch": 10.005555555555556, + "grad_norm": 1.8107878864228215, + "learning_rate": 1.7564186659370267e-07, + "loss": 0.6388, + "step": 5403 + }, + { + "epoch": 10.007407407407408, + "grad_norm": 1.8437931273681092, + "learning_rate": 1.756320581961172e-07, + "loss": 0.6488, + "step": 5404 + }, + { + "epoch": 10.00925925925926, + "grad_norm": 1.399536817733659, + "learning_rate": 1.7562224809811768e-07, + "loss": 0.6201, + "step": 5405 + }, + { + "epoch": 10.011111111111111, + "grad_norm": 1.3749930273296522, + "learning_rate": 1.7561243629992464e-07, + "loss": 0.6225, + "step": 5406 + }, + { + "epoch": 10.012962962962963, + "grad_norm": 1.6733135468429279, + "learning_rate": 1.7560262280175868e-07, + "loss": 0.6356, + "step": 5407 + }, + { + "epoch": 10.014814814814814, + "grad_norm": 1.3853374438282582, + "learning_rate": 1.7559280760384038e-07, + "loss": 0.6259, + "step": 5408 + }, + { + "epoch": 10.016666666666667, + "grad_norm": 1.7394206880810161, + "learning_rate": 1.7558299070639052e-07, + "loss": 0.6159, + "step": 5409 + }, + { + "epoch": 10.018518518518519, + "grad_norm": 1.9959962114385614, + "learning_rate": 1.755731721096297e-07, + "loss": 0.6591, + "step": 5410 + }, + { + "epoch": 10.02037037037037, + "grad_norm": 3.1104791556069147, + "learning_rate": 1.7556335181377872e-07, + "loss": 0.6816, + "step": 5411 + }, + { + "epoch": 10.022222222222222, + "grad_norm": 1.569045766373436, + "learning_rate": 1.7555352981905838e-07, + "loss": 0.6263, + "step": 5412 + }, + { + "epoch": 10.024074074074074, + "grad_norm": 1.574548976842346, + "learning_rate": 1.7554370612568946e-07, + "loss": 0.6361, + "step": 5413 + }, + { + "epoch": 10.025925925925925, + "grad_norm": 1.5966426102196793, + "learning_rate": 1.755338807338929e-07, + "loss": 0.6626, + "step": 5414 + }, + { + "epoch": 10.027777777777779, + "grad_norm": 1.6311327368399573, + "learning_rate": 1.7552405364388952e-07, + "loss": 0.6615, + "step": 5415 + }, + { + "epoch": 10.02962962962963, + "grad_norm": 1.4758553669940013, + "learning_rate": 1.755142248559003e-07, + "loss": 0.5891, + "step": 5416 + }, + { + "epoch": 10.031481481481482, + "grad_norm": 1.6698168419392883, + "learning_rate": 1.7550439437014623e-07, + "loss": 0.6646, + "step": 5417 + }, + { + "epoch": 10.033333333333333, + "grad_norm": 1.6827280407422223, + "learning_rate": 1.7549456218684832e-07, + "loss": 0.6154, + "step": 5418 + }, + { + "epoch": 10.035185185185185, + "grad_norm": 1.7781165273835078, + "learning_rate": 1.7548472830622756e-07, + "loss": 0.6568, + "step": 5419 + }, + { + "epoch": 10.037037037037036, + "grad_norm": 1.5133939721052163, + "learning_rate": 1.7547489272850512e-07, + "loss": 0.6142, + "step": 5420 + }, + { + "epoch": 10.03888888888889, + "grad_norm": 1.806437339783129, + "learning_rate": 1.7546505545390212e-07, + "loss": 0.6297, + "step": 5421 + }, + { + "epoch": 10.040740740740741, + "grad_norm": 1.5629938681275963, + "learning_rate": 1.7545521648263969e-07, + "loss": 0.6022, + "step": 5422 + }, + { + "epoch": 10.042592592592593, + "grad_norm": 1.9452022259578916, + "learning_rate": 1.7544537581493906e-07, + "loss": 0.5931, + "step": 5423 + }, + { + "epoch": 10.044444444444444, + "grad_norm": 1.5695491944405213, + "learning_rate": 1.754355334510215e-07, + "loss": 0.629, + "step": 5424 + }, + { + "epoch": 10.046296296296296, + "grad_norm": 1.5529083569681585, + "learning_rate": 1.7542568939110823e-07, + "loss": 0.62, + "step": 5425 + }, + { + "epoch": 10.048148148148147, + "grad_norm": 1.5028566603604283, + "learning_rate": 1.7541584363542064e-07, + "loss": 0.6398, + "step": 5426 + }, + { + "epoch": 10.05, + "grad_norm": 1.7213447199997993, + "learning_rate": 1.7540599618418005e-07, + "loss": 0.6043, + "step": 5427 + }, + { + "epoch": 10.051851851851852, + "grad_norm": 1.5159099790962234, + "learning_rate": 1.7539614703760786e-07, + "loss": 0.6235, + "step": 5428 + }, + { + "epoch": 10.053703703703704, + "grad_norm": 1.528433817159488, + "learning_rate": 1.7538629619592552e-07, + "loss": 0.6684, + "step": 5429 + }, + { + "epoch": 10.055555555555555, + "grad_norm": 1.5031113085458192, + "learning_rate": 1.753764436593545e-07, + "loss": 0.6907, + "step": 5430 + }, + { + "epoch": 10.057407407407407, + "grad_norm": 1.8304059969812667, + "learning_rate": 1.7536658942811632e-07, + "loss": 0.624, + "step": 5431 + }, + { + "epoch": 10.059259259259258, + "grad_norm": 1.606070419953307, + "learning_rate": 1.7535673350243245e-07, + "loss": 0.6269, + "step": 5432 + }, + { + "epoch": 10.061111111111112, + "grad_norm": 1.4545851842975484, + "learning_rate": 1.7534687588252462e-07, + "loss": 0.6631, + "step": 5433 + }, + { + "epoch": 10.062962962962963, + "grad_norm": 1.5851564348222162, + "learning_rate": 1.7533701656861434e-07, + "loss": 0.6289, + "step": 5434 + }, + { + "epoch": 10.064814814814815, + "grad_norm": 1.8128035180214417, + "learning_rate": 1.7532715556092335e-07, + "loss": 0.6251, + "step": 5435 + }, + { + "epoch": 10.066666666666666, + "grad_norm": 1.7227149687478553, + "learning_rate": 1.7531729285967328e-07, + "loss": 0.6262, + "step": 5436 + }, + { + "epoch": 10.068518518518518, + "grad_norm": 1.6464925960931711, + "learning_rate": 1.7530742846508593e-07, + "loss": 0.6109, + "step": 5437 + }, + { + "epoch": 10.07037037037037, + "grad_norm": 1.3926493573294632, + "learning_rate": 1.7529756237738305e-07, + "loss": 0.6139, + "step": 5438 + }, + { + "epoch": 10.072222222222223, + "grad_norm": 3.0896580145107557, + "learning_rate": 1.7528769459678646e-07, + "loss": 0.591, + "step": 5439 + }, + { + "epoch": 10.074074074074074, + "grad_norm": 1.6418635059843978, + "learning_rate": 1.7527782512351804e-07, + "loss": 0.6718, + "step": 5440 + }, + { + "epoch": 10.075925925925926, + "grad_norm": 1.380776230271832, + "learning_rate": 1.752679539577996e-07, + "loss": 0.6071, + "step": 5441 + }, + { + "epoch": 10.077777777777778, + "grad_norm": 1.5018014045666626, + "learning_rate": 1.752580810998532e-07, + "loss": 0.5957, + "step": 5442 + }, + { + "epoch": 10.079629629629629, + "grad_norm": 1.5644677745854392, + "learning_rate": 1.752482065499007e-07, + "loss": 0.6367, + "step": 5443 + }, + { + "epoch": 10.081481481481482, + "grad_norm": 1.589005607459597, + "learning_rate": 1.7523833030816414e-07, + "loss": 0.6243, + "step": 5444 + }, + { + "epoch": 10.083333333333334, + "grad_norm": 1.4444195485919817, + "learning_rate": 1.752284523748656e-07, + "loss": 0.6538, + "step": 5445 + }, + { + "epoch": 10.085185185185185, + "grad_norm": 1.5359609767188398, + "learning_rate": 1.7521857275022712e-07, + "loss": 0.5972, + "step": 5446 + }, + { + "epoch": 10.087037037037037, + "grad_norm": 1.410038322340428, + "learning_rate": 1.7520869143447084e-07, + "loss": 0.6285, + "step": 5447 + }, + { + "epoch": 10.088888888888889, + "grad_norm": 1.7825376721065744, + "learning_rate": 1.7519880842781892e-07, + "loss": 0.6047, + "step": 5448 + }, + { + "epoch": 10.09074074074074, + "grad_norm": 1.7384020488052006, + "learning_rate": 1.7518892373049353e-07, + "loss": 0.644, + "step": 5449 + }, + { + "epoch": 10.092592592592593, + "grad_norm": 1.6543661230294295, + "learning_rate": 1.7517903734271693e-07, + "loss": 0.6267, + "step": 5450 + }, + { + "epoch": 10.094444444444445, + "grad_norm": 1.539215619690721, + "learning_rate": 1.7516914926471138e-07, + "loss": 0.5882, + "step": 5451 + }, + { + "epoch": 10.096296296296297, + "grad_norm": 1.649971961981206, + "learning_rate": 1.7515925949669922e-07, + "loss": 0.6527, + "step": 5452 + }, + { + "epoch": 10.098148148148148, + "grad_norm": 1.3220548537596515, + "learning_rate": 1.7514936803890275e-07, + "loss": 0.6301, + "step": 5453 + }, + { + "epoch": 10.1, + "grad_norm": 1.5495174583103943, + "learning_rate": 1.751394748915444e-07, + "loss": 0.619, + "step": 5454 + }, + { + "epoch": 10.101851851851851, + "grad_norm": 1.531033106425347, + "learning_rate": 1.7512958005484657e-07, + "loss": 0.6549, + "step": 5455 + }, + { + "epoch": 10.103703703703705, + "grad_norm": 1.559076341876545, + "learning_rate": 1.7511968352903175e-07, + "loss": 0.6354, + "step": 5456 + }, + { + "epoch": 10.105555555555556, + "grad_norm": 1.768673995635982, + "learning_rate": 1.751097853143224e-07, + "loss": 0.6283, + "step": 5457 + }, + { + "epoch": 10.107407407407408, + "grad_norm": 1.45193003555617, + "learning_rate": 1.7509988541094108e-07, + "loss": 0.6338, + "step": 5458 + }, + { + "epoch": 10.10925925925926, + "grad_norm": 1.682710732335579, + "learning_rate": 1.750899838191104e-07, + "loss": 0.6195, + "step": 5459 + }, + { + "epoch": 10.11111111111111, + "grad_norm": 1.5266955215137943, + "learning_rate": 1.750800805390529e-07, + "loss": 0.6191, + "step": 5460 + }, + { + "epoch": 10.112962962962962, + "grad_norm": 1.6048594704958947, + "learning_rate": 1.750701755709913e-07, + "loss": 0.6392, + "step": 5461 + }, + { + "epoch": 10.114814814814816, + "grad_norm": 1.6260188220193799, + "learning_rate": 1.7506026891514827e-07, + "loss": 0.6452, + "step": 5462 + }, + { + "epoch": 10.116666666666667, + "grad_norm": 1.7308424331945567, + "learning_rate": 1.750503605717465e-07, + "loss": 0.5788, + "step": 5463 + }, + { + "epoch": 10.118518518518519, + "grad_norm": 1.956225410862961, + "learning_rate": 1.7504045054100884e-07, + "loss": 0.6594, + "step": 5464 + }, + { + "epoch": 10.12037037037037, + "grad_norm": 1.5532658265287898, + "learning_rate": 1.7503053882315802e-07, + "loss": 0.6293, + "step": 5465 + }, + { + "epoch": 10.122222222222222, + "grad_norm": 3.412774904691906, + "learning_rate": 1.7502062541841692e-07, + "loss": 0.6732, + "step": 5466 + }, + { + "epoch": 10.124074074074073, + "grad_norm": 1.5537896442815855, + "learning_rate": 1.7501071032700838e-07, + "loss": 0.6623, + "step": 5467 + }, + { + "epoch": 10.125925925925927, + "grad_norm": 1.4694556242385732, + "learning_rate": 1.7500079354915535e-07, + "loss": 0.6628, + "step": 5468 + }, + { + "epoch": 10.127777777777778, + "grad_norm": 1.6229229898959951, + "learning_rate": 1.749908750850808e-07, + "loss": 0.604, + "step": 5469 + }, + { + "epoch": 10.12962962962963, + "grad_norm": 1.792271862927219, + "learning_rate": 1.7498095493500774e-07, + "loss": 0.6358, + "step": 5470 + }, + { + "epoch": 10.131481481481481, + "grad_norm": 1.766618300118669, + "learning_rate": 1.749710330991591e-07, + "loss": 0.6485, + "step": 5471 + }, + { + "epoch": 10.133333333333333, + "grad_norm": 1.978909843463822, + "learning_rate": 1.7496110957775808e-07, + "loss": 0.6566, + "step": 5472 + }, + { + "epoch": 10.135185185185184, + "grad_norm": 1.5660534328838762, + "learning_rate": 1.7495118437102766e-07, + "loss": 0.6437, + "step": 5473 + }, + { + "epoch": 10.137037037037038, + "grad_norm": 1.8229889221917583, + "learning_rate": 1.7494125747919109e-07, + "loss": 0.6508, + "step": 5474 + }, + { + "epoch": 10.13888888888889, + "grad_norm": 1.6391646748274622, + "learning_rate": 1.749313289024715e-07, + "loss": 0.6481, + "step": 5475 + }, + { + "epoch": 10.14074074074074, + "grad_norm": 2.029009025041515, + "learning_rate": 1.7492139864109216e-07, + "loss": 0.6257, + "step": 5476 + }, + { + "epoch": 10.142592592592592, + "grad_norm": 1.3708176002505783, + "learning_rate": 1.7491146669527624e-07, + "loss": 0.6122, + "step": 5477 + }, + { + "epoch": 10.144444444444444, + "grad_norm": 1.5143378257760454, + "learning_rate": 1.7490153306524712e-07, + "loss": 0.6557, + "step": 5478 + }, + { + "epoch": 10.146296296296295, + "grad_norm": 1.7993450592830094, + "learning_rate": 1.7489159775122812e-07, + "loss": 0.6053, + "step": 5479 + }, + { + "epoch": 10.148148148148149, + "grad_norm": 1.824416286209922, + "learning_rate": 1.7488166075344258e-07, + "loss": 0.6475, + "step": 5480 + }, + { + "epoch": 10.15, + "grad_norm": 1.5359348155297259, + "learning_rate": 1.7487172207211393e-07, + "loss": 0.6438, + "step": 5481 + }, + { + "epoch": 10.151851851851852, + "grad_norm": 2.181821129056764, + "learning_rate": 1.7486178170746563e-07, + "loss": 0.6386, + "step": 5482 + }, + { + "epoch": 10.153703703703703, + "grad_norm": 1.4890011471993854, + "learning_rate": 1.7485183965972117e-07, + "loss": 0.6863, + "step": 5483 + }, + { + "epoch": 10.155555555555555, + "grad_norm": 1.8035843706339572, + "learning_rate": 1.7484189592910402e-07, + "loss": 0.6343, + "step": 5484 + }, + { + "epoch": 10.157407407407407, + "grad_norm": 1.4596870801261845, + "learning_rate": 1.7483195051583781e-07, + "loss": 0.6442, + "step": 5485 + }, + { + "epoch": 10.15925925925926, + "grad_norm": 1.6045386504073424, + "learning_rate": 1.748220034201461e-07, + "loss": 0.6483, + "step": 5486 + }, + { + "epoch": 10.161111111111111, + "grad_norm": 1.443354283286236, + "learning_rate": 1.7481205464225254e-07, + "loss": 0.6324, + "step": 5487 + }, + { + "epoch": 10.162962962962963, + "grad_norm": 2.5576780325241186, + "learning_rate": 1.7480210418238084e-07, + "loss": 0.6107, + "step": 5488 + }, + { + "epoch": 10.164814814814815, + "grad_norm": 1.6782656832253358, + "learning_rate": 1.7479215204075466e-07, + "loss": 0.6011, + "step": 5489 + }, + { + "epoch": 10.166666666666666, + "grad_norm": 1.497783018464314, + "learning_rate": 1.7478219821759775e-07, + "loss": 0.5917, + "step": 5490 + }, + { + "epoch": 10.168518518518518, + "grad_norm": 1.4028401969888447, + "learning_rate": 1.747722427131339e-07, + "loss": 0.5957, + "step": 5491 + }, + { + "epoch": 10.170370370370371, + "grad_norm": 1.7500533193866945, + "learning_rate": 1.74762285527587e-07, + "loss": 0.6411, + "step": 5492 + }, + { + "epoch": 10.172222222222222, + "grad_norm": 2.536000570561538, + "learning_rate": 1.7475232666118086e-07, + "loss": 0.6671, + "step": 5493 + }, + { + "epoch": 10.174074074074074, + "grad_norm": 1.6274632337016584, + "learning_rate": 1.747423661141394e-07, + "loss": 0.6275, + "step": 5494 + }, + { + "epoch": 10.175925925925926, + "grad_norm": 1.7027974976720113, + "learning_rate": 1.7473240388668654e-07, + "loss": 0.6125, + "step": 5495 + }, + { + "epoch": 10.177777777777777, + "grad_norm": 1.6481532436486679, + "learning_rate": 1.7472243997904626e-07, + "loss": 0.6063, + "step": 5496 + }, + { + "epoch": 10.17962962962963, + "grad_norm": 1.9632694961044845, + "learning_rate": 1.747124743914426e-07, + "loss": 0.6325, + "step": 5497 + }, + { + "epoch": 10.181481481481482, + "grad_norm": 1.6392790247435027, + "learning_rate": 1.747025071240996e-07, + "loss": 0.6108, + "step": 5498 + }, + { + "epoch": 10.183333333333334, + "grad_norm": 1.8209886790195473, + "learning_rate": 1.7469253817724133e-07, + "loss": 0.6047, + "step": 5499 + }, + { + "epoch": 10.185185185185185, + "grad_norm": 1.904186901142017, + "learning_rate": 1.7468256755109196e-07, + "loss": 0.5867, + "step": 5500 + }, + { + "epoch": 10.187037037037037, + "grad_norm": 1.5541576103066044, + "learning_rate": 1.7467259524587564e-07, + "loss": 0.6415, + "step": 5501 + }, + { + "epoch": 10.188888888888888, + "grad_norm": 2.507280815088853, + "learning_rate": 1.7466262126181657e-07, + "loss": 0.5996, + "step": 5502 + }, + { + "epoch": 10.190740740740742, + "grad_norm": 1.546303851731707, + "learning_rate": 1.7465264559913897e-07, + "loss": 0.6551, + "step": 5503 + }, + { + "epoch": 10.192592592592593, + "grad_norm": 1.5705319691122943, + "learning_rate": 1.7464266825806717e-07, + "loss": 0.6469, + "step": 5504 + }, + { + "epoch": 10.194444444444445, + "grad_norm": 1.5193132845643262, + "learning_rate": 1.7463268923882545e-07, + "loss": 0.6011, + "step": 5505 + }, + { + "epoch": 10.196296296296296, + "grad_norm": 1.6649974369870668, + "learning_rate": 1.746227085416382e-07, + "loss": 0.6289, + "step": 5506 + }, + { + "epoch": 10.198148148148148, + "grad_norm": 1.6219604889928134, + "learning_rate": 1.7461272616672974e-07, + "loss": 0.5924, + "step": 5507 + }, + { + "epoch": 10.2, + "grad_norm": 1.319202947556063, + "learning_rate": 1.746027421143246e-07, + "loss": 0.6358, + "step": 5508 + }, + { + "epoch": 10.201851851851853, + "grad_norm": 1.39331139402287, + "learning_rate": 1.7459275638464716e-07, + "loss": 0.591, + "step": 5509 + }, + { + "epoch": 10.203703703703704, + "grad_norm": 1.581059424479091, + "learning_rate": 1.74582768977922e-07, + "loss": 0.5983, + "step": 5510 + }, + { + "epoch": 10.205555555555556, + "grad_norm": 1.7827362833276263, + "learning_rate": 1.745727798943736e-07, + "loss": 0.6486, + "step": 5511 + }, + { + "epoch": 10.207407407407407, + "grad_norm": 1.676344083964432, + "learning_rate": 1.7456278913422659e-07, + "loss": 0.6447, + "step": 5512 + }, + { + "epoch": 10.209259259259259, + "grad_norm": 1.7267224663675524, + "learning_rate": 1.7455279669770558e-07, + "loss": 0.6147, + "step": 5513 + }, + { + "epoch": 10.21111111111111, + "grad_norm": 1.4654553691012164, + "learning_rate": 1.745428025850352e-07, + "loss": 0.6496, + "step": 5514 + }, + { + "epoch": 10.212962962962964, + "grad_norm": 2.2067321695075512, + "learning_rate": 1.7453280679644018e-07, + "loss": 0.6167, + "step": 5515 + }, + { + "epoch": 10.214814814814815, + "grad_norm": 1.5442336004459, + "learning_rate": 1.7452280933214523e-07, + "loss": 0.6437, + "step": 5516 + }, + { + "epoch": 10.216666666666667, + "grad_norm": 1.6302117104545533, + "learning_rate": 1.745128101923751e-07, + "loss": 0.6386, + "step": 5517 + }, + { + "epoch": 10.218518518518518, + "grad_norm": 1.863647711142828, + "learning_rate": 1.7450280937735464e-07, + "loss": 0.6131, + "step": 5518 + }, + { + "epoch": 10.22037037037037, + "grad_norm": 1.6803911691655162, + "learning_rate": 1.7449280688730869e-07, + "loss": 0.6177, + "step": 5519 + }, + { + "epoch": 10.222222222222221, + "grad_norm": 1.4533503352272066, + "learning_rate": 1.744828027224621e-07, + "loss": 0.6401, + "step": 5520 + }, + { + "epoch": 10.224074074074075, + "grad_norm": 4.717111465419407, + "learning_rate": 1.7447279688303982e-07, + "loss": 0.6392, + "step": 5521 + }, + { + "epoch": 10.225925925925926, + "grad_norm": 1.4044209547715456, + "learning_rate": 1.7446278936926681e-07, + "loss": 0.6368, + "step": 5522 + }, + { + "epoch": 10.227777777777778, + "grad_norm": 1.8145224553152064, + "learning_rate": 1.7445278018136805e-07, + "loss": 0.6355, + "step": 5523 + }, + { + "epoch": 10.22962962962963, + "grad_norm": 1.421038230452516, + "learning_rate": 1.744427693195686e-07, + "loss": 0.611, + "step": 5524 + }, + { + "epoch": 10.231481481481481, + "grad_norm": 1.3321516046857718, + "learning_rate": 1.7443275678409348e-07, + "loss": 0.5833, + "step": 5525 + }, + { + "epoch": 10.233333333333333, + "grad_norm": 1.4950411109701227, + "learning_rate": 1.7442274257516781e-07, + "loss": 0.5672, + "step": 5526 + }, + { + "epoch": 10.235185185185186, + "grad_norm": 1.5837934135649152, + "learning_rate": 1.7441272669301684e-07, + "loss": 0.6305, + "step": 5527 + }, + { + "epoch": 10.237037037037037, + "grad_norm": 1.5805286526469449, + "learning_rate": 1.744027091378656e-07, + "loss": 0.6363, + "step": 5528 + }, + { + "epoch": 10.238888888888889, + "grad_norm": 2.5909401288928353, + "learning_rate": 1.7439268990993944e-07, + "loss": 0.646, + "step": 5529 + }, + { + "epoch": 10.24074074074074, + "grad_norm": 1.3892380470013994, + "learning_rate": 1.7438266900946353e-07, + "loss": 0.6475, + "step": 5530 + }, + { + "epoch": 10.242592592592592, + "grad_norm": 1.7218547567376672, + "learning_rate": 1.7437264643666317e-07, + "loss": 0.6554, + "step": 5531 + }, + { + "epoch": 10.244444444444444, + "grad_norm": 1.9632540575265303, + "learning_rate": 1.7436262219176378e-07, + "loss": 0.649, + "step": 5532 + }, + { + "epoch": 10.246296296296297, + "grad_norm": 1.862987607175111, + "learning_rate": 1.7435259627499064e-07, + "loss": 0.5927, + "step": 5533 + }, + { + "epoch": 10.248148148148148, + "grad_norm": 1.5017592493813212, + "learning_rate": 1.7434256868656924e-07, + "loss": 0.6283, + "step": 5534 + }, + { + "epoch": 10.25, + "grad_norm": 1.7186720218181681, + "learning_rate": 1.7433253942672495e-07, + "loss": 0.6077, + "step": 5535 + }, + { + "epoch": 10.251851851851852, + "grad_norm": 1.4885062582660054, + "learning_rate": 1.743225084956833e-07, + "loss": 0.6157, + "step": 5536 + }, + { + "epoch": 10.253703703703703, + "grad_norm": 1.80912790417594, + "learning_rate": 1.743124758936698e-07, + "loss": 0.6485, + "step": 5537 + }, + { + "epoch": 10.255555555555556, + "grad_norm": 1.3966421337815214, + "learning_rate": 1.7430244162091002e-07, + "loss": 0.6237, + "step": 5538 + }, + { + "epoch": 10.257407407407408, + "grad_norm": 1.5076718384621615, + "learning_rate": 1.7429240567762953e-07, + "loss": 0.6205, + "step": 5539 + }, + { + "epoch": 10.25925925925926, + "grad_norm": 1.6343518892967013, + "learning_rate": 1.74282368064054e-07, + "loss": 0.6438, + "step": 5540 + }, + { + "epoch": 10.261111111111111, + "grad_norm": 1.6166409071831247, + "learning_rate": 1.742723287804091e-07, + "loss": 0.6507, + "step": 5541 + }, + { + "epoch": 10.262962962962963, + "grad_norm": 1.5338285153498954, + "learning_rate": 1.742622878269205e-07, + "loss": 0.6277, + "step": 5542 + }, + { + "epoch": 10.264814814814814, + "grad_norm": 1.5009842496230708, + "learning_rate": 1.7425224520381397e-07, + "loss": 0.6639, + "step": 5543 + }, + { + "epoch": 10.266666666666667, + "grad_norm": 1.5348752243529402, + "learning_rate": 1.7424220091131535e-07, + "loss": 0.661, + "step": 5544 + }, + { + "epoch": 10.268518518518519, + "grad_norm": 3.1096101725094165, + "learning_rate": 1.7423215494965038e-07, + "loss": 0.6, + "step": 5545 + }, + { + "epoch": 10.27037037037037, + "grad_norm": 1.582307403119614, + "learning_rate": 1.7422210731904497e-07, + "loss": 0.6297, + "step": 5546 + }, + { + "epoch": 10.272222222222222, + "grad_norm": 1.589105049232147, + "learning_rate": 1.7421205801972499e-07, + "loss": 0.7048, + "step": 5547 + }, + { + "epoch": 10.274074074074074, + "grad_norm": 1.5981124117530174, + "learning_rate": 1.7420200705191638e-07, + "loss": 0.6651, + "step": 5548 + }, + { + "epoch": 10.275925925925925, + "grad_norm": 1.8703227544827403, + "learning_rate": 1.7419195441584513e-07, + "loss": 0.6231, + "step": 5549 + }, + { + "epoch": 10.277777777777779, + "grad_norm": 1.472798494260182, + "learning_rate": 1.7418190011173728e-07, + "loss": 0.6568, + "step": 5550 + }, + { + "epoch": 10.27962962962963, + "grad_norm": 4.887250444622802, + "learning_rate": 1.7417184413981877e-07, + "loss": 0.6312, + "step": 5551 + }, + { + "epoch": 10.281481481481482, + "grad_norm": 1.4889889626658819, + "learning_rate": 1.741617865003158e-07, + "loss": 0.64, + "step": 5552 + }, + { + "epoch": 10.283333333333333, + "grad_norm": 1.686091022171425, + "learning_rate": 1.7415172719345446e-07, + "loss": 0.6263, + "step": 5553 + }, + { + "epoch": 10.285185185185185, + "grad_norm": 1.664668886009631, + "learning_rate": 1.741416662194609e-07, + "loss": 0.6431, + "step": 5554 + }, + { + "epoch": 10.287037037037036, + "grad_norm": 1.7054550805628932, + "learning_rate": 1.741316035785613e-07, + "loss": 0.6317, + "step": 5555 + }, + { + "epoch": 10.28888888888889, + "grad_norm": 1.4083990332617717, + "learning_rate": 1.741215392709819e-07, + "loss": 0.6087, + "step": 5556 + }, + { + "epoch": 10.290740740740741, + "grad_norm": 2.4208191751102075, + "learning_rate": 1.7411147329694898e-07, + "loss": 0.6478, + "step": 5557 + }, + { + "epoch": 10.292592592592593, + "grad_norm": 1.7109823064464107, + "learning_rate": 1.741014056566889e-07, + "loss": 0.6182, + "step": 5558 + }, + { + "epoch": 10.294444444444444, + "grad_norm": 1.5950325601201951, + "learning_rate": 1.7409133635042795e-07, + "loss": 0.656, + "step": 5559 + }, + { + "epoch": 10.296296296296296, + "grad_norm": 1.4501350924335195, + "learning_rate": 1.740812653783925e-07, + "loss": 0.6322, + "step": 5560 + }, + { + "epoch": 10.298148148148147, + "grad_norm": 1.6473791712265817, + "learning_rate": 1.7407119274080904e-07, + "loss": 0.6046, + "step": 5561 + }, + { + "epoch": 10.3, + "grad_norm": 1.568230333855012, + "learning_rate": 1.7406111843790398e-07, + "loss": 0.639, + "step": 5562 + }, + { + "epoch": 10.301851851851852, + "grad_norm": 1.996711631565388, + "learning_rate": 1.740510424699038e-07, + "loss": 0.6124, + "step": 5563 + }, + { + "epoch": 10.303703703703704, + "grad_norm": 1.4386105777575093, + "learning_rate": 1.7404096483703512e-07, + "loss": 0.671, + "step": 5564 + }, + { + "epoch": 10.305555555555555, + "grad_norm": 1.506472839373819, + "learning_rate": 1.7403088553952442e-07, + "loss": 0.6233, + "step": 5565 + }, + { + "epoch": 10.307407407407407, + "grad_norm": 1.5168896943004355, + "learning_rate": 1.7402080457759836e-07, + "loss": 0.6057, + "step": 5566 + }, + { + "epoch": 10.309259259259258, + "grad_norm": 1.6397689211594215, + "learning_rate": 1.7401072195148357e-07, + "loss": 0.6669, + "step": 5567 + }, + { + "epoch": 10.311111111111112, + "grad_norm": 1.8173012835702376, + "learning_rate": 1.7400063766140677e-07, + "loss": 0.6176, + "step": 5568 + }, + { + "epoch": 10.312962962962963, + "grad_norm": 1.8232472211201778, + "learning_rate": 1.7399055170759462e-07, + "loss": 0.6303, + "step": 5569 + }, + { + "epoch": 10.314814814814815, + "grad_norm": 1.8704698502710055, + "learning_rate": 1.739804640902739e-07, + "loss": 0.6187, + "step": 5570 + }, + { + "epoch": 10.316666666666666, + "grad_norm": 1.447568043820789, + "learning_rate": 1.7397037480967146e-07, + "loss": 0.6411, + "step": 5571 + }, + { + "epoch": 10.318518518518518, + "grad_norm": 1.84325159171906, + "learning_rate": 1.7396028386601408e-07, + "loss": 0.6292, + "step": 5572 + }, + { + "epoch": 10.32037037037037, + "grad_norm": 1.8953526264542464, + "learning_rate": 1.7395019125952864e-07, + "loss": 0.6542, + "step": 5573 + }, + { + "epoch": 10.322222222222223, + "grad_norm": 1.776440025858486, + "learning_rate": 1.7394009699044204e-07, + "loss": 0.6341, + "step": 5574 + }, + { + "epoch": 10.324074074074074, + "grad_norm": 1.3759567847625795, + "learning_rate": 1.7393000105898126e-07, + "loss": 0.6077, + "step": 5575 + }, + { + "epoch": 10.325925925925926, + "grad_norm": 1.718473701447944, + "learning_rate": 1.7391990346537327e-07, + "loss": 0.6443, + "step": 5576 + }, + { + "epoch": 10.327777777777778, + "grad_norm": 3.1139820921655565, + "learning_rate": 1.7390980420984506e-07, + "loss": 0.6682, + "step": 5577 + }, + { + "epoch": 10.329629629629629, + "grad_norm": 2.1036378025468214, + "learning_rate": 1.7389970329262375e-07, + "loss": 0.6181, + "step": 5578 + }, + { + "epoch": 10.331481481481482, + "grad_norm": 1.697639162940773, + "learning_rate": 1.7388960071393637e-07, + "loss": 0.6514, + "step": 5579 + }, + { + "epoch": 10.333333333333334, + "grad_norm": 1.5413222475807067, + "learning_rate": 1.738794964740101e-07, + "loss": 0.6014, + "step": 5580 + }, + { + "epoch": 10.335185185185185, + "grad_norm": 1.4884563975506548, + "learning_rate": 1.738693905730721e-07, + "loss": 0.5975, + "step": 5581 + }, + { + "epoch": 10.337037037037037, + "grad_norm": 1.4849875083875392, + "learning_rate": 1.7385928301134957e-07, + "loss": 0.6463, + "step": 5582 + }, + { + "epoch": 10.338888888888889, + "grad_norm": 1.7043338941610162, + "learning_rate": 1.7384917378906977e-07, + "loss": 0.5953, + "step": 5583 + }, + { + "epoch": 10.34074074074074, + "grad_norm": 1.4547009229947307, + "learning_rate": 1.7383906290645994e-07, + "loss": 0.6641, + "step": 5584 + }, + { + "epoch": 10.342592592592593, + "grad_norm": 1.6012708629009933, + "learning_rate": 1.7382895036374746e-07, + "loss": 0.6118, + "step": 5585 + }, + { + "epoch": 10.344444444444445, + "grad_norm": 1.4828069656544804, + "learning_rate": 1.7381883616115966e-07, + "loss": 0.6028, + "step": 5586 + }, + { + "epoch": 10.346296296296297, + "grad_norm": 1.4452112374201271, + "learning_rate": 1.7380872029892392e-07, + "loss": 0.5754, + "step": 5587 + }, + { + "epoch": 10.348148148148148, + "grad_norm": 1.5868306913319499, + "learning_rate": 1.7379860277726768e-07, + "loss": 0.5982, + "step": 5588 + }, + { + "epoch": 10.35, + "grad_norm": 1.7519571473136228, + "learning_rate": 1.7378848359641845e-07, + "loss": 0.6333, + "step": 5589 + }, + { + "epoch": 10.351851851851851, + "grad_norm": 1.7493028509881299, + "learning_rate": 1.7377836275660368e-07, + "loss": 0.5886, + "step": 5590 + }, + { + "epoch": 10.353703703703705, + "grad_norm": 1.711461949245045, + "learning_rate": 1.737682402580509e-07, + "loss": 0.6272, + "step": 5591 + }, + { + "epoch": 10.355555555555556, + "grad_norm": 1.5582610946205975, + "learning_rate": 1.7375811610098778e-07, + "loss": 0.644, + "step": 5592 + }, + { + "epoch": 10.357407407407408, + "grad_norm": 1.8660506883667933, + "learning_rate": 1.737479902856419e-07, + "loss": 0.6436, + "step": 5593 + }, + { + "epoch": 10.35925925925926, + "grad_norm": 1.5297975489859659, + "learning_rate": 1.7373786281224082e-07, + "loss": 0.6, + "step": 5594 + }, + { + "epoch": 10.36111111111111, + "grad_norm": 1.4456178552997858, + "learning_rate": 1.737277336810124e-07, + "loss": 0.617, + "step": 5595 + }, + { + "epoch": 10.362962962962962, + "grad_norm": 1.6777621052576792, + "learning_rate": 1.7371760289218426e-07, + "loss": 0.5846, + "step": 5596 + }, + { + "epoch": 10.364814814814816, + "grad_norm": 3.0504269261398242, + "learning_rate": 1.7370747044598417e-07, + "loss": 0.6568, + "step": 5597 + }, + { + "epoch": 10.366666666666667, + "grad_norm": 1.5898361692025798, + "learning_rate": 1.7369733634264e-07, + "loss": 0.6459, + "step": 5598 + }, + { + "epoch": 10.368518518518519, + "grad_norm": 1.5635668893132577, + "learning_rate": 1.736872005823795e-07, + "loss": 0.6297, + "step": 5599 + }, + { + "epoch": 10.37037037037037, + "grad_norm": 1.5461290969141985, + "learning_rate": 1.736770631654306e-07, + "loss": 0.621, + "step": 5600 + }, + { + "epoch": 10.372222222222222, + "grad_norm": 1.4832596952264288, + "learning_rate": 1.7366692409202127e-07, + "loss": 0.6278, + "step": 5601 + }, + { + "epoch": 10.374074074074073, + "grad_norm": 1.8431309223150518, + "learning_rate": 1.736567833623794e-07, + "loss": 0.6339, + "step": 5602 + }, + { + "epoch": 10.375925925925927, + "grad_norm": 1.453667585698071, + "learning_rate": 1.7364664097673295e-07, + "loss": 0.6327, + "step": 5603 + }, + { + "epoch": 10.377777777777778, + "grad_norm": 1.4787439998930816, + "learning_rate": 1.7363649693530999e-07, + "loss": 0.5829, + "step": 5604 + }, + { + "epoch": 10.37962962962963, + "grad_norm": 1.815058250165841, + "learning_rate": 1.736263512383386e-07, + "loss": 0.6344, + "step": 5605 + }, + { + "epoch": 10.381481481481481, + "grad_norm": 1.5130607145217372, + "learning_rate": 1.7361620388604687e-07, + "loss": 0.6057, + "step": 5606 + }, + { + "epoch": 10.383333333333333, + "grad_norm": 1.5348777031692438, + "learning_rate": 1.7360605487866297e-07, + "loss": 0.6713, + "step": 5607 + }, + { + "epoch": 10.385185185185184, + "grad_norm": 1.7045118441962281, + "learning_rate": 1.7359590421641502e-07, + "loss": 0.5957, + "step": 5608 + }, + { + "epoch": 10.387037037037038, + "grad_norm": 1.6201407612969965, + "learning_rate": 1.7358575189953123e-07, + "loss": 0.6239, + "step": 5609 + }, + { + "epoch": 10.38888888888889, + "grad_norm": 1.8643119793238843, + "learning_rate": 1.7357559792823991e-07, + "loss": 0.6457, + "step": 5610 + }, + { + "epoch": 10.39074074074074, + "grad_norm": 2.1230210261776192, + "learning_rate": 1.7356544230276933e-07, + "loss": 0.6267, + "step": 5611 + }, + { + "epoch": 10.392592592592592, + "grad_norm": 1.6433140724899655, + "learning_rate": 1.735552850233478e-07, + "loss": 0.6299, + "step": 5612 + }, + { + "epoch": 10.394444444444444, + "grad_norm": 1.5505719414011734, + "learning_rate": 1.7354512609020365e-07, + "loss": 0.6614, + "step": 5613 + }, + { + "epoch": 10.396296296296295, + "grad_norm": 1.6085290625304194, + "learning_rate": 1.7353496550356538e-07, + "loss": 0.6335, + "step": 5614 + }, + { + "epoch": 10.398148148148149, + "grad_norm": 1.7753758012329388, + "learning_rate": 1.7352480326366133e-07, + "loss": 0.6409, + "step": 5615 + }, + { + "epoch": 10.4, + "grad_norm": 1.6355605303039746, + "learning_rate": 1.7351463937072005e-07, + "loss": 0.6592, + "step": 5616 + }, + { + "epoch": 10.401851851851852, + "grad_norm": 1.3998280358780555, + "learning_rate": 1.7350447382496997e-07, + "loss": 0.6363, + "step": 5617 + }, + { + "epoch": 10.403703703703703, + "grad_norm": 1.6072157041710209, + "learning_rate": 1.7349430662663972e-07, + "loss": 0.6104, + "step": 5618 + }, + { + "epoch": 10.405555555555555, + "grad_norm": 1.4306611968293306, + "learning_rate": 1.7348413777595782e-07, + "loss": 0.6098, + "step": 5619 + }, + { + "epoch": 10.407407407407407, + "grad_norm": 1.804048104237655, + "learning_rate": 1.7347396727315293e-07, + "loss": 0.6416, + "step": 5620 + }, + { + "epoch": 10.40925925925926, + "grad_norm": 1.3984142985830783, + "learning_rate": 1.7346379511845371e-07, + "loss": 0.6243, + "step": 5621 + }, + { + "epoch": 10.411111111111111, + "grad_norm": 1.4859437010383754, + "learning_rate": 1.7345362131208885e-07, + "loss": 0.6492, + "step": 5622 + }, + { + "epoch": 10.412962962962963, + "grad_norm": 1.6291128754008786, + "learning_rate": 1.7344344585428707e-07, + "loss": 0.6321, + "step": 5623 + }, + { + "epoch": 10.414814814814815, + "grad_norm": 1.8993033776711115, + "learning_rate": 1.7343326874527717e-07, + "loss": 0.6746, + "step": 5624 + }, + { + "epoch": 10.416666666666666, + "grad_norm": 1.4486317926151744, + "learning_rate": 1.7342308998528795e-07, + "loss": 0.5651, + "step": 5625 + }, + { + "epoch": 10.418518518518518, + "grad_norm": 1.3376030482538848, + "learning_rate": 1.7341290957454825e-07, + "loss": 0.6434, + "step": 5626 + }, + { + "epoch": 10.420370370370371, + "grad_norm": 1.6244110671107352, + "learning_rate": 1.7340272751328695e-07, + "loss": 0.5995, + "step": 5627 + }, + { + "epoch": 10.422222222222222, + "grad_norm": 1.6323174797127578, + "learning_rate": 1.73392543801733e-07, + "loss": 0.6639, + "step": 5628 + }, + { + "epoch": 10.424074074074074, + "grad_norm": 2.0721136553374593, + "learning_rate": 1.7338235844011534e-07, + "loss": 0.6388, + "step": 5629 + }, + { + "epoch": 10.425925925925926, + "grad_norm": 1.7090250952487094, + "learning_rate": 1.7337217142866292e-07, + "loss": 0.6143, + "step": 5630 + }, + { + "epoch": 10.427777777777777, + "grad_norm": 1.874663675977847, + "learning_rate": 1.7336198276760484e-07, + "loss": 0.6422, + "step": 5631 + }, + { + "epoch": 10.42962962962963, + "grad_norm": 1.989424696624551, + "learning_rate": 1.7335179245717011e-07, + "loss": 0.6638, + "step": 5632 + }, + { + "epoch": 10.431481481481482, + "grad_norm": 1.4182583820806682, + "learning_rate": 1.7334160049758789e-07, + "loss": 0.596, + "step": 5633 + }, + { + "epoch": 10.433333333333334, + "grad_norm": 1.7317402853825656, + "learning_rate": 1.7333140688908728e-07, + "loss": 0.6233, + "step": 5634 + }, + { + "epoch": 10.435185185185185, + "grad_norm": 1.9737827564139214, + "learning_rate": 1.7332121163189748e-07, + "loss": 0.6323, + "step": 5635 + }, + { + "epoch": 10.437037037037037, + "grad_norm": 1.6008264958018399, + "learning_rate": 1.7331101472624772e-07, + "loss": 0.6386, + "step": 5636 + }, + { + "epoch": 10.438888888888888, + "grad_norm": 1.4406854334249448, + "learning_rate": 1.7330081617236725e-07, + "loss": 0.6482, + "step": 5637 + }, + { + "epoch": 10.440740740740742, + "grad_norm": 1.403411125963815, + "learning_rate": 1.7329061597048532e-07, + "loss": 0.618, + "step": 5638 + }, + { + "epoch": 10.442592592592593, + "grad_norm": 1.8285387661774546, + "learning_rate": 1.732804141208313e-07, + "loss": 0.6322, + "step": 5639 + }, + { + "epoch": 10.444444444444445, + "grad_norm": 1.6346942011361483, + "learning_rate": 1.7327021062363456e-07, + "loss": 0.6356, + "step": 5640 + }, + { + "epoch": 10.446296296296296, + "grad_norm": 1.4195442215071123, + "learning_rate": 1.7326000547912447e-07, + "loss": 0.6489, + "step": 5641 + }, + { + "epoch": 10.448148148148148, + "grad_norm": 1.4869905493495832, + "learning_rate": 1.732497986875305e-07, + "loss": 0.5884, + "step": 5642 + }, + { + "epoch": 10.45, + "grad_norm": 1.7577182872924617, + "learning_rate": 1.7323959024908209e-07, + "loss": 0.6078, + "step": 5643 + }, + { + "epoch": 10.451851851851853, + "grad_norm": 1.4224877283071675, + "learning_rate": 1.732293801640088e-07, + "loss": 0.5932, + "step": 5644 + }, + { + "epoch": 10.453703703703704, + "grad_norm": 1.3154210498915635, + "learning_rate": 1.732191684325401e-07, + "loss": 0.6175, + "step": 5645 + }, + { + "epoch": 10.455555555555556, + "grad_norm": 1.634533795488109, + "learning_rate": 1.732089550549057e-07, + "loss": 0.6838, + "step": 5646 + }, + { + "epoch": 10.457407407407407, + "grad_norm": 1.4610794154473281, + "learning_rate": 1.7319874003133513e-07, + "loss": 0.614, + "step": 5647 + }, + { + "epoch": 10.459259259259259, + "grad_norm": 1.4141051217475848, + "learning_rate": 1.7318852336205806e-07, + "loss": 0.6489, + "step": 5648 + }, + { + "epoch": 10.46111111111111, + "grad_norm": 1.557943309573647, + "learning_rate": 1.7317830504730426e-07, + "loss": 0.6452, + "step": 5649 + }, + { + "epoch": 10.462962962962964, + "grad_norm": 1.474957456195557, + "learning_rate": 1.7316808508730336e-07, + "loss": 0.6344, + "step": 5650 + }, + { + "epoch": 10.464814814814815, + "grad_norm": 2.331331729557835, + "learning_rate": 1.7315786348228523e-07, + "loss": 0.6216, + "step": 5651 + }, + { + "epoch": 10.466666666666667, + "grad_norm": 1.4983931908397468, + "learning_rate": 1.731476402324796e-07, + "loss": 0.641, + "step": 5652 + }, + { + "epoch": 10.468518518518518, + "grad_norm": 2.9008750975051107, + "learning_rate": 1.7313741533811638e-07, + "loss": 0.6115, + "step": 5653 + }, + { + "epoch": 10.47037037037037, + "grad_norm": 1.6952039861295856, + "learning_rate": 1.7312718879942541e-07, + "loss": 0.6399, + "step": 5654 + }, + { + "epoch": 10.472222222222221, + "grad_norm": 1.489424301538423, + "learning_rate": 1.7311696061663664e-07, + "loss": 0.5911, + "step": 5655 + }, + { + "epoch": 10.474074074074075, + "grad_norm": 1.7277111543420562, + "learning_rate": 1.7310673078997997e-07, + "loss": 0.6915, + "step": 5656 + }, + { + "epoch": 10.475925925925926, + "grad_norm": 1.6628745760137444, + "learning_rate": 1.7309649931968547e-07, + "loss": 0.6315, + "step": 5657 + }, + { + "epoch": 10.477777777777778, + "grad_norm": 1.533700252751984, + "learning_rate": 1.7308626620598314e-07, + "loss": 0.6502, + "step": 5658 + }, + { + "epoch": 10.47962962962963, + "grad_norm": 1.66732880869917, + "learning_rate": 1.7307603144910305e-07, + "loss": 0.6413, + "step": 5659 + }, + { + "epoch": 10.481481481481481, + "grad_norm": 1.520279876691348, + "learning_rate": 1.730657950492753e-07, + "loss": 0.6429, + "step": 5660 + }, + { + "epoch": 10.483333333333333, + "grad_norm": 1.4445882224270583, + "learning_rate": 1.7305555700673001e-07, + "loss": 0.6364, + "step": 5661 + }, + { + "epoch": 10.485185185185186, + "grad_norm": 1.8669137043784276, + "learning_rate": 1.7304531732169738e-07, + "loss": 0.5668, + "step": 5662 + }, + { + "epoch": 10.487037037037037, + "grad_norm": 1.6118966459197406, + "learning_rate": 1.7303507599440765e-07, + "loss": 0.6213, + "step": 5663 + }, + { + "epoch": 10.488888888888889, + "grad_norm": 1.4928326356671953, + "learning_rate": 1.7302483302509105e-07, + "loss": 0.6851, + "step": 5664 + }, + { + "epoch": 10.49074074074074, + "grad_norm": 1.590635808259245, + "learning_rate": 1.7301458841397788e-07, + "loss": 0.6228, + "step": 5665 + }, + { + "epoch": 10.492592592592592, + "grad_norm": 2.1337903928787227, + "learning_rate": 1.7300434216129847e-07, + "loss": 0.6367, + "step": 5666 + }, + { + "epoch": 10.494444444444444, + "grad_norm": 1.598801441389825, + "learning_rate": 1.7299409426728314e-07, + "loss": 0.6091, + "step": 5667 + }, + { + "epoch": 10.496296296296297, + "grad_norm": 1.5868704317155193, + "learning_rate": 1.729838447321623e-07, + "loss": 0.6569, + "step": 5668 + }, + { + "epoch": 10.498148148148148, + "grad_norm": 1.4258910093065258, + "learning_rate": 1.7297359355616644e-07, + "loss": 0.6369, + "step": 5669 + }, + { + "epoch": 10.5, + "grad_norm": 1.3472595043845652, + "learning_rate": 1.7296334073952604e-07, + "loss": 0.5985, + "step": 5670 + }, + { + "epoch": 10.501851851851852, + "grad_norm": 2.3170841769906643, + "learning_rate": 1.729530862824715e-07, + "loss": 0.6263, + "step": 5671 + }, + { + "epoch": 10.503703703703703, + "grad_norm": 1.7231266880447955, + "learning_rate": 1.7294283018523351e-07, + "loss": 0.5966, + "step": 5672 + }, + { + "epoch": 10.505555555555556, + "grad_norm": 1.5144459103973558, + "learning_rate": 1.7293257244804256e-07, + "loss": 0.6486, + "step": 5673 + }, + { + "epoch": 10.507407407407408, + "grad_norm": 1.594242607675054, + "learning_rate": 1.729223130711293e-07, + "loss": 0.6169, + "step": 5674 + }, + { + "epoch": 10.50925925925926, + "grad_norm": 1.936084630040173, + "learning_rate": 1.729120520547244e-07, + "loss": 0.6232, + "step": 5675 + }, + { + "epoch": 10.511111111111111, + "grad_norm": 1.571595714329208, + "learning_rate": 1.729017893990585e-07, + "loss": 0.6038, + "step": 5676 + }, + { + "epoch": 10.512962962962963, + "grad_norm": 1.3521668243150566, + "learning_rate": 1.7289152510436243e-07, + "loss": 0.6291, + "step": 5677 + }, + { + "epoch": 10.514814814814814, + "grad_norm": 1.6619348080724383, + "learning_rate": 1.7288125917086686e-07, + "loss": 0.6381, + "step": 5678 + }, + { + "epoch": 10.516666666666667, + "grad_norm": 1.6376784420357435, + "learning_rate": 1.7287099159880264e-07, + "loss": 0.589, + "step": 5679 + }, + { + "epoch": 10.518518518518519, + "grad_norm": 1.625498230622435, + "learning_rate": 1.7286072238840066e-07, + "loss": 0.5898, + "step": 5680 + }, + { + "epoch": 10.52037037037037, + "grad_norm": 3.0222733102603594, + "learning_rate": 1.7285045153989172e-07, + "loss": 0.6327, + "step": 5681 + }, + { + "epoch": 10.522222222222222, + "grad_norm": 2.762616949259456, + "learning_rate": 1.7284017905350677e-07, + "loss": 0.5789, + "step": 5682 + }, + { + "epoch": 10.524074074074074, + "grad_norm": 1.6590225967526497, + "learning_rate": 1.728299049294768e-07, + "loss": 0.5932, + "step": 5683 + }, + { + "epoch": 10.525925925925925, + "grad_norm": 1.8747846440555545, + "learning_rate": 1.7281962916803274e-07, + "loss": 0.5865, + "step": 5684 + }, + { + "epoch": 10.527777777777779, + "grad_norm": 1.9302440237974134, + "learning_rate": 1.7280935176940563e-07, + "loss": 0.612, + "step": 5685 + }, + { + "epoch": 10.52962962962963, + "grad_norm": 1.6468961884768185, + "learning_rate": 1.7279907273382658e-07, + "loss": 0.6485, + "step": 5686 + }, + { + "epoch": 10.531481481481482, + "grad_norm": 1.639521751589659, + "learning_rate": 1.727887920615266e-07, + "loss": 0.6149, + "step": 5687 + }, + { + "epoch": 10.533333333333333, + "grad_norm": 1.6145716487603634, + "learning_rate": 1.7277850975273693e-07, + "loss": 0.6682, + "step": 5688 + }, + { + "epoch": 10.535185185185185, + "grad_norm": 1.6953885478437642, + "learning_rate": 1.7276822580768871e-07, + "loss": 0.6303, + "step": 5689 + }, + { + "epoch": 10.537037037037036, + "grad_norm": 1.6567657720319768, + "learning_rate": 1.727579402266131e-07, + "loss": 0.6471, + "step": 5690 + }, + { + "epoch": 10.53888888888889, + "grad_norm": 1.5145919834696557, + "learning_rate": 1.7274765300974137e-07, + "loss": 0.6277, + "step": 5691 + }, + { + "epoch": 10.540740740740741, + "grad_norm": 1.6714727130434472, + "learning_rate": 1.7273736415730486e-07, + "loss": 0.688, + "step": 5692 + }, + { + "epoch": 10.542592592592593, + "grad_norm": 1.4287889166568395, + "learning_rate": 1.7272707366953485e-07, + "loss": 0.5754, + "step": 5693 + }, + { + "epoch": 10.544444444444444, + "grad_norm": 1.5359363884638486, + "learning_rate": 1.7271678154666267e-07, + "loss": 0.6593, + "step": 5694 + }, + { + "epoch": 10.546296296296296, + "grad_norm": 1.6102839148142356, + "learning_rate": 1.7270648778891977e-07, + "loss": 0.6192, + "step": 5695 + }, + { + "epoch": 10.548148148148147, + "grad_norm": 1.5233105207479953, + "learning_rate": 1.7269619239653756e-07, + "loss": 0.6478, + "step": 5696 + }, + { + "epoch": 10.55, + "grad_norm": 1.467582711316881, + "learning_rate": 1.726858953697475e-07, + "loss": 0.628, + "step": 5697 + }, + { + "epoch": 10.551851851851852, + "grad_norm": 1.8767352136281532, + "learning_rate": 1.726755967087811e-07, + "loss": 0.6115, + "step": 5698 + }, + { + "epoch": 10.553703703703704, + "grad_norm": 1.4990376673647625, + "learning_rate": 1.7266529641386989e-07, + "loss": 0.6566, + "step": 5699 + }, + { + "epoch": 10.555555555555555, + "grad_norm": 1.569918657540948, + "learning_rate": 1.7265499448524547e-07, + "loss": 0.6156, + "step": 5700 + }, + { + "epoch": 10.557407407407407, + "grad_norm": 1.3450302475744602, + "learning_rate": 1.7264469092313945e-07, + "loss": 0.6125, + "step": 5701 + }, + { + "epoch": 10.559259259259258, + "grad_norm": 1.475919061188164, + "learning_rate": 1.7263438572778347e-07, + "loss": 0.5925, + "step": 5702 + }, + { + "epoch": 10.561111111111112, + "grad_norm": 1.4708315566287036, + "learning_rate": 1.7262407889940925e-07, + "loss": 0.6282, + "step": 5703 + }, + { + "epoch": 10.562962962962963, + "grad_norm": 1.6717141845646533, + "learning_rate": 1.7261377043824844e-07, + "loss": 0.6169, + "step": 5704 + }, + { + "epoch": 10.564814814814815, + "grad_norm": 1.6198375480838834, + "learning_rate": 1.726034603445329e-07, + "loss": 0.6206, + "step": 5705 + }, + { + "epoch": 10.566666666666666, + "grad_norm": 1.636936084274294, + "learning_rate": 1.7259314861849438e-07, + "loss": 0.6336, + "step": 5706 + }, + { + "epoch": 10.568518518518518, + "grad_norm": 1.487970724137033, + "learning_rate": 1.725828352603647e-07, + "loss": 0.6703, + "step": 5707 + }, + { + "epoch": 10.57037037037037, + "grad_norm": 2.3176179585290857, + "learning_rate": 1.7257252027037577e-07, + "loss": 0.6244, + "step": 5708 + }, + { + "epoch": 10.572222222222223, + "grad_norm": 1.550963673680878, + "learning_rate": 1.7256220364875945e-07, + "loss": 0.6422, + "step": 5709 + }, + { + "epoch": 10.574074074074074, + "grad_norm": 1.5167872392552564, + "learning_rate": 1.7255188539574774e-07, + "loss": 0.6428, + "step": 5710 + }, + { + "epoch": 10.575925925925926, + "grad_norm": 1.5793100816140422, + "learning_rate": 1.7254156551157258e-07, + "loss": 0.5963, + "step": 5711 + }, + { + "epoch": 10.577777777777778, + "grad_norm": 1.4774129154914482, + "learning_rate": 1.7253124399646603e-07, + "loss": 0.5963, + "step": 5712 + }, + { + "epoch": 10.579629629629629, + "grad_norm": 1.5535667891951814, + "learning_rate": 1.7252092085066012e-07, + "loss": 0.6168, + "step": 5713 + }, + { + "epoch": 10.581481481481482, + "grad_norm": 1.536798729302369, + "learning_rate": 1.7251059607438696e-07, + "loss": 0.6407, + "step": 5714 + }, + { + "epoch": 10.583333333333334, + "grad_norm": 1.8388621059463646, + "learning_rate": 1.7250026966787866e-07, + "loss": 0.616, + "step": 5715 + }, + { + "epoch": 10.585185185185185, + "grad_norm": 1.5777272704786476, + "learning_rate": 1.7248994163136735e-07, + "loss": 0.6322, + "step": 5716 + }, + { + "epoch": 10.587037037037037, + "grad_norm": 1.7491826442959557, + "learning_rate": 1.7247961196508532e-07, + "loss": 0.6216, + "step": 5717 + }, + { + "epoch": 10.588888888888889, + "grad_norm": 1.493476496291092, + "learning_rate": 1.7246928066926473e-07, + "loss": 0.5683, + "step": 5718 + }, + { + "epoch": 10.59074074074074, + "grad_norm": 1.4899634244819258, + "learning_rate": 1.724589477441379e-07, + "loss": 0.6373, + "step": 5719 + }, + { + "epoch": 10.592592592592592, + "grad_norm": 1.7165190032007915, + "learning_rate": 1.7244861318993712e-07, + "loss": 0.6147, + "step": 5720 + }, + { + "epoch": 10.594444444444445, + "grad_norm": 1.4410333142919904, + "learning_rate": 1.7243827700689474e-07, + "loss": 0.6469, + "step": 5721 + }, + { + "epoch": 10.596296296296297, + "grad_norm": 1.5160481873400484, + "learning_rate": 1.724279391952432e-07, + "loss": 0.6318, + "step": 5722 + }, + { + "epoch": 10.598148148148148, + "grad_norm": 1.8621485526873551, + "learning_rate": 1.7241759975521484e-07, + "loss": 0.6337, + "step": 5723 + }, + { + "epoch": 10.6, + "grad_norm": 1.6650670094474247, + "learning_rate": 1.7240725868704216e-07, + "loss": 0.6251, + "step": 5724 + }, + { + "epoch": 10.601851851851851, + "grad_norm": 1.482307189758289, + "learning_rate": 1.7239691599095762e-07, + "loss": 0.6694, + "step": 5725 + }, + { + "epoch": 10.603703703703705, + "grad_norm": 1.581866869248674, + "learning_rate": 1.7238657166719382e-07, + "loss": 0.6438, + "step": 5726 + }, + { + "epoch": 10.605555555555556, + "grad_norm": 1.4675099285895001, + "learning_rate": 1.7237622571598327e-07, + "loss": 0.6506, + "step": 5727 + }, + { + "epoch": 10.607407407407408, + "grad_norm": 1.7920583366206484, + "learning_rate": 1.7236587813755863e-07, + "loss": 0.6525, + "step": 5728 + }, + { + "epoch": 10.60925925925926, + "grad_norm": 1.4665089772309414, + "learning_rate": 1.7235552893215247e-07, + "loss": 0.6294, + "step": 5729 + }, + { + "epoch": 10.61111111111111, + "grad_norm": 1.466159662623476, + "learning_rate": 1.7234517809999752e-07, + "loss": 0.6321, + "step": 5730 + }, + { + "epoch": 10.612962962962962, + "grad_norm": 1.4962356532612326, + "learning_rate": 1.7233482564132646e-07, + "loss": 0.6242, + "step": 5731 + }, + { + "epoch": 10.614814814814816, + "grad_norm": 1.650078858730073, + "learning_rate": 1.723244715563721e-07, + "loss": 0.5948, + "step": 5732 + }, + { + "epoch": 10.616666666666667, + "grad_norm": 1.6468620810718793, + "learning_rate": 1.7231411584536714e-07, + "loss": 0.6371, + "step": 5733 + }, + { + "epoch": 10.618518518518519, + "grad_norm": 1.815103855647812, + "learning_rate": 1.723037585085445e-07, + "loss": 0.6111, + "step": 5734 + }, + { + "epoch": 10.62037037037037, + "grad_norm": 1.6811049186790263, + "learning_rate": 1.7229339954613694e-07, + "loss": 0.6129, + "step": 5735 + }, + { + "epoch": 10.622222222222222, + "grad_norm": 1.4564777650193013, + "learning_rate": 1.7228303895837745e-07, + "loss": 0.5777, + "step": 5736 + }, + { + "epoch": 10.624074074074073, + "grad_norm": 1.3821775978180875, + "learning_rate": 1.7227267674549894e-07, + "loss": 0.6312, + "step": 5737 + }, + { + "epoch": 10.625925925925927, + "grad_norm": 1.5338686232048901, + "learning_rate": 1.7226231290773435e-07, + "loss": 0.6593, + "step": 5738 + }, + { + "epoch": 10.627777777777778, + "grad_norm": 2.0359658791350177, + "learning_rate": 1.722519474453167e-07, + "loss": 0.6529, + "step": 5739 + }, + { + "epoch": 10.62962962962963, + "grad_norm": 1.757406146016718, + "learning_rate": 1.7224158035847903e-07, + "loss": 0.6288, + "step": 5740 + }, + { + "epoch": 10.631481481481481, + "grad_norm": 1.314803996865729, + "learning_rate": 1.7223121164745446e-07, + "loss": 0.6225, + "step": 5741 + }, + { + "epoch": 10.633333333333333, + "grad_norm": 1.606924771041962, + "learning_rate": 1.7222084131247606e-07, + "loss": 0.6266, + "step": 5742 + }, + { + "epoch": 10.635185185185184, + "grad_norm": 1.4731856113852748, + "learning_rate": 1.7221046935377698e-07, + "loss": 0.6235, + "step": 5743 + }, + { + "epoch": 10.637037037037038, + "grad_norm": 2.0233919686711443, + "learning_rate": 1.7220009577159047e-07, + "loss": 0.6523, + "step": 5744 + }, + { + "epoch": 10.63888888888889, + "grad_norm": 1.8177760915677952, + "learning_rate": 1.7218972056614967e-07, + "loss": 0.6492, + "step": 5745 + }, + { + "epoch": 10.64074074074074, + "grad_norm": 1.5758205075256775, + "learning_rate": 1.721793437376879e-07, + "loss": 0.6311, + "step": 5746 + }, + { + "epoch": 10.642592592592592, + "grad_norm": 1.576370034078635, + "learning_rate": 1.7216896528643847e-07, + "loss": 0.6539, + "step": 5747 + }, + { + "epoch": 10.644444444444444, + "grad_norm": 1.5142773170881674, + "learning_rate": 1.7215858521263465e-07, + "loss": 0.6399, + "step": 5748 + }, + { + "epoch": 10.646296296296295, + "grad_norm": 1.694950143845439, + "learning_rate": 1.721482035165099e-07, + "loss": 0.6753, + "step": 5749 + }, + { + "epoch": 10.648148148148149, + "grad_norm": 1.4014231838157987, + "learning_rate": 1.7213782019829754e-07, + "loss": 0.6303, + "step": 5750 + }, + { + "epoch": 10.65, + "grad_norm": 1.598392630736057, + "learning_rate": 1.721274352582311e-07, + "loss": 0.6241, + "step": 5751 + }, + { + "epoch": 10.651851851851852, + "grad_norm": 1.3685750005585275, + "learning_rate": 1.7211704869654399e-07, + "loss": 0.6327, + "step": 5752 + }, + { + "epoch": 10.653703703703703, + "grad_norm": 1.9655545281334443, + "learning_rate": 1.721066605134698e-07, + "loss": 0.6537, + "step": 5753 + }, + { + "epoch": 10.655555555555555, + "grad_norm": 1.5167603329062058, + "learning_rate": 1.7209627070924198e-07, + "loss": 0.6374, + "step": 5754 + }, + { + "epoch": 10.657407407407408, + "grad_norm": 1.7027541708785325, + "learning_rate": 1.7208587928409422e-07, + "loss": 0.5996, + "step": 5755 + }, + { + "epoch": 10.65925925925926, + "grad_norm": 1.6180101606610673, + "learning_rate": 1.7207548623826012e-07, + "loss": 0.6302, + "step": 5756 + }, + { + "epoch": 10.661111111111111, + "grad_norm": 1.7882501712593906, + "learning_rate": 1.7206509157197333e-07, + "loss": 0.6742, + "step": 5757 + }, + { + "epoch": 10.662962962962963, + "grad_norm": 1.6386248611687573, + "learning_rate": 1.7205469528546751e-07, + "loss": 0.6479, + "step": 5758 + }, + { + "epoch": 10.664814814814815, + "grad_norm": 1.4597843438575, + "learning_rate": 1.720442973789765e-07, + "loss": 0.6559, + "step": 5759 + }, + { + "epoch": 10.666666666666666, + "grad_norm": 1.5974565688131441, + "learning_rate": 1.72033897852734e-07, + "loss": 0.6493, + "step": 5760 + }, + { + "epoch": 10.668518518518518, + "grad_norm": 1.6548839630022452, + "learning_rate": 1.7202349670697383e-07, + "loss": 0.5894, + "step": 5761 + }, + { + "epoch": 10.670370370370371, + "grad_norm": 1.604307414003859, + "learning_rate": 1.7201309394192985e-07, + "loss": 0.5961, + "step": 5762 + }, + { + "epoch": 10.672222222222222, + "grad_norm": 1.5628062678756305, + "learning_rate": 1.7200268955783593e-07, + "loss": 0.6429, + "step": 5763 + }, + { + "epoch": 10.674074074074074, + "grad_norm": 1.8009219845733089, + "learning_rate": 1.7199228355492598e-07, + "loss": 0.5842, + "step": 5764 + }, + { + "epoch": 10.675925925925926, + "grad_norm": 1.642301145030084, + "learning_rate": 1.71981875933434e-07, + "loss": 0.6378, + "step": 5765 + }, + { + "epoch": 10.677777777777777, + "grad_norm": 1.8840836079181922, + "learning_rate": 1.719714666935939e-07, + "loss": 0.6246, + "step": 5766 + }, + { + "epoch": 10.67962962962963, + "grad_norm": 1.6072095113297304, + "learning_rate": 1.7196105583563977e-07, + "loss": 0.6515, + "step": 5767 + }, + { + "epoch": 10.681481481481482, + "grad_norm": 1.5274787248533939, + "learning_rate": 1.719506433598057e-07, + "loss": 0.6703, + "step": 5768 + }, + { + "epoch": 10.683333333333334, + "grad_norm": 1.5825837593269407, + "learning_rate": 1.7194022926632571e-07, + "loss": 0.5988, + "step": 5769 + }, + { + "epoch": 10.685185185185185, + "grad_norm": 1.6630970104806428, + "learning_rate": 1.7192981355543398e-07, + "loss": 0.6266, + "step": 5770 + }, + { + "epoch": 10.687037037037037, + "grad_norm": 1.4298503678880312, + "learning_rate": 1.719193962273647e-07, + "loss": 0.602, + "step": 5771 + }, + { + "epoch": 10.688888888888888, + "grad_norm": 1.5686657351578537, + "learning_rate": 1.7190897728235208e-07, + "loss": 0.5993, + "step": 5772 + }, + { + "epoch": 10.690740740740742, + "grad_norm": 1.5180783702596443, + "learning_rate": 1.7189855672063033e-07, + "loss": 0.6315, + "step": 5773 + }, + { + "epoch": 10.692592592592593, + "grad_norm": 2.135821158532443, + "learning_rate": 1.7188813454243375e-07, + "loss": 0.596, + "step": 5774 + }, + { + "epoch": 10.694444444444445, + "grad_norm": 1.5792821781851023, + "learning_rate": 1.7187771074799667e-07, + "loss": 0.6045, + "step": 5775 + }, + { + "epoch": 10.696296296296296, + "grad_norm": 1.487706572139368, + "learning_rate": 1.7186728533755344e-07, + "loss": 0.6608, + "step": 5776 + }, + { + "epoch": 10.698148148148148, + "grad_norm": 1.4857700394336428, + "learning_rate": 1.718568583113384e-07, + "loss": 0.6295, + "step": 5777 + }, + { + "epoch": 10.7, + "grad_norm": 1.5599470941653037, + "learning_rate": 1.7184642966958607e-07, + "loss": 0.6593, + "step": 5778 + }, + { + "epoch": 10.701851851851853, + "grad_norm": 1.771713445759415, + "learning_rate": 1.7183599941253085e-07, + "loss": 0.6233, + "step": 5779 + }, + { + "epoch": 10.703703703703704, + "grad_norm": 1.5448332302225707, + "learning_rate": 1.7182556754040727e-07, + "loss": 0.6518, + "step": 5780 + }, + { + "epoch": 10.705555555555556, + "grad_norm": 1.5666295817756888, + "learning_rate": 1.7181513405344986e-07, + "loss": 0.6013, + "step": 5781 + }, + { + "epoch": 10.707407407407407, + "grad_norm": 1.464570150520032, + "learning_rate": 1.7180469895189318e-07, + "loss": 0.6614, + "step": 5782 + }, + { + "epoch": 10.709259259259259, + "grad_norm": 1.4885006350243783, + "learning_rate": 1.7179426223597187e-07, + "loss": 0.6304, + "step": 5783 + }, + { + "epoch": 10.71111111111111, + "grad_norm": 1.751847051815446, + "learning_rate": 1.7178382390592055e-07, + "loss": 0.6198, + "step": 5784 + }, + { + "epoch": 10.712962962962964, + "grad_norm": 1.4465671912199698, + "learning_rate": 1.717733839619739e-07, + "loss": 0.6121, + "step": 5785 + }, + { + "epoch": 10.714814814814815, + "grad_norm": 1.7351913173179225, + "learning_rate": 1.7176294240436662e-07, + "loss": 0.6296, + "step": 5786 + }, + { + "epoch": 10.716666666666667, + "grad_norm": 1.6720490000121169, + "learning_rate": 1.7175249923333351e-07, + "loss": 0.6136, + "step": 5787 + }, + { + "epoch": 10.718518518518518, + "grad_norm": 1.4161435815053065, + "learning_rate": 1.7174205444910935e-07, + "loss": 0.6599, + "step": 5788 + }, + { + "epoch": 10.72037037037037, + "grad_norm": 1.547747891511945, + "learning_rate": 1.7173160805192892e-07, + "loss": 0.6271, + "step": 5789 + }, + { + "epoch": 10.722222222222221, + "grad_norm": 1.8505863824612934, + "learning_rate": 1.7172116004202716e-07, + "loss": 0.6554, + "step": 5790 + }, + { + "epoch": 10.724074074074075, + "grad_norm": 1.5591947262246384, + "learning_rate": 1.7171071041963892e-07, + "loss": 0.618, + "step": 5791 + }, + { + "epoch": 10.725925925925926, + "grad_norm": 1.4283225260216625, + "learning_rate": 1.7170025918499916e-07, + "loss": 0.6655, + "step": 5792 + }, + { + "epoch": 10.727777777777778, + "grad_norm": 1.649494296468817, + "learning_rate": 1.7168980633834282e-07, + "loss": 0.6089, + "step": 5793 + }, + { + "epoch": 10.72962962962963, + "grad_norm": 1.6474498661389734, + "learning_rate": 1.7167935187990493e-07, + "loss": 0.6465, + "step": 5794 + }, + { + "epoch": 10.731481481481481, + "grad_norm": 1.7337928549553054, + "learning_rate": 1.7166889580992053e-07, + "loss": 0.7171, + "step": 5795 + }, + { + "epoch": 10.733333333333333, + "grad_norm": 1.6481526050767683, + "learning_rate": 1.716584381286247e-07, + "loss": 0.6055, + "step": 5796 + }, + { + "epoch": 10.735185185185186, + "grad_norm": 1.5917129592174264, + "learning_rate": 1.7164797883625255e-07, + "loss": 0.6149, + "step": 5797 + }, + { + "epoch": 10.737037037037037, + "grad_norm": 1.7334651950372966, + "learning_rate": 1.7163751793303926e-07, + "loss": 0.6086, + "step": 5798 + }, + { + "epoch": 10.738888888888889, + "grad_norm": 1.5973658229373153, + "learning_rate": 1.7162705541922002e-07, + "loss": 0.6547, + "step": 5799 + }, + { + "epoch": 10.74074074074074, + "grad_norm": 1.7060440299390252, + "learning_rate": 1.7161659129503e-07, + "loss": 0.6168, + "step": 5800 + }, + { + "epoch": 10.742592592592592, + "grad_norm": 2.829532026594358, + "learning_rate": 1.716061255607045e-07, + "loss": 0.6111, + "step": 5801 + }, + { + "epoch": 10.744444444444444, + "grad_norm": 2.1300543124298885, + "learning_rate": 1.7159565821647886e-07, + "loss": 0.6051, + "step": 5802 + }, + { + "epoch": 10.746296296296297, + "grad_norm": 1.5018888552860983, + "learning_rate": 1.7158518926258835e-07, + "loss": 0.63, + "step": 5803 + }, + { + "epoch": 10.748148148148148, + "grad_norm": 1.7256997847152065, + "learning_rate": 1.7157471869926835e-07, + "loss": 0.6034, + "step": 5804 + }, + { + "epoch": 10.75, + "grad_norm": 1.7022402193355048, + "learning_rate": 1.7156424652675428e-07, + "loss": 0.6287, + "step": 5805 + }, + { + "epoch": 10.751851851851852, + "grad_norm": 1.6337867590197357, + "learning_rate": 1.715537727452816e-07, + "loss": 0.594, + "step": 5806 + }, + { + "epoch": 10.753703703703703, + "grad_norm": 1.7034862539137605, + "learning_rate": 1.7154329735508578e-07, + "loss": 0.6017, + "step": 5807 + }, + { + "epoch": 10.755555555555556, + "grad_norm": 1.5604583378884518, + "learning_rate": 1.7153282035640228e-07, + "loss": 0.6057, + "step": 5808 + }, + { + "epoch": 10.757407407407408, + "grad_norm": 1.5128060791383016, + "learning_rate": 1.7152234174946675e-07, + "loss": 0.6479, + "step": 5809 + }, + { + "epoch": 10.75925925925926, + "grad_norm": 1.7800225785340646, + "learning_rate": 1.7151186153451472e-07, + "loss": 0.6667, + "step": 5810 + }, + { + "epoch": 10.761111111111111, + "grad_norm": 1.8018205137617924, + "learning_rate": 1.715013797117818e-07, + "loss": 0.6167, + "step": 5811 + }, + { + "epoch": 10.762962962962963, + "grad_norm": 1.6405391843331543, + "learning_rate": 1.7149089628150367e-07, + "loss": 0.6408, + "step": 5812 + }, + { + "epoch": 10.764814814814814, + "grad_norm": 1.6362062112278146, + "learning_rate": 1.7148041124391603e-07, + "loss": 0.6462, + "step": 5813 + }, + { + "epoch": 10.766666666666667, + "grad_norm": 2.8559182148846984, + "learning_rate": 1.714699245992546e-07, + "loss": 0.6353, + "step": 5814 + }, + { + "epoch": 10.768518518518519, + "grad_norm": 1.5107189926914173, + "learning_rate": 1.7145943634775517e-07, + "loss": 0.6699, + "step": 5815 + }, + { + "epoch": 10.77037037037037, + "grad_norm": 1.5374575422975876, + "learning_rate": 1.7144894648965353e-07, + "loss": 0.5766, + "step": 5816 + }, + { + "epoch": 10.772222222222222, + "grad_norm": 1.4958817630047532, + "learning_rate": 1.714384550251855e-07, + "loss": 0.6388, + "step": 5817 + }, + { + "epoch": 10.774074074074074, + "grad_norm": 1.599019371328936, + "learning_rate": 1.71427961954587e-07, + "loss": 0.5986, + "step": 5818 + }, + { + "epoch": 10.775925925925925, + "grad_norm": 1.6231399426538982, + "learning_rate": 1.7141746727809392e-07, + "loss": 0.6438, + "step": 5819 + }, + { + "epoch": 10.777777777777779, + "grad_norm": 1.580626800626447, + "learning_rate": 1.7140697099594218e-07, + "loss": 0.6156, + "step": 5820 + }, + { + "epoch": 10.77962962962963, + "grad_norm": 1.4687918546796297, + "learning_rate": 1.713964731083678e-07, + "loss": 0.6253, + "step": 5821 + }, + { + "epoch": 10.781481481481482, + "grad_norm": 1.6858469050498297, + "learning_rate": 1.7138597361560682e-07, + "loss": 0.6467, + "step": 5822 + }, + { + "epoch": 10.783333333333333, + "grad_norm": 1.8416731824236705, + "learning_rate": 1.7137547251789527e-07, + "loss": 0.6079, + "step": 5823 + }, + { + "epoch": 10.785185185185185, + "grad_norm": 1.5240326735536087, + "learning_rate": 1.7136496981546924e-07, + "loss": 0.6057, + "step": 5824 + }, + { + "epoch": 10.787037037037036, + "grad_norm": 1.5893687083399342, + "learning_rate": 1.7135446550856486e-07, + "loss": 0.6172, + "step": 5825 + }, + { + "epoch": 10.78888888888889, + "grad_norm": 1.416083158743161, + "learning_rate": 1.7134395959741828e-07, + "loss": 0.6092, + "step": 5826 + }, + { + "epoch": 10.790740740740741, + "grad_norm": 1.4278803242233795, + "learning_rate": 1.7133345208226576e-07, + "loss": 0.6214, + "step": 5827 + }, + { + "epoch": 10.792592592592593, + "grad_norm": 1.4683796597877252, + "learning_rate": 1.7132294296334347e-07, + "loss": 0.6597, + "step": 5828 + }, + { + "epoch": 10.794444444444444, + "grad_norm": 1.5004471842097729, + "learning_rate": 1.713124322408877e-07, + "loss": 0.672, + "step": 5829 + }, + { + "epoch": 10.796296296296296, + "grad_norm": 1.7601918721067162, + "learning_rate": 1.713019199151348e-07, + "loss": 0.6402, + "step": 5830 + }, + { + "epoch": 10.798148148148147, + "grad_norm": 1.7234651239570857, + "learning_rate": 1.712914059863211e-07, + "loss": 0.6067, + "step": 5831 + }, + { + "epoch": 10.8, + "grad_norm": 1.3879055873838537, + "learning_rate": 1.7128089045468293e-07, + "loss": 0.6249, + "step": 5832 + }, + { + "epoch": 10.801851851851852, + "grad_norm": 1.7422974737478263, + "learning_rate": 1.7127037332045676e-07, + "loss": 0.6132, + "step": 5833 + }, + { + "epoch": 10.803703703703704, + "grad_norm": 1.5972577099884724, + "learning_rate": 1.7125985458387904e-07, + "loss": 0.6533, + "step": 5834 + }, + { + "epoch": 10.805555555555555, + "grad_norm": 1.639882097607804, + "learning_rate": 1.7124933424518623e-07, + "loss": 0.6491, + "step": 5835 + }, + { + "epoch": 10.807407407407407, + "grad_norm": 1.4821054277619221, + "learning_rate": 1.712388123046149e-07, + "loss": 0.6205, + "step": 5836 + }, + { + "epoch": 10.809259259259258, + "grad_norm": 3.6329563508240157, + "learning_rate": 1.7122828876240156e-07, + "loss": 0.615, + "step": 5837 + }, + { + "epoch": 10.811111111111112, + "grad_norm": 1.6592183703060885, + "learning_rate": 1.7121776361878285e-07, + "loss": 0.673, + "step": 5838 + }, + { + "epoch": 10.812962962962963, + "grad_norm": 1.5405493700107749, + "learning_rate": 1.712072368739954e-07, + "loss": 0.6037, + "step": 5839 + }, + { + "epoch": 10.814814814814815, + "grad_norm": 1.4304083923326163, + "learning_rate": 1.7119670852827586e-07, + "loss": 0.6308, + "step": 5840 + }, + { + "epoch": 10.816666666666666, + "grad_norm": 3.3523033067518977, + "learning_rate": 1.7118617858186096e-07, + "loss": 0.5988, + "step": 5841 + }, + { + "epoch": 10.818518518518518, + "grad_norm": 1.6273275768169826, + "learning_rate": 1.711756470349874e-07, + "loss": 0.6384, + "step": 5842 + }, + { + "epoch": 10.82037037037037, + "grad_norm": 1.9555284104965578, + "learning_rate": 1.7116511388789201e-07, + "loss": 0.5834, + "step": 5843 + }, + { + "epoch": 10.822222222222223, + "grad_norm": 1.7907661489313798, + "learning_rate": 1.7115457914081156e-07, + "loss": 0.609, + "step": 5844 + }, + { + "epoch": 10.824074074074074, + "grad_norm": 1.5144972052140446, + "learning_rate": 1.7114404279398294e-07, + "loss": 0.6217, + "step": 5845 + }, + { + "epoch": 10.825925925925926, + "grad_norm": 1.5312180497361965, + "learning_rate": 1.7113350484764297e-07, + "loss": 0.6492, + "step": 5846 + }, + { + "epoch": 10.827777777777778, + "grad_norm": 1.6680249175255166, + "learning_rate": 1.7112296530202865e-07, + "loss": 0.6299, + "step": 5847 + }, + { + "epoch": 10.829629629629629, + "grad_norm": 1.5485044703697364, + "learning_rate": 1.711124241573769e-07, + "loss": 0.6662, + "step": 5848 + }, + { + "epoch": 10.831481481481482, + "grad_norm": 1.4810048006627499, + "learning_rate": 1.711018814139247e-07, + "loss": 0.6021, + "step": 5849 + }, + { + "epoch": 10.833333333333334, + "grad_norm": 1.899062348275314, + "learning_rate": 1.710913370719091e-07, + "loss": 0.6239, + "step": 5850 + }, + { + "epoch": 10.835185185185185, + "grad_norm": 1.945187880470277, + "learning_rate": 1.710807911315672e-07, + "loss": 0.6081, + "step": 5851 + }, + { + "epoch": 10.837037037037037, + "grad_norm": 1.5811830615277291, + "learning_rate": 1.7107024359313603e-07, + "loss": 0.651, + "step": 5852 + }, + { + "epoch": 10.838888888888889, + "grad_norm": 1.5500753098618565, + "learning_rate": 1.7105969445685275e-07, + "loss": 0.6798, + "step": 5853 + }, + { + "epoch": 10.84074074074074, + "grad_norm": 1.9419497627389486, + "learning_rate": 1.7104914372295456e-07, + "loss": 0.6469, + "step": 5854 + }, + { + "epoch": 10.842592592592592, + "grad_norm": 1.462000541562202, + "learning_rate": 1.7103859139167867e-07, + "loss": 0.6413, + "step": 5855 + }, + { + "epoch": 10.844444444444445, + "grad_norm": 1.5288525604190968, + "learning_rate": 1.7102803746326227e-07, + "loss": 0.6319, + "step": 5856 + }, + { + "epoch": 10.846296296296297, + "grad_norm": 1.488358505757009, + "learning_rate": 1.710174819379427e-07, + "loss": 0.6145, + "step": 5857 + }, + { + "epoch": 10.848148148148148, + "grad_norm": 2.3875678304550707, + "learning_rate": 1.7100692481595727e-07, + "loss": 0.6432, + "step": 5858 + }, + { + "epoch": 10.85, + "grad_norm": 1.361304194274407, + "learning_rate": 1.7099636609754326e-07, + "loss": 0.6048, + "step": 5859 + }, + { + "epoch": 10.851851851851851, + "grad_norm": 1.463855705963754, + "learning_rate": 1.7098580578293818e-07, + "loss": 0.5892, + "step": 5860 + }, + { + "epoch": 10.853703703703705, + "grad_norm": 1.5644114917869907, + "learning_rate": 1.7097524387237936e-07, + "loss": 0.6075, + "step": 5861 + }, + { + "epoch": 10.855555555555556, + "grad_norm": 1.3640782096032182, + "learning_rate": 1.709646803661043e-07, + "loss": 0.6224, + "step": 5862 + }, + { + "epoch": 10.857407407407408, + "grad_norm": 1.5128254741137421, + "learning_rate": 1.709541152643505e-07, + "loss": 0.6027, + "step": 5863 + }, + { + "epoch": 10.85925925925926, + "grad_norm": 1.4619139483410486, + "learning_rate": 1.7094354856735548e-07, + "loss": 0.6242, + "step": 5864 + }, + { + "epoch": 10.86111111111111, + "grad_norm": 1.3849844022499764, + "learning_rate": 1.7093298027535677e-07, + "loss": 0.6249, + "step": 5865 + }, + { + "epoch": 10.862962962962962, + "grad_norm": 1.668214596545056, + "learning_rate": 1.7092241038859207e-07, + "loss": 0.6404, + "step": 5866 + }, + { + "epoch": 10.864814814814816, + "grad_norm": 1.344688229689299, + "learning_rate": 1.709118389072989e-07, + "loss": 0.6471, + "step": 5867 + }, + { + "epoch": 10.866666666666667, + "grad_norm": 1.4592079445921544, + "learning_rate": 1.7090126583171503e-07, + "loss": 0.6518, + "step": 5868 + }, + { + "epoch": 10.868518518518519, + "grad_norm": 1.5195929428178105, + "learning_rate": 1.7089069116207812e-07, + "loss": 0.6428, + "step": 5869 + }, + { + "epoch": 10.87037037037037, + "grad_norm": 1.7047514622055526, + "learning_rate": 1.7088011489862598e-07, + "loss": 0.6277, + "step": 5870 + }, + { + "epoch": 10.872222222222222, + "grad_norm": 1.660986182730027, + "learning_rate": 1.7086953704159631e-07, + "loss": 0.6414, + "step": 5871 + }, + { + "epoch": 10.874074074074073, + "grad_norm": 1.4967613780406208, + "learning_rate": 1.7085895759122697e-07, + "loss": 0.6457, + "step": 5872 + }, + { + "epoch": 10.875925925925927, + "grad_norm": 1.6871884626761262, + "learning_rate": 1.7084837654775583e-07, + "loss": 0.6527, + "step": 5873 + }, + { + "epoch": 10.877777777777778, + "grad_norm": 2.2725637080703462, + "learning_rate": 1.7083779391142077e-07, + "loss": 0.6361, + "step": 5874 + }, + { + "epoch": 10.87962962962963, + "grad_norm": 1.7463092073225324, + "learning_rate": 1.7082720968245973e-07, + "loss": 0.6215, + "step": 5875 + }, + { + "epoch": 10.881481481481481, + "grad_norm": 1.555899650861377, + "learning_rate": 1.708166238611106e-07, + "loss": 0.6331, + "step": 5876 + }, + { + "epoch": 10.883333333333333, + "grad_norm": 2.3952046897198165, + "learning_rate": 1.7080603644761146e-07, + "loss": 0.5985, + "step": 5877 + }, + { + "epoch": 10.885185185185184, + "grad_norm": 1.5667752679022278, + "learning_rate": 1.707954474422003e-07, + "loss": 0.6049, + "step": 5878 + }, + { + "epoch": 10.887037037037038, + "grad_norm": 1.523051244182511, + "learning_rate": 1.7078485684511522e-07, + "loss": 0.6246, + "step": 5879 + }, + { + "epoch": 10.88888888888889, + "grad_norm": 1.5159082377614266, + "learning_rate": 1.7077426465659431e-07, + "loss": 0.583, + "step": 5880 + }, + { + "epoch": 10.89074074074074, + "grad_norm": 1.3717961459429246, + "learning_rate": 1.707636708768757e-07, + "loss": 0.6163, + "step": 5881 + }, + { + "epoch": 10.892592592592592, + "grad_norm": 1.6332154906681302, + "learning_rate": 1.7075307550619758e-07, + "loss": 0.658, + "step": 5882 + }, + { + "epoch": 10.894444444444444, + "grad_norm": 1.4130658756923027, + "learning_rate": 1.7074247854479814e-07, + "loss": 0.6392, + "step": 5883 + }, + { + "epoch": 10.896296296296295, + "grad_norm": 1.5593087207293561, + "learning_rate": 1.7073187999291567e-07, + "loss": 0.6316, + "step": 5884 + }, + { + "epoch": 10.898148148148149, + "grad_norm": 1.5107680102122947, + "learning_rate": 1.7072127985078844e-07, + "loss": 0.6098, + "step": 5885 + }, + { + "epoch": 10.9, + "grad_norm": 1.5810094424258323, + "learning_rate": 1.7071067811865473e-07, + "loss": 0.6747, + "step": 5886 + }, + { + "epoch": 10.901851851851852, + "grad_norm": 1.4391330277702756, + "learning_rate": 1.7070007479675296e-07, + "loss": 0.5995, + "step": 5887 + }, + { + "epoch": 10.903703703703703, + "grad_norm": 1.58909736533067, + "learning_rate": 1.706894698853215e-07, + "loss": 0.6196, + "step": 5888 + }, + { + "epoch": 10.905555555555555, + "grad_norm": 1.6160992523099549, + "learning_rate": 1.7067886338459875e-07, + "loss": 0.6746, + "step": 5889 + }, + { + "epoch": 10.907407407407408, + "grad_norm": 1.7125014057466894, + "learning_rate": 1.7066825529482319e-07, + "loss": 0.575, + "step": 5890 + }, + { + "epoch": 10.90925925925926, + "grad_norm": 1.6978009646612635, + "learning_rate": 1.706576456162333e-07, + "loss": 0.6328, + "step": 5891 + }, + { + "epoch": 10.911111111111111, + "grad_norm": 1.6722955547612481, + "learning_rate": 1.7064703434906767e-07, + "loss": 0.6444, + "step": 5892 + }, + { + "epoch": 10.912962962962963, + "grad_norm": 1.7627915282042608, + "learning_rate": 1.7063642149356481e-07, + "loss": 0.6066, + "step": 5893 + }, + { + "epoch": 10.914814814814815, + "grad_norm": 1.7990006325096803, + "learning_rate": 1.7062580704996335e-07, + "loss": 0.6122, + "step": 5894 + }, + { + "epoch": 10.916666666666666, + "grad_norm": 2.801482688622267, + "learning_rate": 1.7061519101850195e-07, + "loss": 0.6251, + "step": 5895 + }, + { + "epoch": 10.918518518518518, + "grad_norm": 1.664448645333804, + "learning_rate": 1.7060457339941926e-07, + "loss": 0.6406, + "step": 5896 + }, + { + "epoch": 10.920370370370371, + "grad_norm": 1.4676581184883395, + "learning_rate": 1.7059395419295396e-07, + "loss": 0.6705, + "step": 5897 + }, + { + "epoch": 10.922222222222222, + "grad_norm": 1.6122419827824905, + "learning_rate": 1.705833333993449e-07, + "loss": 0.6384, + "step": 5898 + }, + { + "epoch": 10.924074074074074, + "grad_norm": 1.8826563430347905, + "learning_rate": 1.7057271101883077e-07, + "loss": 0.6345, + "step": 5899 + }, + { + "epoch": 10.925925925925926, + "grad_norm": 1.7319788554003726, + "learning_rate": 1.7056208705165041e-07, + "loss": 0.6643, + "step": 5900 + }, + { + "epoch": 10.927777777777777, + "grad_norm": 2.009172340771583, + "learning_rate": 1.7055146149804273e-07, + "loss": 0.6551, + "step": 5901 + }, + { + "epoch": 10.92962962962963, + "grad_norm": 1.7828116051587684, + "learning_rate": 1.7054083435824657e-07, + "loss": 0.5951, + "step": 5902 + }, + { + "epoch": 10.931481481481482, + "grad_norm": 1.8879528408979587, + "learning_rate": 1.7053020563250087e-07, + "loss": 0.6283, + "step": 5903 + }, + { + "epoch": 10.933333333333334, + "grad_norm": 1.4773106532183309, + "learning_rate": 1.7051957532104459e-07, + "loss": 0.615, + "step": 5904 + }, + { + "epoch": 10.935185185185185, + "grad_norm": 1.4529833705674395, + "learning_rate": 1.705089434241167e-07, + "loss": 0.5874, + "step": 5905 + }, + { + "epoch": 10.937037037037037, + "grad_norm": 1.7386534208172533, + "learning_rate": 1.7049830994195626e-07, + "loss": 0.6616, + "step": 5906 + }, + { + "epoch": 10.938888888888888, + "grad_norm": 1.480171742846781, + "learning_rate": 1.7048767487480239e-07, + "loss": 0.6581, + "step": 5907 + }, + { + "epoch": 10.940740740740742, + "grad_norm": 2.4007596250601653, + "learning_rate": 1.7047703822289413e-07, + "loss": 0.6331, + "step": 5908 + }, + { + "epoch": 10.942592592592593, + "grad_norm": 1.6496927901459952, + "learning_rate": 1.7046639998647062e-07, + "loss": 0.6502, + "step": 5909 + }, + { + "epoch": 10.944444444444445, + "grad_norm": 1.537044194589527, + "learning_rate": 1.7045576016577102e-07, + "loss": 0.6481, + "step": 5910 + }, + { + "epoch": 10.946296296296296, + "grad_norm": 1.4557301788407626, + "learning_rate": 1.7044511876103464e-07, + "loss": 0.6144, + "step": 5911 + }, + { + "epoch": 10.948148148148148, + "grad_norm": 1.6435139556027867, + "learning_rate": 1.7043447577250064e-07, + "loss": 0.6402, + "step": 5912 + }, + { + "epoch": 10.95, + "grad_norm": 1.7564321079667797, + "learning_rate": 1.7042383120040833e-07, + "loss": 0.6436, + "step": 5913 + }, + { + "epoch": 10.951851851851853, + "grad_norm": 1.5061460319688613, + "learning_rate": 1.7041318504499703e-07, + "loss": 0.6294, + "step": 5914 + }, + { + "epoch": 10.953703703703704, + "grad_norm": 2.5184878638740904, + "learning_rate": 1.7040253730650608e-07, + "loss": 0.6538, + "step": 5915 + }, + { + "epoch": 10.955555555555556, + "grad_norm": 1.5949283379518575, + "learning_rate": 1.703918879851749e-07, + "loss": 0.6158, + "step": 5916 + }, + { + "epoch": 10.957407407407407, + "grad_norm": 1.752464075535082, + "learning_rate": 1.7038123708124288e-07, + "loss": 0.6423, + "step": 5917 + }, + { + "epoch": 10.959259259259259, + "grad_norm": 1.5609427691533169, + "learning_rate": 1.703705845949495e-07, + "loss": 0.6132, + "step": 5918 + }, + { + "epoch": 10.96111111111111, + "grad_norm": 1.8231154624004224, + "learning_rate": 1.7035993052653427e-07, + "loss": 0.631, + "step": 5919 + }, + { + "epoch": 10.962962962962964, + "grad_norm": 1.5501861038383078, + "learning_rate": 1.7034927487623668e-07, + "loss": 0.6409, + "step": 5920 + }, + { + "epoch": 10.964814814814815, + "grad_norm": 1.8829032410868844, + "learning_rate": 1.7033861764429634e-07, + "loss": 0.6023, + "step": 5921 + }, + { + "epoch": 10.966666666666667, + "grad_norm": 1.9486797269862017, + "learning_rate": 1.7032795883095284e-07, + "loss": 0.6345, + "step": 5922 + }, + { + "epoch": 10.968518518518518, + "grad_norm": 2.5087948213646323, + "learning_rate": 1.7031729843644582e-07, + "loss": 0.5874, + "step": 5923 + }, + { + "epoch": 10.97037037037037, + "grad_norm": 1.4019500252654198, + "learning_rate": 1.7030663646101493e-07, + "loss": 0.6489, + "step": 5924 + }, + { + "epoch": 10.972222222222221, + "grad_norm": 2.002097245795404, + "learning_rate": 1.7029597290489994e-07, + "loss": 0.5809, + "step": 5925 + }, + { + "epoch": 10.974074074074075, + "grad_norm": 1.7688245375718317, + "learning_rate": 1.7028530776834056e-07, + "loss": 0.6196, + "step": 5926 + }, + { + "epoch": 10.975925925925926, + "grad_norm": 1.449846185082219, + "learning_rate": 1.7027464105157653e-07, + "loss": 0.608, + "step": 5927 + }, + { + "epoch": 10.977777777777778, + "grad_norm": 1.5576471207792844, + "learning_rate": 1.7026397275484773e-07, + "loss": 0.6157, + "step": 5928 + }, + { + "epoch": 10.97962962962963, + "grad_norm": 1.944666932166961, + "learning_rate": 1.70253302878394e-07, + "loss": 0.6052, + "step": 5929 + }, + { + "epoch": 10.981481481481481, + "grad_norm": 1.7634398773183664, + "learning_rate": 1.7024263142245515e-07, + "loss": 0.629, + "step": 5930 + }, + { + "epoch": 10.983333333333333, + "grad_norm": 1.698089668444641, + "learning_rate": 1.7023195838727124e-07, + "loss": 0.6443, + "step": 5931 + }, + { + "epoch": 10.985185185185186, + "grad_norm": 1.4705764205025331, + "learning_rate": 1.7022128377308215e-07, + "loss": 0.5944, + "step": 5932 + }, + { + "epoch": 10.987037037037037, + "grad_norm": 1.5898779509725238, + "learning_rate": 1.7021060758012785e-07, + "loss": 0.6544, + "step": 5933 + }, + { + "epoch": 10.988888888888889, + "grad_norm": 1.425544155866549, + "learning_rate": 1.7019992980864844e-07, + "loss": 0.6397, + "step": 5934 + }, + { + "epoch": 10.99074074074074, + "grad_norm": 1.5403476289747633, + "learning_rate": 1.7018925045888393e-07, + "loss": 0.63, + "step": 5935 + }, + { + "epoch": 10.992592592592592, + "grad_norm": 1.546030320999319, + "learning_rate": 1.7017856953107441e-07, + "loss": 0.5924, + "step": 5936 + }, + { + "epoch": 10.994444444444444, + "grad_norm": 1.7900087734556676, + "learning_rate": 1.701678870254601e-07, + "loss": 0.5829, + "step": 5937 + }, + { + "epoch": 10.996296296296297, + "grad_norm": 1.5863577571041783, + "learning_rate": 1.7015720294228109e-07, + "loss": 0.5829, + "step": 5938 + }, + { + "epoch": 10.998148148148148, + "grad_norm": 1.7033146766985157, + "learning_rate": 1.701465172817776e-07, + "loss": 0.6205, + "step": 5939 + }, + { + "epoch": 11.0, + "grad_norm": 1.4980773676279124, + "learning_rate": 1.7013583004418992e-07, + "loss": 0.6477, + "step": 5940 + }, + { + "epoch": 11.001851851851852, + "grad_norm": 1.7757209065045465, + "learning_rate": 1.7012514122975827e-07, + "loss": 0.6264, + "step": 5941 + }, + { + "epoch": 11.003703703703703, + "grad_norm": 1.5566347525346038, + "learning_rate": 1.7011445083872297e-07, + "loss": 0.6158, + "step": 5942 + }, + { + "epoch": 11.005555555555556, + "grad_norm": 1.8648723244427359, + "learning_rate": 1.7010375887132442e-07, + "loss": 0.6361, + "step": 5943 + }, + { + "epoch": 11.007407407407408, + "grad_norm": 1.6622766515932923, + "learning_rate": 1.7009306532780295e-07, + "loss": 0.6351, + "step": 5944 + }, + { + "epoch": 11.00925925925926, + "grad_norm": 1.4281526003066156, + "learning_rate": 1.7008237020839903e-07, + "loss": 0.6298, + "step": 5945 + }, + { + "epoch": 11.011111111111111, + "grad_norm": 1.615771249114317, + "learning_rate": 1.7007167351335307e-07, + "loss": 0.6221, + "step": 5946 + }, + { + "epoch": 11.012962962962963, + "grad_norm": 1.426716798274646, + "learning_rate": 1.7006097524290556e-07, + "loss": 0.6147, + "step": 5947 + }, + { + "epoch": 11.014814814814814, + "grad_norm": 2.0816540916383426, + "learning_rate": 1.7005027539729708e-07, + "loss": 0.6317, + "step": 5948 + }, + { + "epoch": 11.016666666666667, + "grad_norm": 1.5724228171986836, + "learning_rate": 1.7003957397676816e-07, + "loss": 0.646, + "step": 5949 + }, + { + "epoch": 11.018518518518519, + "grad_norm": 1.5234184039832246, + "learning_rate": 1.7002887098155933e-07, + "loss": 0.6359, + "step": 5950 + }, + { + "epoch": 11.02037037037037, + "grad_norm": 1.7935653084244785, + "learning_rate": 1.7001816641191133e-07, + "loss": 0.6182, + "step": 5951 + }, + { + "epoch": 11.022222222222222, + "grad_norm": 1.6395628034279153, + "learning_rate": 1.7000746026806477e-07, + "loss": 0.6449, + "step": 5952 + }, + { + "epoch": 11.024074074074074, + "grad_norm": 1.8087511888882657, + "learning_rate": 1.699967525502604e-07, + "loss": 0.6233, + "step": 5953 + }, + { + "epoch": 11.025925925925925, + "grad_norm": 1.5036446125841607, + "learning_rate": 1.699860432587389e-07, + "loss": 0.6061, + "step": 5954 + }, + { + "epoch": 11.027777777777779, + "grad_norm": 1.3716273776904815, + "learning_rate": 1.6997533239374104e-07, + "loss": 0.625, + "step": 5955 + }, + { + "epoch": 11.02962962962963, + "grad_norm": 1.5582109838612122, + "learning_rate": 1.699646199555077e-07, + "loss": 0.6222, + "step": 5956 + }, + { + "epoch": 11.031481481481482, + "grad_norm": 1.5441614423046037, + "learning_rate": 1.6995390594427965e-07, + "loss": 0.6588, + "step": 5957 + }, + { + "epoch": 11.033333333333333, + "grad_norm": 1.547692424104394, + "learning_rate": 1.6994319036029783e-07, + "loss": 0.6078, + "step": 5958 + }, + { + "epoch": 11.035185185185185, + "grad_norm": 1.5502695973119278, + "learning_rate": 1.6993247320380313e-07, + "loss": 0.6319, + "step": 5959 + }, + { + "epoch": 11.037037037037036, + "grad_norm": 1.5349932782159459, + "learning_rate": 1.6992175447503647e-07, + "loss": 0.6348, + "step": 5960 + }, + { + "epoch": 11.03888888888889, + "grad_norm": 1.8132430452866657, + "learning_rate": 1.699110341742389e-07, + "loss": 0.6373, + "step": 5961 + }, + { + "epoch": 11.040740740740741, + "grad_norm": 1.4455850069334868, + "learning_rate": 1.6990031230165135e-07, + "loss": 0.6484, + "step": 5962 + }, + { + "epoch": 11.042592592592593, + "grad_norm": 1.709489917083784, + "learning_rate": 1.69889588857515e-07, + "loss": 0.6083, + "step": 5963 + }, + { + "epoch": 11.044444444444444, + "grad_norm": 1.7275875346330496, + "learning_rate": 1.6987886384207083e-07, + "loss": 0.6763, + "step": 5964 + }, + { + "epoch": 11.046296296296296, + "grad_norm": 1.6342008958016723, + "learning_rate": 1.6986813725556002e-07, + "loss": 0.6267, + "step": 5965 + }, + { + "epoch": 11.048148148148147, + "grad_norm": 1.6178084650958968, + "learning_rate": 1.6985740909822375e-07, + "loss": 0.6068, + "step": 5966 + }, + { + "epoch": 11.05, + "grad_norm": 1.528443996617462, + "learning_rate": 1.6984667937030316e-07, + "loss": 0.5982, + "step": 5967 + }, + { + "epoch": 11.051851851851852, + "grad_norm": 1.729968006765182, + "learning_rate": 1.6983594807203953e-07, + "loss": 0.6391, + "step": 5968 + }, + { + "epoch": 11.053703703703704, + "grad_norm": 1.4396639839859426, + "learning_rate": 1.6982521520367411e-07, + "loss": 0.5996, + "step": 5969 + }, + { + "epoch": 11.055555555555555, + "grad_norm": 1.9726778631487216, + "learning_rate": 1.6981448076544825e-07, + "loss": 0.6212, + "step": 5970 + }, + { + "epoch": 11.057407407407407, + "grad_norm": 1.6734363757810804, + "learning_rate": 1.6980374475760318e-07, + "loss": 0.6378, + "step": 5971 + }, + { + "epoch": 11.059259259259258, + "grad_norm": 1.6628775109845524, + "learning_rate": 1.6979300718038042e-07, + "loss": 0.6211, + "step": 5972 + }, + { + "epoch": 11.061111111111112, + "grad_norm": 1.6559022609778822, + "learning_rate": 1.6978226803402123e-07, + "loss": 0.6389, + "step": 5973 + }, + { + "epoch": 11.062962962962963, + "grad_norm": 1.477957725222872, + "learning_rate": 1.697715273187672e-07, + "loss": 0.6663, + "step": 5974 + }, + { + "epoch": 11.064814814814815, + "grad_norm": 1.7470628626633204, + "learning_rate": 1.697607850348597e-07, + "loss": 0.6123, + "step": 5975 + }, + { + "epoch": 11.066666666666666, + "grad_norm": 1.5236742126645169, + "learning_rate": 1.6975004118254027e-07, + "loss": 0.6234, + "step": 5976 + }, + { + "epoch": 11.068518518518518, + "grad_norm": 1.4456330907270265, + "learning_rate": 1.697392957620505e-07, + "loss": 0.6046, + "step": 5977 + }, + { + "epoch": 11.07037037037037, + "grad_norm": 1.417079906843894, + "learning_rate": 1.6972854877363193e-07, + "loss": 0.6234, + "step": 5978 + }, + { + "epoch": 11.072222222222223, + "grad_norm": 1.5028361842938198, + "learning_rate": 1.6971780021752624e-07, + "loss": 0.6607, + "step": 5979 + }, + { + "epoch": 11.074074074074074, + "grad_norm": 1.5608358950670342, + "learning_rate": 1.6970705009397502e-07, + "loss": 0.6133, + "step": 5980 + }, + { + "epoch": 11.075925925925926, + "grad_norm": 1.727954642805163, + "learning_rate": 1.6969629840322e-07, + "loss": 0.6048, + "step": 5981 + }, + { + "epoch": 11.077777777777778, + "grad_norm": 3.716198016960567, + "learning_rate": 1.6968554514550291e-07, + "loss": 0.6014, + "step": 5982 + }, + { + "epoch": 11.079629629629629, + "grad_norm": 1.409378465235887, + "learning_rate": 1.6967479032106548e-07, + "loss": 0.6276, + "step": 5983 + }, + { + "epoch": 11.081481481481482, + "grad_norm": 1.7093019742636377, + "learning_rate": 1.6966403393014955e-07, + "loss": 0.599, + "step": 5984 + }, + { + "epoch": 11.083333333333334, + "grad_norm": 1.416583372862473, + "learning_rate": 1.6965327597299695e-07, + "loss": 0.6833, + "step": 5985 + }, + { + "epoch": 11.085185185185185, + "grad_norm": 1.5970954100280967, + "learning_rate": 1.6964251644984948e-07, + "loss": 0.6433, + "step": 5986 + }, + { + "epoch": 11.087037037037037, + "grad_norm": 1.5081580156050232, + "learning_rate": 1.6963175536094915e-07, + "loss": 0.6016, + "step": 5987 + }, + { + "epoch": 11.088888888888889, + "grad_norm": 1.4807640555911499, + "learning_rate": 1.696209927065378e-07, + "loss": 0.6372, + "step": 5988 + }, + { + "epoch": 11.09074074074074, + "grad_norm": 1.6124370048274683, + "learning_rate": 1.6961022848685747e-07, + "loss": 0.6614, + "step": 5989 + }, + { + "epoch": 11.092592592592593, + "grad_norm": 1.7178997830146034, + "learning_rate": 1.6959946270215012e-07, + "loss": 0.6196, + "step": 5990 + }, + { + "epoch": 11.094444444444445, + "grad_norm": 1.7679883986376386, + "learning_rate": 1.6958869535265786e-07, + "loss": 0.6376, + "step": 5991 + }, + { + "epoch": 11.096296296296297, + "grad_norm": 1.688111609031246, + "learning_rate": 1.6957792643862268e-07, + "loss": 0.6204, + "step": 5992 + }, + { + "epoch": 11.098148148148148, + "grad_norm": 1.5947660269786659, + "learning_rate": 1.6956715596028676e-07, + "loss": 0.6142, + "step": 5993 + }, + { + "epoch": 11.1, + "grad_norm": 1.4654212159204272, + "learning_rate": 1.6955638391789226e-07, + "loss": 0.6232, + "step": 5994 + }, + { + "epoch": 11.101851851851851, + "grad_norm": 1.956301815811382, + "learning_rate": 1.6954561031168127e-07, + "loss": 0.6134, + "step": 5995 + }, + { + "epoch": 11.103703703703705, + "grad_norm": 1.5222159748830348, + "learning_rate": 1.6953483514189615e-07, + "loss": 0.6458, + "step": 5996 + }, + { + "epoch": 11.105555555555556, + "grad_norm": 1.3575981078835275, + "learning_rate": 1.6952405840877906e-07, + "loss": 0.6493, + "step": 5997 + }, + { + "epoch": 11.107407407407408, + "grad_norm": 1.8044274184417082, + "learning_rate": 1.6951328011257232e-07, + "loss": 0.6314, + "step": 5998 + }, + { + "epoch": 11.10925925925926, + "grad_norm": 2.1737412828499827, + "learning_rate": 1.6950250025351825e-07, + "loss": 0.6257, + "step": 5999 + }, + { + "epoch": 11.11111111111111, + "grad_norm": 1.651597664799581, + "learning_rate": 1.6949171883185918e-07, + "loss": 0.5951, + "step": 6000 + }, + { + "epoch": 11.112962962962962, + "grad_norm": 2.7260874808956967, + "learning_rate": 1.6948093584783753e-07, + "loss": 0.6081, + "step": 6001 + }, + { + "epoch": 11.114814814814816, + "grad_norm": 1.610616137968889, + "learning_rate": 1.6947015130169576e-07, + "loss": 0.6384, + "step": 6002 + }, + { + "epoch": 11.116666666666667, + "grad_norm": 1.7683623770826475, + "learning_rate": 1.694593651936763e-07, + "loss": 0.6259, + "step": 6003 + }, + { + "epoch": 11.118518518518519, + "grad_norm": 1.7011868717006917, + "learning_rate": 1.6944857752402167e-07, + "loss": 0.5941, + "step": 6004 + }, + { + "epoch": 11.12037037037037, + "grad_norm": 1.5887048445925007, + "learning_rate": 1.694377882929744e-07, + "loss": 0.6372, + "step": 6005 + }, + { + "epoch": 11.122222222222222, + "grad_norm": 1.576354478728774, + "learning_rate": 1.6942699750077706e-07, + "loss": 0.6225, + "step": 6006 + }, + { + "epoch": 11.124074074074073, + "grad_norm": 1.609499867610394, + "learning_rate": 1.6941620514767224e-07, + "loss": 0.6076, + "step": 6007 + }, + { + "epoch": 11.125925925925927, + "grad_norm": 1.5540112440566092, + "learning_rate": 1.694054112339026e-07, + "loss": 0.6244, + "step": 6008 + }, + { + "epoch": 11.127777777777778, + "grad_norm": 1.6166724770567855, + "learning_rate": 1.693946157597108e-07, + "loss": 0.571, + "step": 6009 + }, + { + "epoch": 11.12962962962963, + "grad_norm": 1.6899808787818047, + "learning_rate": 1.6938381872533957e-07, + "loss": 0.6218, + "step": 6010 + }, + { + "epoch": 11.131481481481481, + "grad_norm": 2.1029401542675665, + "learning_rate": 1.693730201310317e-07, + "loss": 0.6292, + "step": 6011 + }, + { + "epoch": 11.133333333333333, + "grad_norm": 1.7258477364108313, + "learning_rate": 1.6936221997702988e-07, + "loss": 0.6695, + "step": 6012 + }, + { + "epoch": 11.135185185185184, + "grad_norm": 1.559478884145518, + "learning_rate": 1.6935141826357699e-07, + "loss": 0.6415, + "step": 6013 + }, + { + "epoch": 11.137037037037038, + "grad_norm": 1.5928205579841872, + "learning_rate": 1.6934061499091582e-07, + "loss": 0.6267, + "step": 6014 + }, + { + "epoch": 11.13888888888889, + "grad_norm": 1.4078854556125695, + "learning_rate": 1.6932981015928936e-07, + "loss": 0.6024, + "step": 6015 + }, + { + "epoch": 11.14074074074074, + "grad_norm": 1.492493794216963, + "learning_rate": 1.6931900376894042e-07, + "loss": 0.6426, + "step": 6016 + }, + { + "epoch": 11.142592592592592, + "grad_norm": 1.7437166584427275, + "learning_rate": 1.6930819582011203e-07, + "loss": 0.5964, + "step": 6017 + }, + { + "epoch": 11.144444444444444, + "grad_norm": 2.081434451286772, + "learning_rate": 1.6929738631304716e-07, + "loss": 0.6193, + "step": 6018 + }, + { + "epoch": 11.146296296296295, + "grad_norm": 1.5904257002379132, + "learning_rate": 1.6928657524798882e-07, + "loss": 0.64, + "step": 6019 + }, + { + "epoch": 11.148148148148149, + "grad_norm": 1.651310416420862, + "learning_rate": 1.692757626251801e-07, + "loss": 0.6465, + "step": 6020 + }, + { + "epoch": 11.15, + "grad_norm": 1.5319695012607, + "learning_rate": 1.692649484448641e-07, + "loss": 0.6729, + "step": 6021 + }, + { + "epoch": 11.151851851851852, + "grad_norm": 1.604386973204612, + "learning_rate": 1.6925413270728394e-07, + "loss": 0.5817, + "step": 6022 + }, + { + "epoch": 11.153703703703703, + "grad_norm": 1.623496519385197, + "learning_rate": 1.6924331541268274e-07, + "loss": 0.6162, + "step": 6023 + }, + { + "epoch": 11.155555555555555, + "grad_norm": 1.6107820446110659, + "learning_rate": 1.692324965613038e-07, + "loss": 0.6301, + "step": 6024 + }, + { + "epoch": 11.157407407407407, + "grad_norm": 1.881423819384251, + "learning_rate": 1.6922167615339028e-07, + "loss": 0.6136, + "step": 6025 + }, + { + "epoch": 11.15925925925926, + "grad_norm": 1.7450790994419552, + "learning_rate": 1.6921085418918546e-07, + "loss": 0.6168, + "step": 6026 + }, + { + "epoch": 11.161111111111111, + "grad_norm": 3.034118473800667, + "learning_rate": 1.6920003066893269e-07, + "loss": 0.6211, + "step": 6027 + }, + { + "epoch": 11.162962962962963, + "grad_norm": 1.538125211259693, + "learning_rate": 1.6918920559287528e-07, + "loss": 0.6375, + "step": 6028 + }, + { + "epoch": 11.164814814814815, + "grad_norm": 2.0464684899286696, + "learning_rate": 1.6917837896125661e-07, + "loss": 0.6308, + "step": 6029 + }, + { + "epoch": 11.166666666666666, + "grad_norm": 1.5928481151469456, + "learning_rate": 1.6916755077432012e-07, + "loss": 0.6434, + "step": 6030 + }, + { + "epoch": 11.168518518518518, + "grad_norm": 1.6956289759861047, + "learning_rate": 1.691567210323092e-07, + "loss": 0.6713, + "step": 6031 + }, + { + "epoch": 11.170370370370371, + "grad_norm": 2.1460977688328287, + "learning_rate": 1.6914588973546737e-07, + "loss": 0.6309, + "step": 6032 + }, + { + "epoch": 11.172222222222222, + "grad_norm": 1.4513230539235706, + "learning_rate": 1.6913505688403818e-07, + "loss": 0.5987, + "step": 6033 + }, + { + "epoch": 11.174074074074074, + "grad_norm": 1.6312318700342592, + "learning_rate": 1.6912422247826507e-07, + "loss": 0.6139, + "step": 6034 + }, + { + "epoch": 11.175925925925926, + "grad_norm": 1.5773566082549608, + "learning_rate": 1.6911338651839174e-07, + "loss": 0.5903, + "step": 6035 + }, + { + "epoch": 11.177777777777777, + "grad_norm": 1.6141722374712115, + "learning_rate": 1.691025490046618e-07, + "loss": 0.6339, + "step": 6036 + }, + { + "epoch": 11.17962962962963, + "grad_norm": 1.8830565967122879, + "learning_rate": 1.6909170993731882e-07, + "loss": 0.618, + "step": 6037 + }, + { + "epoch": 11.181481481481482, + "grad_norm": 1.8457149045199799, + "learning_rate": 1.690808693166066e-07, + "loss": 0.6602, + "step": 6038 + }, + { + "epoch": 11.183333333333334, + "grad_norm": 1.6331941515412243, + "learning_rate": 1.690700271427688e-07, + "loss": 0.6108, + "step": 6039 + }, + { + "epoch": 11.185185185185185, + "grad_norm": 1.5435956183842898, + "learning_rate": 1.690591834160492e-07, + "loss": 0.6495, + "step": 6040 + }, + { + "epoch": 11.187037037037037, + "grad_norm": 1.6286746877249632, + "learning_rate": 1.6904833813669156e-07, + "loss": 0.624, + "step": 6041 + }, + { + "epoch": 11.188888888888888, + "grad_norm": 1.4693250721860047, + "learning_rate": 1.6903749130493982e-07, + "loss": 0.6234, + "step": 6042 + }, + { + "epoch": 11.190740740740742, + "grad_norm": 1.7945023863999972, + "learning_rate": 1.6902664292103773e-07, + "loss": 0.6385, + "step": 6043 + }, + { + "epoch": 11.192592592592593, + "grad_norm": 1.7565048693465957, + "learning_rate": 1.690157929852292e-07, + "loss": 0.5955, + "step": 6044 + }, + { + "epoch": 11.194444444444445, + "grad_norm": 1.7556448003776346, + "learning_rate": 1.6900494149775827e-07, + "loss": 0.6122, + "step": 6045 + }, + { + "epoch": 11.196296296296296, + "grad_norm": 1.5193049872043598, + "learning_rate": 1.689940884588688e-07, + "loss": 0.6485, + "step": 6046 + }, + { + "epoch": 11.198148148148148, + "grad_norm": 1.3727336371787522, + "learning_rate": 1.6898323386880481e-07, + "loss": 0.6386, + "step": 6047 + }, + { + "epoch": 11.2, + "grad_norm": 1.325615420928115, + "learning_rate": 1.6897237772781044e-07, + "loss": 0.619, + "step": 6048 + }, + { + "epoch": 11.201851851851853, + "grad_norm": 1.7221032047844296, + "learning_rate": 1.6896152003612961e-07, + "loss": 0.6392, + "step": 6049 + }, + { + "epoch": 11.203703703703704, + "grad_norm": 2.2990644557164694, + "learning_rate": 1.6895066079400657e-07, + "loss": 0.6027, + "step": 6050 + }, + { + "epoch": 11.205555555555556, + "grad_norm": 1.7923598744216604, + "learning_rate": 1.6893980000168538e-07, + "loss": 0.5988, + "step": 6051 + }, + { + "epoch": 11.207407407407407, + "grad_norm": 1.4668453900910225, + "learning_rate": 1.6892893765941025e-07, + "loss": 0.6147, + "step": 6052 + }, + { + "epoch": 11.209259259259259, + "grad_norm": 1.4629318980876314, + "learning_rate": 1.689180737674254e-07, + "loss": 0.6128, + "step": 6053 + }, + { + "epoch": 11.21111111111111, + "grad_norm": 1.588515348146373, + "learning_rate": 1.689072083259751e-07, + "loss": 0.6025, + "step": 6054 + }, + { + "epoch": 11.212962962962964, + "grad_norm": 1.4833052016151294, + "learning_rate": 1.688963413353036e-07, + "loss": 0.5896, + "step": 6055 + }, + { + "epoch": 11.214814814814815, + "grad_norm": 2.245265316933493, + "learning_rate": 1.688854727956552e-07, + "loss": 0.5985, + "step": 6056 + }, + { + "epoch": 11.216666666666667, + "grad_norm": 1.6043734876256155, + "learning_rate": 1.6887460270727428e-07, + "loss": 0.6847, + "step": 6057 + }, + { + "epoch": 11.218518518518518, + "grad_norm": 1.6640803034910947, + "learning_rate": 1.6886373107040525e-07, + "loss": 0.6216, + "step": 6058 + }, + { + "epoch": 11.22037037037037, + "grad_norm": 1.6486378694629042, + "learning_rate": 1.6885285788529252e-07, + "loss": 0.62, + "step": 6059 + }, + { + "epoch": 11.222222222222221, + "grad_norm": 1.5718836868416153, + "learning_rate": 1.6884198315218052e-07, + "loss": 0.5949, + "step": 6060 + }, + { + "epoch": 11.224074074074075, + "grad_norm": 1.5495588270168725, + "learning_rate": 1.6883110687131378e-07, + "loss": 0.6695, + "step": 6061 + }, + { + "epoch": 11.225925925925926, + "grad_norm": 1.5390156570672184, + "learning_rate": 1.688202290429368e-07, + "loss": 0.6187, + "step": 6062 + }, + { + "epoch": 11.227777777777778, + "grad_norm": 1.876350223610275, + "learning_rate": 1.6880934966729415e-07, + "loss": 0.5815, + "step": 6063 + }, + { + "epoch": 11.22962962962963, + "grad_norm": 1.6117198714866232, + "learning_rate": 1.6879846874463048e-07, + "loss": 0.6148, + "step": 6064 + }, + { + "epoch": 11.231481481481481, + "grad_norm": 1.4746371150648672, + "learning_rate": 1.6878758627519036e-07, + "loss": 0.6279, + "step": 6065 + }, + { + "epoch": 11.233333333333333, + "grad_norm": 1.5090741366129812, + "learning_rate": 1.6877670225921845e-07, + "loss": 0.5765, + "step": 6066 + }, + { + "epoch": 11.235185185185186, + "grad_norm": 1.3697860633718788, + "learning_rate": 1.6876581669695953e-07, + "loss": 0.5757, + "step": 6067 + }, + { + "epoch": 11.237037037037037, + "grad_norm": 1.9070652412239084, + "learning_rate": 1.6875492958865824e-07, + "loss": 0.6314, + "step": 6068 + }, + { + "epoch": 11.238888888888889, + "grad_norm": 1.443061466517777, + "learning_rate": 1.6874404093455943e-07, + "loss": 0.6683, + "step": 6069 + }, + { + "epoch": 11.24074074074074, + "grad_norm": 1.5473290418442325, + "learning_rate": 1.6873315073490786e-07, + "loss": 0.6192, + "step": 6070 + }, + { + "epoch": 11.242592592592592, + "grad_norm": 1.877640418874891, + "learning_rate": 1.6872225898994837e-07, + "loss": 0.6346, + "step": 6071 + }, + { + "epoch": 11.244444444444444, + "grad_norm": 1.425461386317159, + "learning_rate": 1.6871136569992586e-07, + "loss": 0.6857, + "step": 6072 + }, + { + "epoch": 11.246296296296297, + "grad_norm": 1.5116383844670422, + "learning_rate": 1.6870047086508526e-07, + "loss": 0.6308, + "step": 6073 + }, + { + "epoch": 11.248148148148148, + "grad_norm": 1.5012634895031323, + "learning_rate": 1.6868957448567146e-07, + "loss": 0.6155, + "step": 6074 + }, + { + "epoch": 11.25, + "grad_norm": 1.9203060105502727, + "learning_rate": 1.6867867656192945e-07, + "loss": 0.5898, + "step": 6075 + }, + { + "epoch": 11.251851851851852, + "grad_norm": 1.4486571292866743, + "learning_rate": 1.686677770941043e-07, + "loss": 0.6292, + "step": 6076 + }, + { + "epoch": 11.253703703703703, + "grad_norm": 1.436259377297359, + "learning_rate": 1.68656876082441e-07, + "loss": 0.6217, + "step": 6077 + }, + { + "epoch": 11.255555555555556, + "grad_norm": 1.5332591755143443, + "learning_rate": 1.6864597352718467e-07, + "loss": 0.6712, + "step": 6078 + }, + { + "epoch": 11.257407407407408, + "grad_norm": 1.568702262379678, + "learning_rate": 1.686350694285804e-07, + "loss": 0.6143, + "step": 6079 + }, + { + "epoch": 11.25925925925926, + "grad_norm": 1.5198962337249173, + "learning_rate": 1.6862416378687336e-07, + "loss": 0.6492, + "step": 6080 + }, + { + "epoch": 11.261111111111111, + "grad_norm": 1.4832586179153602, + "learning_rate": 1.6861325660230872e-07, + "loss": 0.6537, + "step": 6081 + }, + { + "epoch": 11.262962962962963, + "grad_norm": 1.6735327668796576, + "learning_rate": 1.6860234787513176e-07, + "loss": 0.6396, + "step": 6082 + }, + { + "epoch": 11.264814814814814, + "grad_norm": 1.497741682575467, + "learning_rate": 1.685914376055877e-07, + "loss": 0.6042, + "step": 6083 + }, + { + "epoch": 11.266666666666667, + "grad_norm": 1.6488951038822184, + "learning_rate": 1.685805257939218e-07, + "loss": 0.6068, + "step": 6084 + }, + { + "epoch": 11.268518518518519, + "grad_norm": 2.1806082758097993, + "learning_rate": 1.6856961244037945e-07, + "loss": 0.6594, + "step": 6085 + }, + { + "epoch": 11.27037037037037, + "grad_norm": 2.5688896822217604, + "learning_rate": 1.6855869754520592e-07, + "loss": 0.5993, + "step": 6086 + }, + { + "epoch": 11.272222222222222, + "grad_norm": 1.4049906507190306, + "learning_rate": 1.6854778110864673e-07, + "loss": 0.6356, + "step": 6087 + }, + { + "epoch": 11.274074074074074, + "grad_norm": 1.6487246924078636, + "learning_rate": 1.6853686313094722e-07, + "loss": 0.6563, + "step": 6088 + }, + { + "epoch": 11.275925925925925, + "grad_norm": 1.694617583194286, + "learning_rate": 1.685259436123529e-07, + "loss": 0.67, + "step": 6089 + }, + { + "epoch": 11.277777777777779, + "grad_norm": 1.8596727786777985, + "learning_rate": 1.6851502255310923e-07, + "loss": 0.6749, + "step": 6090 + }, + { + "epoch": 11.27962962962963, + "grad_norm": 1.570895438672205, + "learning_rate": 1.685040999534618e-07, + "loss": 0.6384, + "step": 6091 + }, + { + "epoch": 11.281481481481482, + "grad_norm": 1.6792584951084863, + "learning_rate": 1.6849317581365612e-07, + "loss": 0.6068, + "step": 6092 + }, + { + "epoch": 11.283333333333333, + "grad_norm": 1.6733517599722272, + "learning_rate": 1.6848225013393785e-07, + "loss": 0.6563, + "step": 6093 + }, + { + "epoch": 11.285185185185185, + "grad_norm": 1.538314953048454, + "learning_rate": 1.6847132291455257e-07, + "loss": 0.6499, + "step": 6094 + }, + { + "epoch": 11.287037037037036, + "grad_norm": 1.553033337510129, + "learning_rate": 1.68460394155746e-07, + "loss": 0.6426, + "step": 6095 + }, + { + "epoch": 11.28888888888889, + "grad_norm": 1.5090041667604468, + "learning_rate": 1.6844946385776384e-07, + "loss": 0.6343, + "step": 6096 + }, + { + "epoch": 11.290740740740741, + "grad_norm": 1.917276757344583, + "learning_rate": 1.6843853202085182e-07, + "loss": 0.6182, + "step": 6097 + }, + { + "epoch": 11.292592592592593, + "grad_norm": 1.4630878373724512, + "learning_rate": 1.6842759864525573e-07, + "loss": 0.6098, + "step": 6098 + }, + { + "epoch": 11.294444444444444, + "grad_norm": 1.5734014950144677, + "learning_rate": 1.6841666373122138e-07, + "loss": 0.6259, + "step": 6099 + }, + { + "epoch": 11.296296296296296, + "grad_norm": 1.6067452972836314, + "learning_rate": 1.684057272789946e-07, + "loss": 0.6042, + "step": 6100 + }, + { + "epoch": 11.298148148148147, + "grad_norm": 1.5320340246940198, + "learning_rate": 1.683947892888213e-07, + "loss": 0.5977, + "step": 6101 + }, + { + "epoch": 11.3, + "grad_norm": 1.8003599727709685, + "learning_rate": 1.6838384976094736e-07, + "loss": 0.6112, + "step": 6102 + }, + { + "epoch": 11.301851851851852, + "grad_norm": 2.138378519944151, + "learning_rate": 1.6837290869561875e-07, + "loss": 0.6351, + "step": 6103 + }, + { + "epoch": 11.303703703703704, + "grad_norm": 1.5250346964659607, + "learning_rate": 1.6836196609308147e-07, + "loss": 0.6307, + "step": 6104 + }, + { + "epoch": 11.305555555555555, + "grad_norm": 1.5296776369039202, + "learning_rate": 1.683510219535815e-07, + "loss": 0.6403, + "step": 6105 + }, + { + "epoch": 11.307407407407407, + "grad_norm": 1.6509854862658215, + "learning_rate": 1.6834007627736496e-07, + "loss": 0.6304, + "step": 6106 + }, + { + "epoch": 11.309259259259258, + "grad_norm": 1.5471030484463637, + "learning_rate": 1.6832912906467786e-07, + "loss": 0.6282, + "step": 6107 + }, + { + "epoch": 11.311111111111112, + "grad_norm": 1.6338150097128694, + "learning_rate": 1.6831818031576637e-07, + "loss": 0.6256, + "step": 6108 + }, + { + "epoch": 11.312962962962963, + "grad_norm": 2.1878779858980777, + "learning_rate": 1.6830723003087664e-07, + "loss": 0.649, + "step": 6109 + }, + { + "epoch": 11.314814814814815, + "grad_norm": 1.4498875861507257, + "learning_rate": 1.6829627821025488e-07, + "loss": 0.6225, + "step": 6110 + }, + { + "epoch": 11.316666666666666, + "grad_norm": 2.785420159099274, + "learning_rate": 1.6828532485414724e-07, + "loss": 0.6413, + "step": 6111 + }, + { + "epoch": 11.318518518518518, + "grad_norm": 1.7407396121593037, + "learning_rate": 1.6827436996280007e-07, + "loss": 0.6234, + "step": 6112 + }, + { + "epoch": 11.32037037037037, + "grad_norm": 1.8272448599153395, + "learning_rate": 1.6826341353645965e-07, + "loss": 0.5869, + "step": 6113 + }, + { + "epoch": 11.322222222222223, + "grad_norm": 1.7073131512091284, + "learning_rate": 1.6825245557537228e-07, + "loss": 0.6266, + "step": 6114 + }, + { + "epoch": 11.324074074074074, + "grad_norm": 2.004167948563305, + "learning_rate": 1.6824149607978434e-07, + "loss": 0.6524, + "step": 6115 + }, + { + "epoch": 11.325925925925926, + "grad_norm": 1.5501093589508403, + "learning_rate": 1.6823053504994223e-07, + "loss": 0.6034, + "step": 6116 + }, + { + "epoch": 11.327777777777778, + "grad_norm": 1.7120958481742412, + "learning_rate": 1.6821957248609232e-07, + "loss": 0.6271, + "step": 6117 + }, + { + "epoch": 11.329629629629629, + "grad_norm": 1.6471374294874674, + "learning_rate": 1.6820860838848123e-07, + "loss": 0.6021, + "step": 6118 + }, + { + "epoch": 11.331481481481482, + "grad_norm": 1.6595997833019283, + "learning_rate": 1.6819764275735533e-07, + "loss": 0.6322, + "step": 6119 + }, + { + "epoch": 11.333333333333334, + "grad_norm": 1.487895686112585, + "learning_rate": 1.681866755929612e-07, + "loss": 0.6074, + "step": 6120 + }, + { + "epoch": 11.335185185185185, + "grad_norm": 1.5076138127711285, + "learning_rate": 1.6817570689554539e-07, + "loss": 0.6221, + "step": 6121 + }, + { + "epoch": 11.337037037037037, + "grad_norm": 1.3930525592568264, + "learning_rate": 1.6816473666535456e-07, + "loss": 0.6103, + "step": 6122 + }, + { + "epoch": 11.338888888888889, + "grad_norm": 1.8740855015633813, + "learning_rate": 1.6815376490263532e-07, + "loss": 0.6063, + "step": 6123 + }, + { + "epoch": 11.34074074074074, + "grad_norm": 1.4405692945863406, + "learning_rate": 1.6814279160763429e-07, + "loss": 0.6441, + "step": 6124 + }, + { + "epoch": 11.342592592592593, + "grad_norm": 1.6457619080252759, + "learning_rate": 1.6813181678059828e-07, + "loss": 0.5759, + "step": 6125 + }, + { + "epoch": 11.344444444444445, + "grad_norm": 1.7675443842494571, + "learning_rate": 1.68120840421774e-07, + "loss": 0.5969, + "step": 6126 + }, + { + "epoch": 11.346296296296297, + "grad_norm": 1.6809304572503099, + "learning_rate": 1.6810986253140819e-07, + "loss": 0.6271, + "step": 6127 + }, + { + "epoch": 11.348148148148148, + "grad_norm": 1.5544472841501098, + "learning_rate": 1.6809888310974768e-07, + "loss": 0.6307, + "step": 6128 + }, + { + "epoch": 11.35, + "grad_norm": 1.894822185973263, + "learning_rate": 1.6808790215703933e-07, + "loss": 0.6596, + "step": 6129 + }, + { + "epoch": 11.351851851851851, + "grad_norm": 3.549222274338036, + "learning_rate": 1.6807691967353003e-07, + "loss": 0.6223, + "step": 6130 + }, + { + "epoch": 11.353703703703705, + "grad_norm": 3.347767675781517, + "learning_rate": 1.6806593565946665e-07, + "loss": 0.6302, + "step": 6131 + }, + { + "epoch": 11.355555555555556, + "grad_norm": 1.7769978880340598, + "learning_rate": 1.6805495011509617e-07, + "loss": 0.6655, + "step": 6132 + }, + { + "epoch": 11.357407407407408, + "grad_norm": 1.4754736066991214, + "learning_rate": 1.680439630406656e-07, + "loss": 0.604, + "step": 6133 + }, + { + "epoch": 11.35925925925926, + "grad_norm": 1.5397310461845914, + "learning_rate": 1.6803297443642194e-07, + "loss": 0.6106, + "step": 6134 + }, + { + "epoch": 11.36111111111111, + "grad_norm": 2.111810597903459, + "learning_rate": 1.6802198430261222e-07, + "loss": 0.6363, + "step": 6135 + }, + { + "epoch": 11.362962962962962, + "grad_norm": 2.2462278560262567, + "learning_rate": 1.6801099263948357e-07, + "loss": 0.6385, + "step": 6136 + }, + { + "epoch": 11.364814814814816, + "grad_norm": 1.591875192794211, + "learning_rate": 1.6799999944728307e-07, + "loss": 0.6142, + "step": 6137 + }, + { + "epoch": 11.366666666666667, + "grad_norm": 1.4351378320019608, + "learning_rate": 1.679890047262579e-07, + "loss": 0.6516, + "step": 6138 + }, + { + "epoch": 11.368518518518519, + "grad_norm": 1.4311572403363906, + "learning_rate": 1.6797800847665527e-07, + "loss": 0.636, + "step": 6139 + }, + { + "epoch": 11.37037037037037, + "grad_norm": 1.3740052489398664, + "learning_rate": 1.6796701069872237e-07, + "loss": 0.5795, + "step": 6140 + }, + { + "epoch": 11.372222222222222, + "grad_norm": 1.5811911655716102, + "learning_rate": 1.6795601139270646e-07, + "loss": 0.5956, + "step": 6141 + }, + { + "epoch": 11.374074074074073, + "grad_norm": 1.4595634081644167, + "learning_rate": 1.6794501055885486e-07, + "loss": 0.6327, + "step": 6142 + }, + { + "epoch": 11.375925925925927, + "grad_norm": 2.65516898242403, + "learning_rate": 1.679340081974149e-07, + "loss": 0.6157, + "step": 6143 + }, + { + "epoch": 11.377777777777778, + "grad_norm": 1.6217768608074086, + "learning_rate": 1.6792300430863393e-07, + "loss": 0.5961, + "step": 6144 + }, + { + "epoch": 11.37962962962963, + "grad_norm": 1.5262965329906493, + "learning_rate": 1.679119988927593e-07, + "loss": 0.5999, + "step": 6145 + }, + { + "epoch": 11.381481481481481, + "grad_norm": 1.4136164959799644, + "learning_rate": 1.6790099195003854e-07, + "loss": 0.6184, + "step": 6146 + }, + { + "epoch": 11.383333333333333, + "grad_norm": 1.558848943983646, + "learning_rate": 1.6788998348071904e-07, + "loss": 0.7145, + "step": 6147 + }, + { + "epoch": 11.385185185185184, + "grad_norm": 1.8412967081728602, + "learning_rate": 1.6787897348504836e-07, + "loss": 0.6834, + "step": 6148 + }, + { + "epoch": 11.387037037037038, + "grad_norm": 1.5336734531670488, + "learning_rate": 1.6786796196327397e-07, + "loss": 0.6088, + "step": 6149 + }, + { + "epoch": 11.38888888888889, + "grad_norm": 4.010885856768812, + "learning_rate": 1.6785694891564345e-07, + "loss": 0.672, + "step": 6150 + }, + { + "epoch": 11.39074074074074, + "grad_norm": 1.4811792940981883, + "learning_rate": 1.6784593434240447e-07, + "loss": 0.6105, + "step": 6151 + }, + { + "epoch": 11.392592592592592, + "grad_norm": 1.9036486955386922, + "learning_rate": 1.6783491824380458e-07, + "loss": 0.5868, + "step": 6152 + }, + { + "epoch": 11.394444444444444, + "grad_norm": 1.458907816283198, + "learning_rate": 1.6782390062009148e-07, + "loss": 0.6029, + "step": 6153 + }, + { + "epoch": 11.396296296296295, + "grad_norm": 1.9278317363167718, + "learning_rate": 1.6781288147151295e-07, + "loss": 0.6459, + "step": 6154 + }, + { + "epoch": 11.398148148148149, + "grad_norm": 1.4924844805487636, + "learning_rate": 1.6780186079831662e-07, + "loss": 0.6504, + "step": 6155 + }, + { + "epoch": 11.4, + "grad_norm": 1.8339537601646378, + "learning_rate": 1.6779083860075033e-07, + "loss": 0.6086, + "step": 6156 + }, + { + "epoch": 11.401851851851852, + "grad_norm": 1.495523066543893, + "learning_rate": 1.6777981487906185e-07, + "loss": 0.5938, + "step": 6157 + }, + { + "epoch": 11.403703703703703, + "grad_norm": 1.5234592004415843, + "learning_rate": 1.6776878963349907e-07, + "loss": 0.6172, + "step": 6158 + }, + { + "epoch": 11.405555555555555, + "grad_norm": 1.525890855458275, + "learning_rate": 1.677577628643098e-07, + "loss": 0.6087, + "step": 6159 + }, + { + "epoch": 11.407407407407407, + "grad_norm": 1.5035218318707106, + "learning_rate": 1.6774673457174204e-07, + "loss": 0.6284, + "step": 6160 + }, + { + "epoch": 11.40925925925926, + "grad_norm": 2.3349176082004197, + "learning_rate": 1.6773570475604366e-07, + "loss": 0.62, + "step": 6161 + }, + { + "epoch": 11.411111111111111, + "grad_norm": 1.4882429513321542, + "learning_rate": 1.6772467341746266e-07, + "loss": 0.6622, + "step": 6162 + }, + { + "epoch": 11.412962962962963, + "grad_norm": 1.4547862053085523, + "learning_rate": 1.677136405562471e-07, + "loss": 0.599, + "step": 6163 + }, + { + "epoch": 11.414814814814815, + "grad_norm": 1.492584721341252, + "learning_rate": 1.6770260617264498e-07, + "loss": 0.6381, + "step": 6164 + }, + { + "epoch": 11.416666666666666, + "grad_norm": 1.569951233310649, + "learning_rate": 1.676915702669044e-07, + "loss": 0.6676, + "step": 6165 + }, + { + "epoch": 11.418518518518518, + "grad_norm": 1.6098508910170555, + "learning_rate": 1.6768053283927346e-07, + "loss": 0.628, + "step": 6166 + }, + { + "epoch": 11.420370370370371, + "grad_norm": 1.4088206554000426, + "learning_rate": 1.6766949389000034e-07, + "loss": 0.6037, + "step": 6167 + }, + { + "epoch": 11.422222222222222, + "grad_norm": 1.382604391730993, + "learning_rate": 1.6765845341933317e-07, + "loss": 0.5952, + "step": 6168 + }, + { + "epoch": 11.424074074074074, + "grad_norm": 1.719526958831739, + "learning_rate": 1.6764741142752027e-07, + "loss": 0.6198, + "step": 6169 + }, + { + "epoch": 11.425925925925926, + "grad_norm": 1.7198045333229666, + "learning_rate": 1.6763636791480978e-07, + "loss": 0.6039, + "step": 6170 + }, + { + "epoch": 11.427777777777777, + "grad_norm": 1.445981010339226, + "learning_rate": 1.676253228814501e-07, + "loss": 0.5835, + "step": 6171 + }, + { + "epoch": 11.42962962962963, + "grad_norm": 1.5541143313150898, + "learning_rate": 1.6761427632768944e-07, + "loss": 0.5839, + "step": 6172 + }, + { + "epoch": 11.431481481481482, + "grad_norm": 1.6137516598086628, + "learning_rate": 1.676032282537762e-07, + "loss": 0.5998, + "step": 6173 + }, + { + "epoch": 11.433333333333334, + "grad_norm": 1.559032224186864, + "learning_rate": 1.6759217865995883e-07, + "loss": 0.6313, + "step": 6174 + }, + { + "epoch": 11.435185185185185, + "grad_norm": 1.588872350946832, + "learning_rate": 1.675811275464857e-07, + "loss": 0.632, + "step": 6175 + }, + { + "epoch": 11.437037037037037, + "grad_norm": 1.5727046238120939, + "learning_rate": 1.6757007491360525e-07, + "loss": 0.5951, + "step": 6176 + }, + { + "epoch": 11.438888888888888, + "grad_norm": 1.4631903414229273, + "learning_rate": 1.6755902076156602e-07, + "loss": 0.6328, + "step": 6177 + }, + { + "epoch": 11.440740740740742, + "grad_norm": 1.8119439849567267, + "learning_rate": 1.675479650906165e-07, + "loss": 0.5837, + "step": 6178 + }, + { + "epoch": 11.442592592592593, + "grad_norm": 1.4689622619878557, + "learning_rate": 1.6753690790100528e-07, + "loss": 0.6366, + "step": 6179 + }, + { + "epoch": 11.444444444444445, + "grad_norm": 1.372096008281409, + "learning_rate": 1.6752584919298091e-07, + "loss": 0.6041, + "step": 6180 + }, + { + "epoch": 11.446296296296296, + "grad_norm": 1.3933387475011356, + "learning_rate": 1.675147889667921e-07, + "loss": 0.6458, + "step": 6181 + }, + { + "epoch": 11.448148148148148, + "grad_norm": 1.8071017325743397, + "learning_rate": 1.6750372722268743e-07, + "loss": 0.6249, + "step": 6182 + }, + { + "epoch": 11.45, + "grad_norm": 1.4771975121232566, + "learning_rate": 1.6749266396091568e-07, + "loss": 0.6295, + "step": 6183 + }, + { + "epoch": 11.451851851851853, + "grad_norm": 1.7932795890712372, + "learning_rate": 1.674815991817255e-07, + "loss": 0.6615, + "step": 6184 + }, + { + "epoch": 11.453703703703704, + "grad_norm": 1.6257021907341753, + "learning_rate": 1.674705328853657e-07, + "loss": 0.6371, + "step": 6185 + }, + { + "epoch": 11.455555555555556, + "grad_norm": 1.5260879723417615, + "learning_rate": 1.6745946507208508e-07, + "loss": 0.6589, + "step": 6186 + }, + { + "epoch": 11.457407407407407, + "grad_norm": 1.884503395385978, + "learning_rate": 1.6744839574213248e-07, + "loss": 0.6101, + "step": 6187 + }, + { + "epoch": 11.459259259259259, + "grad_norm": 1.3965438083799675, + "learning_rate": 1.674373248957567e-07, + "loss": 0.6009, + "step": 6188 + }, + { + "epoch": 11.46111111111111, + "grad_norm": 2.144178611322647, + "learning_rate": 1.6742625253320676e-07, + "loss": 0.6172, + "step": 6189 + }, + { + "epoch": 11.462962962962964, + "grad_norm": 1.4520723812939664, + "learning_rate": 1.674151786547315e-07, + "loss": 0.627, + "step": 6190 + }, + { + "epoch": 11.464814814814815, + "grad_norm": 1.6803221694476724, + "learning_rate": 1.6740410326057993e-07, + "loss": 0.6103, + "step": 6191 + }, + { + "epoch": 11.466666666666667, + "grad_norm": 1.8766644744783456, + "learning_rate": 1.6739302635100107e-07, + "loss": 0.5688, + "step": 6192 + }, + { + "epoch": 11.468518518518518, + "grad_norm": 1.6000935884917922, + "learning_rate": 1.6738194792624395e-07, + "loss": 0.594, + "step": 6193 + }, + { + "epoch": 11.47037037037037, + "grad_norm": 1.9444780893315359, + "learning_rate": 1.673708679865576e-07, + "loss": 0.6138, + "step": 6194 + }, + { + "epoch": 11.472222222222221, + "grad_norm": 1.3439450826248156, + "learning_rate": 1.6735978653219117e-07, + "loss": 0.6113, + "step": 6195 + }, + { + "epoch": 11.474074074074075, + "grad_norm": 1.8537407461103759, + "learning_rate": 1.6734870356339383e-07, + "loss": 0.6546, + "step": 6196 + }, + { + "epoch": 11.475925925925926, + "grad_norm": 1.62397838566437, + "learning_rate": 1.673376190804147e-07, + "loss": 0.5989, + "step": 6197 + }, + { + "epoch": 11.477777777777778, + "grad_norm": 1.8456980229644713, + "learning_rate": 1.6732653308350297e-07, + "loss": 0.5897, + "step": 6198 + }, + { + "epoch": 11.47962962962963, + "grad_norm": 1.5739771894128676, + "learning_rate": 1.6731544557290795e-07, + "loss": 0.5644, + "step": 6199 + }, + { + "epoch": 11.481481481481481, + "grad_norm": 1.460695117502979, + "learning_rate": 1.6730435654887888e-07, + "loss": 0.6162, + "step": 6200 + }, + { + "epoch": 11.483333333333333, + "grad_norm": 1.3832157119355897, + "learning_rate": 1.672932660116651e-07, + "loss": 0.623, + "step": 6201 + }, + { + "epoch": 11.485185185185186, + "grad_norm": 1.5460034997117744, + "learning_rate": 1.672821739615159e-07, + "loss": 0.6041, + "step": 6202 + }, + { + "epoch": 11.487037037037037, + "grad_norm": 1.698509160794228, + "learning_rate": 1.6727108039868072e-07, + "loss": 0.6197, + "step": 6203 + }, + { + "epoch": 11.488888888888889, + "grad_norm": 1.6067299769982353, + "learning_rate": 1.6725998532340896e-07, + "loss": 0.6304, + "step": 6204 + }, + { + "epoch": 11.49074074074074, + "grad_norm": 1.8519130467011935, + "learning_rate": 1.6724888873595e-07, + "loss": 0.659, + "step": 6205 + }, + { + "epoch": 11.492592592592592, + "grad_norm": 3.33231514201196, + "learning_rate": 1.6723779063655342e-07, + "loss": 0.6273, + "step": 6206 + }, + { + "epoch": 11.494444444444444, + "grad_norm": 1.522498027739761, + "learning_rate": 1.6722669102546868e-07, + "loss": 0.5818, + "step": 6207 + }, + { + "epoch": 11.496296296296297, + "grad_norm": 1.544644123030037, + "learning_rate": 1.6721558990294535e-07, + "loss": 0.5866, + "step": 6208 + }, + { + "epoch": 11.498148148148148, + "grad_norm": 1.3857098635412413, + "learning_rate": 1.67204487269233e-07, + "loss": 0.6215, + "step": 6209 + }, + { + "epoch": 11.5, + "grad_norm": 1.5846332488751016, + "learning_rate": 1.6719338312458123e-07, + "loss": 0.5937, + "step": 6210 + }, + { + "epoch": 11.501851851851852, + "grad_norm": 1.408714697606958, + "learning_rate": 1.6718227746923973e-07, + "loss": 0.5367, + "step": 6211 + }, + { + "epoch": 11.503703703703703, + "grad_norm": 1.458639915016249, + "learning_rate": 1.6717117030345817e-07, + "loss": 0.6252, + "step": 6212 + }, + { + "epoch": 11.505555555555556, + "grad_norm": 1.3651097127448846, + "learning_rate": 1.6716006162748627e-07, + "loss": 0.6101, + "step": 6213 + }, + { + "epoch": 11.507407407407408, + "grad_norm": 1.6055079286355587, + "learning_rate": 1.6714895144157378e-07, + "loss": 0.6243, + "step": 6214 + }, + { + "epoch": 11.50925925925926, + "grad_norm": 1.7737317745031491, + "learning_rate": 1.6713783974597048e-07, + "loss": 0.644, + "step": 6215 + }, + { + "epoch": 11.511111111111111, + "grad_norm": 1.6954493978407092, + "learning_rate": 1.6712672654092622e-07, + "loss": 0.6271, + "step": 6216 + }, + { + "epoch": 11.512962962962963, + "grad_norm": 1.4569796057330202, + "learning_rate": 1.671156118266908e-07, + "loss": 0.6184, + "step": 6217 + }, + { + "epoch": 11.514814814814814, + "grad_norm": 1.8402672842431333, + "learning_rate": 1.6710449560351413e-07, + "loss": 0.6382, + "step": 6218 + }, + { + "epoch": 11.516666666666667, + "grad_norm": 1.4684153407708058, + "learning_rate": 1.670933778716462e-07, + "loss": 0.595, + "step": 6219 + }, + { + "epoch": 11.518518518518519, + "grad_norm": 1.787665606412809, + "learning_rate": 1.670822586313369e-07, + "loss": 0.6503, + "step": 6220 + }, + { + "epoch": 11.52037037037037, + "grad_norm": 1.6469882618413847, + "learning_rate": 1.6707113788283623e-07, + "loss": 0.6431, + "step": 6221 + }, + { + "epoch": 11.522222222222222, + "grad_norm": 1.4877634020193584, + "learning_rate": 1.670600156263942e-07, + "loss": 0.6076, + "step": 6222 + }, + { + "epoch": 11.524074074074074, + "grad_norm": 1.6621082703447536, + "learning_rate": 1.670488918622609e-07, + "loss": 0.6271, + "step": 6223 + }, + { + "epoch": 11.525925925925925, + "grad_norm": 1.4264101397911912, + "learning_rate": 1.670377665906864e-07, + "loss": 0.6387, + "step": 6224 + }, + { + "epoch": 11.527777777777779, + "grad_norm": 1.5288064744522503, + "learning_rate": 1.6702663981192086e-07, + "loss": 0.641, + "step": 6225 + }, + { + "epoch": 11.52962962962963, + "grad_norm": 5.3506404582950715, + "learning_rate": 1.670155115262144e-07, + "loss": 0.633, + "step": 6226 + }, + { + "epoch": 11.531481481481482, + "grad_norm": 1.5853782458286365, + "learning_rate": 1.6700438173381718e-07, + "loss": 0.6663, + "step": 6227 + }, + { + "epoch": 11.533333333333333, + "grad_norm": 1.4896163997112069, + "learning_rate": 1.6699325043497953e-07, + "loss": 0.6218, + "step": 6228 + }, + { + "epoch": 11.535185185185185, + "grad_norm": 1.570512243772648, + "learning_rate": 1.6698211762995167e-07, + "loss": 0.6169, + "step": 6229 + }, + { + "epoch": 11.537037037037036, + "grad_norm": 1.6007949302657458, + "learning_rate": 1.6697098331898386e-07, + "loss": 0.6333, + "step": 6230 + }, + { + "epoch": 11.53888888888889, + "grad_norm": 1.4713092223231632, + "learning_rate": 1.6695984750232647e-07, + "loss": 0.6371, + "step": 6231 + }, + { + "epoch": 11.540740740740741, + "grad_norm": 1.636354499856578, + "learning_rate": 1.6694871018022982e-07, + "loss": 0.6696, + "step": 6232 + }, + { + "epoch": 11.542592592592593, + "grad_norm": 5.147914653041103, + "learning_rate": 1.6693757135294435e-07, + "loss": 0.6079, + "step": 6233 + }, + { + "epoch": 11.544444444444444, + "grad_norm": 2.3274418096888767, + "learning_rate": 1.669264310207205e-07, + "loss": 0.5809, + "step": 6234 + }, + { + "epoch": 11.546296296296296, + "grad_norm": 1.4832605201762499, + "learning_rate": 1.669152891838087e-07, + "loss": 0.6541, + "step": 6235 + }, + { + "epoch": 11.548148148148147, + "grad_norm": 1.5826397384281918, + "learning_rate": 1.669041458424594e-07, + "loss": 0.5781, + "step": 6236 + }, + { + "epoch": 11.55, + "grad_norm": 1.7137126589071063, + "learning_rate": 1.668930009969233e-07, + "loss": 0.6235, + "step": 6237 + }, + { + "epoch": 11.551851851851852, + "grad_norm": 1.349414839609121, + "learning_rate": 1.6688185464745078e-07, + "loss": 0.6113, + "step": 6238 + }, + { + "epoch": 11.553703703703704, + "grad_norm": 1.3085895969654802, + "learning_rate": 1.6687070679429252e-07, + "loss": 0.5878, + "step": 6239 + }, + { + "epoch": 11.555555555555555, + "grad_norm": 1.446915597307635, + "learning_rate": 1.668595574376992e-07, + "loss": 0.6184, + "step": 6240 + }, + { + "epoch": 11.557407407407407, + "grad_norm": 1.5722968347899688, + "learning_rate": 1.6684840657792142e-07, + "loss": 0.6345, + "step": 6241 + }, + { + "epoch": 11.559259259259258, + "grad_norm": 1.5380726948352605, + "learning_rate": 1.668372542152099e-07, + "loss": 0.6087, + "step": 6242 + }, + { + "epoch": 11.561111111111112, + "grad_norm": 1.5650068012898408, + "learning_rate": 1.668261003498154e-07, + "loss": 0.6339, + "step": 6243 + }, + { + "epoch": 11.562962962962963, + "grad_norm": 1.5564925554821079, + "learning_rate": 1.6681494498198863e-07, + "loss": 0.6375, + "step": 6244 + }, + { + "epoch": 11.564814814814815, + "grad_norm": 1.6524571938771602, + "learning_rate": 1.6680378811198044e-07, + "loss": 0.6117, + "step": 6245 + }, + { + "epoch": 11.566666666666666, + "grad_norm": 1.507276010440385, + "learning_rate": 1.6679262974004166e-07, + "loss": 0.6484, + "step": 6246 + }, + { + "epoch": 11.568518518518518, + "grad_norm": 1.3821805948791446, + "learning_rate": 1.6678146986642317e-07, + "loss": 0.6194, + "step": 6247 + }, + { + "epoch": 11.57037037037037, + "grad_norm": 1.4964064575704334, + "learning_rate": 1.6677030849137584e-07, + "loss": 0.6245, + "step": 6248 + }, + { + "epoch": 11.572222222222223, + "grad_norm": 1.5406692356599994, + "learning_rate": 1.6675914561515065e-07, + "loss": 0.6149, + "step": 6249 + }, + { + "epoch": 11.574074074074074, + "grad_norm": 1.5032393608953865, + "learning_rate": 1.6674798123799857e-07, + "loss": 0.6491, + "step": 6250 + }, + { + "epoch": 11.575925925925926, + "grad_norm": 1.5316813856504001, + "learning_rate": 1.6673681536017052e-07, + "loss": 0.5869, + "step": 6251 + }, + { + "epoch": 11.577777777777778, + "grad_norm": 1.5689300372691775, + "learning_rate": 1.6672564798191765e-07, + "loss": 0.5801, + "step": 6252 + }, + { + "epoch": 11.579629629629629, + "grad_norm": 1.5223383963615624, + "learning_rate": 1.66714479103491e-07, + "loss": 0.659, + "step": 6253 + }, + { + "epoch": 11.581481481481482, + "grad_norm": 1.414749129931455, + "learning_rate": 1.667033087251416e-07, + "loss": 0.648, + "step": 6254 + }, + { + "epoch": 11.583333333333334, + "grad_norm": 1.5157401555091015, + "learning_rate": 1.6669213684712076e-07, + "loss": 0.6609, + "step": 6255 + }, + { + "epoch": 11.585185185185185, + "grad_norm": 2.0440044451368022, + "learning_rate": 1.6668096346967948e-07, + "loss": 0.644, + "step": 6256 + }, + { + "epoch": 11.587037037037037, + "grad_norm": 2.0014545065171596, + "learning_rate": 1.6666978859306907e-07, + "loss": 0.6359, + "step": 6257 + }, + { + "epoch": 11.588888888888889, + "grad_norm": 1.591278352066013, + "learning_rate": 1.6665861221754072e-07, + "loss": 0.6195, + "step": 6258 + }, + { + "epoch": 11.59074074074074, + "grad_norm": 1.4294537946652528, + "learning_rate": 1.6664743434334571e-07, + "loss": 0.5972, + "step": 6259 + }, + { + "epoch": 11.592592592592592, + "grad_norm": 1.5720878263636167, + "learning_rate": 1.666362549707354e-07, + "loss": 0.6399, + "step": 6260 + }, + { + "epoch": 11.594444444444445, + "grad_norm": 1.6654548132295792, + "learning_rate": 1.6662507409996106e-07, + "loss": 0.6184, + "step": 6261 + }, + { + "epoch": 11.596296296296297, + "grad_norm": 1.6077963310588668, + "learning_rate": 1.6661389173127416e-07, + "loss": 0.6181, + "step": 6262 + }, + { + "epoch": 11.598148148148148, + "grad_norm": 2.5265451785654993, + "learning_rate": 1.66602707864926e-07, + "loss": 0.6424, + "step": 6263 + }, + { + "epoch": 11.6, + "grad_norm": 1.411964087451029, + "learning_rate": 1.6659152250116811e-07, + "loss": 0.6197, + "step": 6264 + }, + { + "epoch": 11.601851851851851, + "grad_norm": 1.4518107645800067, + "learning_rate": 1.6658033564025193e-07, + "loss": 0.6428, + "step": 6265 + }, + { + "epoch": 11.603703703703705, + "grad_norm": 1.888960815435029, + "learning_rate": 1.6656914728242894e-07, + "loss": 0.6464, + "step": 6266 + }, + { + "epoch": 11.605555555555556, + "grad_norm": 1.532147313626278, + "learning_rate": 1.6655795742795075e-07, + "loss": 0.5987, + "step": 6267 + }, + { + "epoch": 11.607407407407408, + "grad_norm": 1.6863579701247342, + "learning_rate": 1.6654676607706892e-07, + "loss": 0.6284, + "step": 6268 + }, + { + "epoch": 11.60925925925926, + "grad_norm": 1.5920634518351187, + "learning_rate": 1.6653557323003503e-07, + "loss": 0.6303, + "step": 6269 + }, + { + "epoch": 11.61111111111111, + "grad_norm": 1.4814739823819973, + "learning_rate": 1.6652437888710074e-07, + "loss": 0.6129, + "step": 6270 + }, + { + "epoch": 11.612962962962962, + "grad_norm": 1.8392935974418911, + "learning_rate": 1.6651318304851776e-07, + "loss": 0.5781, + "step": 6271 + }, + { + "epoch": 11.614814814814816, + "grad_norm": 1.558156713628762, + "learning_rate": 1.6650198571453777e-07, + "loss": 0.5946, + "step": 6272 + }, + { + "epoch": 11.616666666666667, + "grad_norm": 1.696504430959664, + "learning_rate": 1.6649078688541248e-07, + "loss": 0.6314, + "step": 6273 + }, + { + "epoch": 11.618518518518519, + "grad_norm": 1.4893108789756926, + "learning_rate": 1.6647958656139377e-07, + "loss": 0.6816, + "step": 6274 + }, + { + "epoch": 11.62037037037037, + "grad_norm": 1.492015978183013, + "learning_rate": 1.6646838474273338e-07, + "loss": 0.6302, + "step": 6275 + }, + { + "epoch": 11.622222222222222, + "grad_norm": 1.7225578348589157, + "learning_rate": 1.6645718142968318e-07, + "loss": 0.6325, + "step": 6276 + }, + { + "epoch": 11.624074074074073, + "grad_norm": 1.5960953620892344, + "learning_rate": 1.6644597662249505e-07, + "loss": 0.6014, + "step": 6277 + }, + { + "epoch": 11.625925925925927, + "grad_norm": 2.352903782285803, + "learning_rate": 1.6643477032142087e-07, + "loss": 0.6303, + "step": 6278 + }, + { + "epoch": 11.627777777777778, + "grad_norm": 1.7922186172721324, + "learning_rate": 1.6642356252671265e-07, + "loss": 0.6379, + "step": 6279 + }, + { + "epoch": 11.62962962962963, + "grad_norm": 1.4287209419383993, + "learning_rate": 1.6641235323862234e-07, + "loss": 0.609, + "step": 6280 + }, + { + "epoch": 11.631481481481481, + "grad_norm": 1.8647749058709564, + "learning_rate": 1.6640114245740197e-07, + "loss": 0.6495, + "step": 6281 + }, + { + "epoch": 11.633333333333333, + "grad_norm": 1.4511479062531873, + "learning_rate": 1.6638993018330357e-07, + "loss": 0.6148, + "step": 6282 + }, + { + "epoch": 11.635185185185184, + "grad_norm": 1.5046569050659722, + "learning_rate": 1.6637871641657918e-07, + "loss": 0.6323, + "step": 6283 + }, + { + "epoch": 11.637037037037038, + "grad_norm": 1.410894064080913, + "learning_rate": 1.66367501157481e-07, + "loss": 0.6452, + "step": 6284 + }, + { + "epoch": 11.63888888888889, + "grad_norm": 1.8972649381198574, + "learning_rate": 1.6635628440626115e-07, + "loss": 0.6288, + "step": 6285 + }, + { + "epoch": 11.64074074074074, + "grad_norm": 1.6573103161428755, + "learning_rate": 1.663450661631718e-07, + "loss": 0.6538, + "step": 6286 + }, + { + "epoch": 11.642592592592592, + "grad_norm": 1.7233344138384699, + "learning_rate": 1.6633384642846514e-07, + "loss": 0.5988, + "step": 6287 + }, + { + "epoch": 11.644444444444444, + "grad_norm": 1.6329573323226352, + "learning_rate": 1.6632262520239348e-07, + "loss": 0.6116, + "step": 6288 + }, + { + "epoch": 11.646296296296295, + "grad_norm": 1.8624530315234278, + "learning_rate": 1.6631140248520907e-07, + "loss": 0.6184, + "step": 6289 + }, + { + "epoch": 11.648148148148149, + "grad_norm": 1.6561223108206182, + "learning_rate": 1.6630017827716425e-07, + "loss": 0.6846, + "step": 6290 + }, + { + "epoch": 11.65, + "grad_norm": 1.512509148089906, + "learning_rate": 1.6628895257851135e-07, + "loss": 0.6243, + "step": 6291 + }, + { + "epoch": 11.651851851851852, + "grad_norm": 1.5409569091229534, + "learning_rate": 1.6627772538950274e-07, + "loss": 0.6318, + "step": 6292 + }, + { + "epoch": 11.653703703703703, + "grad_norm": 1.5468195934365614, + "learning_rate": 1.6626649671039085e-07, + "loss": 0.68, + "step": 6293 + }, + { + "epoch": 11.655555555555555, + "grad_norm": 1.4657078003740112, + "learning_rate": 1.6625526654142816e-07, + "loss": 0.6236, + "step": 6294 + }, + { + "epoch": 11.657407407407408, + "grad_norm": 1.574344917277673, + "learning_rate": 1.6624403488286713e-07, + "loss": 0.6011, + "step": 6295 + }, + { + "epoch": 11.65925925925926, + "grad_norm": 1.4207920157995821, + "learning_rate": 1.6623280173496027e-07, + "loss": 0.6454, + "step": 6296 + }, + { + "epoch": 11.661111111111111, + "grad_norm": 1.518694144377378, + "learning_rate": 1.6622156709796014e-07, + "loss": 0.6211, + "step": 6297 + }, + { + "epoch": 11.662962962962963, + "grad_norm": 1.6508759179831987, + "learning_rate": 1.6621033097211932e-07, + "loss": 0.6458, + "step": 6298 + }, + { + "epoch": 11.664814814814815, + "grad_norm": 1.4490673039572237, + "learning_rate": 1.6619909335769046e-07, + "loss": 0.6036, + "step": 6299 + }, + { + "epoch": 11.666666666666666, + "grad_norm": 1.694628014836841, + "learning_rate": 1.6618785425492616e-07, + "loss": 0.6582, + "step": 6300 + }, + { + "epoch": 11.668518518518518, + "grad_norm": 1.341980306918321, + "learning_rate": 1.6617661366407917e-07, + "loss": 0.6234, + "step": 6301 + }, + { + "epoch": 11.670370370370371, + "grad_norm": 1.5219227042071106, + "learning_rate": 1.6616537158540213e-07, + "loss": 0.6429, + "step": 6302 + }, + { + "epoch": 11.672222222222222, + "grad_norm": 1.5803125011634997, + "learning_rate": 1.6615412801914785e-07, + "loss": 0.6202, + "step": 6303 + }, + { + "epoch": 11.674074074074074, + "grad_norm": 1.5761684145214607, + "learning_rate": 1.661428829655691e-07, + "loss": 0.5893, + "step": 6304 + }, + { + "epoch": 11.675925925925926, + "grad_norm": 1.4905278628965921, + "learning_rate": 1.661316364249187e-07, + "loss": 0.5999, + "step": 6305 + }, + { + "epoch": 11.677777777777777, + "grad_norm": 1.3528540921139283, + "learning_rate": 1.6612038839744952e-07, + "loss": 0.6356, + "step": 6306 + }, + { + "epoch": 11.67962962962963, + "grad_norm": 3.4645169879649833, + "learning_rate": 1.661091388834144e-07, + "loss": 0.6071, + "step": 6307 + }, + { + "epoch": 11.681481481481482, + "grad_norm": 1.4887074547331722, + "learning_rate": 1.660978878830663e-07, + "loss": 0.6114, + "step": 6308 + }, + { + "epoch": 11.683333333333334, + "grad_norm": 1.3551729806318744, + "learning_rate": 1.6608663539665817e-07, + "loss": 0.6327, + "step": 6309 + }, + { + "epoch": 11.685185185185185, + "grad_norm": 1.427274671936576, + "learning_rate": 1.6607538142444298e-07, + "loss": 0.6071, + "step": 6310 + }, + { + "epoch": 11.687037037037037, + "grad_norm": 1.4034954578971253, + "learning_rate": 1.6606412596667376e-07, + "loss": 0.6769, + "step": 6311 + }, + { + "epoch": 11.688888888888888, + "grad_norm": 1.4105861666953468, + "learning_rate": 1.6605286902360357e-07, + "loss": 0.5926, + "step": 6312 + }, + { + "epoch": 11.690740740740742, + "grad_norm": 1.7708649118910318, + "learning_rate": 1.6604161059548548e-07, + "loss": 0.627, + "step": 6313 + }, + { + "epoch": 11.692592592592593, + "grad_norm": 2.2537683429355115, + "learning_rate": 1.6603035068257262e-07, + "loss": 0.6577, + "step": 6314 + }, + { + "epoch": 11.694444444444445, + "grad_norm": 1.9011910484698864, + "learning_rate": 1.6601908928511812e-07, + "loss": 0.6134, + "step": 6315 + }, + { + "epoch": 11.696296296296296, + "grad_norm": 1.8244869206006644, + "learning_rate": 1.6600782640337519e-07, + "loss": 0.6474, + "step": 6316 + }, + { + "epoch": 11.698148148148148, + "grad_norm": 1.4216886469031214, + "learning_rate": 1.6599656203759707e-07, + "loss": 0.6465, + "step": 6317 + }, + { + "epoch": 11.7, + "grad_norm": 1.508510795752423, + "learning_rate": 1.6598529618803698e-07, + "loss": 0.6472, + "step": 6318 + }, + { + "epoch": 11.701851851851853, + "grad_norm": 1.3838817303541242, + "learning_rate": 1.6597402885494823e-07, + "loss": 0.6232, + "step": 6319 + }, + { + "epoch": 11.703703703703704, + "grad_norm": 1.5044419428015603, + "learning_rate": 1.659627600385841e-07, + "loss": 0.6671, + "step": 6320 + }, + { + "epoch": 11.705555555555556, + "grad_norm": 1.6664036725195532, + "learning_rate": 1.6595148973919798e-07, + "loss": 0.6084, + "step": 6321 + }, + { + "epoch": 11.707407407407407, + "grad_norm": 1.6377579051601383, + "learning_rate": 1.6594021795704327e-07, + "loss": 0.6343, + "step": 6322 + }, + { + "epoch": 11.709259259259259, + "grad_norm": 1.7718999687809025, + "learning_rate": 1.6592894469237337e-07, + "loss": 0.5925, + "step": 6323 + }, + { + "epoch": 11.71111111111111, + "grad_norm": 1.8104779258889268, + "learning_rate": 1.6591766994544172e-07, + "loss": 0.6122, + "step": 6324 + }, + { + "epoch": 11.712962962962964, + "grad_norm": 1.667405357986239, + "learning_rate": 1.659063937165018e-07, + "loss": 0.62, + "step": 6325 + }, + { + "epoch": 11.714814814814815, + "grad_norm": 1.9286981132301937, + "learning_rate": 1.6589511600580722e-07, + "loss": 0.5899, + "step": 6326 + }, + { + "epoch": 11.716666666666667, + "grad_norm": 1.7416232824369224, + "learning_rate": 1.658838368136114e-07, + "loss": 0.5808, + "step": 6327 + }, + { + "epoch": 11.718518518518518, + "grad_norm": 1.5741337492859668, + "learning_rate": 1.65872556140168e-07, + "loss": 0.6458, + "step": 6328 + }, + { + "epoch": 11.72037037037037, + "grad_norm": 1.5139378624071405, + "learning_rate": 1.6586127398573064e-07, + "loss": 0.6071, + "step": 6329 + }, + { + "epoch": 11.722222222222221, + "grad_norm": 2.786104593113031, + "learning_rate": 1.6584999035055296e-07, + "loss": 0.6413, + "step": 6330 + }, + { + "epoch": 11.724074074074075, + "grad_norm": 1.56208165242815, + "learning_rate": 1.6583870523488864e-07, + "loss": 0.6052, + "step": 6331 + }, + { + "epoch": 11.725925925925926, + "grad_norm": 1.5084508657838658, + "learning_rate": 1.6582741863899147e-07, + "loss": 0.6292, + "step": 6332 + }, + { + "epoch": 11.727777777777778, + "grad_norm": 2.1625393344334496, + "learning_rate": 1.6581613056311508e-07, + "loss": 0.5942, + "step": 6333 + }, + { + "epoch": 11.72962962962963, + "grad_norm": 1.6970789008226272, + "learning_rate": 1.6580484100751334e-07, + "loss": 0.6188, + "step": 6334 + }, + { + "epoch": 11.731481481481481, + "grad_norm": 1.780945080553821, + "learning_rate": 1.6579354997244003e-07, + "loss": 0.6095, + "step": 6335 + }, + { + "epoch": 11.733333333333333, + "grad_norm": 1.6004377846071343, + "learning_rate": 1.6578225745814906e-07, + "loss": 0.639, + "step": 6336 + }, + { + "epoch": 11.735185185185186, + "grad_norm": 3.700453298327079, + "learning_rate": 1.6577096346489425e-07, + "loss": 0.6205, + "step": 6337 + }, + { + "epoch": 11.737037037037037, + "grad_norm": 2.0602346028089475, + "learning_rate": 1.657596679929296e-07, + "loss": 0.5841, + "step": 6338 + }, + { + "epoch": 11.738888888888889, + "grad_norm": 1.7754394071021233, + "learning_rate": 1.6574837104250894e-07, + "loss": 0.6442, + "step": 6339 + }, + { + "epoch": 11.74074074074074, + "grad_norm": 3.5535966782735025, + "learning_rate": 1.6573707261388638e-07, + "loss": 0.6182, + "step": 6340 + }, + { + "epoch": 11.742592592592592, + "grad_norm": 1.5456730744562626, + "learning_rate": 1.6572577270731587e-07, + "loss": 0.6644, + "step": 6341 + }, + { + "epoch": 11.744444444444444, + "grad_norm": 1.8826391635246373, + "learning_rate": 1.657144713230515e-07, + "loss": 0.616, + "step": 6342 + }, + { + "epoch": 11.746296296296297, + "grad_norm": 1.6971899099723748, + "learning_rate": 1.657031684613473e-07, + "loss": 0.6138, + "step": 6343 + }, + { + "epoch": 11.748148148148148, + "grad_norm": 1.7407975194083312, + "learning_rate": 1.656918641224575e-07, + "loss": 0.6241, + "step": 6344 + }, + { + "epoch": 11.75, + "grad_norm": 1.5385458587150036, + "learning_rate": 1.656805583066361e-07, + "loss": 0.5953, + "step": 6345 + }, + { + "epoch": 11.751851851851852, + "grad_norm": 1.4587167880716247, + "learning_rate": 1.656692510141374e-07, + "loss": 0.6087, + "step": 6346 + }, + { + "epoch": 11.753703703703703, + "grad_norm": 1.585098053175455, + "learning_rate": 1.6565794224521558e-07, + "loss": 0.629, + "step": 6347 + }, + { + "epoch": 11.755555555555556, + "grad_norm": 1.5673209210224417, + "learning_rate": 1.6564663200012487e-07, + "loss": 0.635, + "step": 6348 + }, + { + "epoch": 11.757407407407408, + "grad_norm": 1.4972051388538132, + "learning_rate": 1.656353202791196e-07, + "loss": 0.6328, + "step": 6349 + }, + { + "epoch": 11.75925925925926, + "grad_norm": 1.5036532459008225, + "learning_rate": 1.6562400708245405e-07, + "loss": 0.6156, + "step": 6350 + }, + { + "epoch": 11.761111111111111, + "grad_norm": 1.5707943099956645, + "learning_rate": 1.656126924103826e-07, + "loss": 0.6007, + "step": 6351 + }, + { + "epoch": 11.762962962962963, + "grad_norm": 1.4998203632117961, + "learning_rate": 1.6560137626315963e-07, + "loss": 0.6111, + "step": 6352 + }, + { + "epoch": 11.764814814814814, + "grad_norm": 1.307097115620002, + "learning_rate": 1.6559005864103958e-07, + "loss": 0.5775, + "step": 6353 + }, + { + "epoch": 11.766666666666667, + "grad_norm": 1.6128359340919014, + "learning_rate": 1.655787395442768e-07, + "loss": 0.6146, + "step": 6354 + }, + { + "epoch": 11.768518518518519, + "grad_norm": 1.6588181344785153, + "learning_rate": 1.655674189731259e-07, + "loss": 0.6638, + "step": 6355 + }, + { + "epoch": 11.77037037037037, + "grad_norm": 1.5587662598560095, + "learning_rate": 1.6555609692784133e-07, + "loss": 0.6134, + "step": 6356 + }, + { + "epoch": 11.772222222222222, + "grad_norm": 1.4800693262232518, + "learning_rate": 1.6554477340867766e-07, + "loss": 0.6701, + "step": 6357 + }, + { + "epoch": 11.774074074074074, + "grad_norm": 1.5544505860135458, + "learning_rate": 1.6553344841588946e-07, + "loss": 0.6161, + "step": 6358 + }, + { + "epoch": 11.775925925925925, + "grad_norm": 1.6498987677226291, + "learning_rate": 1.6552212194973133e-07, + "loss": 0.6489, + "step": 6359 + }, + { + "epoch": 11.777777777777779, + "grad_norm": 2.2497584476600174, + "learning_rate": 1.6551079401045797e-07, + "loss": 0.6256, + "step": 6360 + }, + { + "epoch": 11.77962962962963, + "grad_norm": 1.5613257878979847, + "learning_rate": 1.6549946459832402e-07, + "loss": 0.6436, + "step": 6361 + }, + { + "epoch": 11.781481481481482, + "grad_norm": 1.3857154439679875, + "learning_rate": 1.6548813371358424e-07, + "loss": 0.6506, + "step": 6362 + }, + { + "epoch": 11.783333333333333, + "grad_norm": 1.8815018983415994, + "learning_rate": 1.6547680135649334e-07, + "loss": 0.6418, + "step": 6363 + }, + { + "epoch": 11.785185185185185, + "grad_norm": 1.6291586631812849, + "learning_rate": 1.6546546752730607e-07, + "loss": 0.6661, + "step": 6364 + }, + { + "epoch": 11.787037037037036, + "grad_norm": 1.7625348141739272, + "learning_rate": 1.6545413222627733e-07, + "loss": 0.6041, + "step": 6365 + }, + { + "epoch": 11.78888888888889, + "grad_norm": 1.4799158452851364, + "learning_rate": 1.6544279545366196e-07, + "loss": 0.6138, + "step": 6366 + }, + { + "epoch": 11.790740740740741, + "grad_norm": 1.5219737854114233, + "learning_rate": 1.6543145720971475e-07, + "loss": 0.6465, + "step": 6367 + }, + { + "epoch": 11.792592592592593, + "grad_norm": 1.8935439148283575, + "learning_rate": 1.654201174946907e-07, + "loss": 0.6542, + "step": 6368 + }, + { + "epoch": 11.794444444444444, + "grad_norm": 1.6798835953498865, + "learning_rate": 1.654087763088447e-07, + "loss": 0.6464, + "step": 6369 + }, + { + "epoch": 11.796296296296296, + "grad_norm": 1.5619914647731885, + "learning_rate": 1.653974336524318e-07, + "loss": 0.5798, + "step": 6370 + }, + { + "epoch": 11.798148148148147, + "grad_norm": 2.2875975378425486, + "learning_rate": 1.6538608952570695e-07, + "loss": 0.6229, + "step": 6371 + }, + { + "epoch": 11.8, + "grad_norm": 1.5132196201133798, + "learning_rate": 1.6537474392892526e-07, + "loss": 0.5765, + "step": 6372 + }, + { + "epoch": 11.801851851851852, + "grad_norm": 1.5367397654286648, + "learning_rate": 1.6536339686234172e-07, + "loss": 0.6321, + "step": 6373 + }, + { + "epoch": 11.803703703703704, + "grad_norm": 1.5893518775819409, + "learning_rate": 1.6535204832621152e-07, + "loss": 0.6147, + "step": 6374 + }, + { + "epoch": 11.805555555555555, + "grad_norm": 1.6905258680258268, + "learning_rate": 1.653406983207898e-07, + "loss": 0.6269, + "step": 6375 + }, + { + "epoch": 11.807407407407407, + "grad_norm": 1.8443198418039966, + "learning_rate": 1.6532934684633165e-07, + "loss": 0.5844, + "step": 6376 + }, + { + "epoch": 11.809259259259258, + "grad_norm": 1.5548908118727325, + "learning_rate": 1.6531799390309242e-07, + "loss": 0.6114, + "step": 6377 + }, + { + "epoch": 11.811111111111112, + "grad_norm": 1.4711181296165419, + "learning_rate": 1.6530663949132726e-07, + "loss": 0.5984, + "step": 6378 + }, + { + "epoch": 11.812962962962963, + "grad_norm": 1.4625822993331419, + "learning_rate": 1.6529528361129147e-07, + "loss": 0.6172, + "step": 6379 + }, + { + "epoch": 11.814814814814815, + "grad_norm": 1.4773707975159378, + "learning_rate": 1.6528392626324038e-07, + "loss": 0.6135, + "step": 6380 + }, + { + "epoch": 11.816666666666666, + "grad_norm": 1.367464785831325, + "learning_rate": 1.652725674474293e-07, + "loss": 0.608, + "step": 6381 + }, + { + "epoch": 11.818518518518518, + "grad_norm": 1.4913928560400456, + "learning_rate": 1.6526120716411364e-07, + "loss": 0.616, + "step": 6382 + }, + { + "epoch": 11.82037037037037, + "grad_norm": 1.4295712749181289, + "learning_rate": 1.6524984541354882e-07, + "loss": 0.6174, + "step": 6383 + }, + { + "epoch": 11.822222222222223, + "grad_norm": 1.4961612361046086, + "learning_rate": 1.652384821959902e-07, + "loss": 0.5906, + "step": 6384 + }, + { + "epoch": 11.824074074074074, + "grad_norm": 1.6656333207454643, + "learning_rate": 1.6522711751169335e-07, + "loss": 0.6463, + "step": 6385 + }, + { + "epoch": 11.825925925925926, + "grad_norm": 1.6934623543651797, + "learning_rate": 1.6521575136091376e-07, + "loss": 0.6087, + "step": 6386 + }, + { + "epoch": 11.827777777777778, + "grad_norm": 1.769327561984438, + "learning_rate": 1.6520438374390694e-07, + "loss": 0.6343, + "step": 6387 + }, + { + "epoch": 11.829629629629629, + "grad_norm": 1.4703750743483683, + "learning_rate": 1.6519301466092847e-07, + "loss": 0.6376, + "step": 6388 + }, + { + "epoch": 11.831481481481482, + "grad_norm": 1.6504806922853015, + "learning_rate": 1.6518164411223402e-07, + "loss": 0.589, + "step": 6389 + }, + { + "epoch": 11.833333333333334, + "grad_norm": 1.673674541719744, + "learning_rate": 1.651702720980791e-07, + "loss": 0.6217, + "step": 6390 + }, + { + "epoch": 11.835185185185185, + "grad_norm": 1.5456181809101863, + "learning_rate": 1.6515889861871953e-07, + "loss": 0.621, + "step": 6391 + }, + { + "epoch": 11.837037037037037, + "grad_norm": 1.9772276254624102, + "learning_rate": 1.6514752367441092e-07, + "loss": 0.642, + "step": 6392 + }, + { + "epoch": 11.838888888888889, + "grad_norm": 1.7345261083362766, + "learning_rate": 1.6513614726540902e-07, + "loss": 0.6316, + "step": 6393 + }, + { + "epoch": 11.84074074074074, + "grad_norm": 1.6205672714711568, + "learning_rate": 1.6512476939196963e-07, + "loss": 0.6411, + "step": 6394 + }, + { + "epoch": 11.842592592592592, + "grad_norm": 1.556559430663859, + "learning_rate": 1.6511339005434856e-07, + "loss": 0.6244, + "step": 6395 + }, + { + "epoch": 11.844444444444445, + "grad_norm": 1.552097635584841, + "learning_rate": 1.6510200925280167e-07, + "loss": 0.6326, + "step": 6396 + }, + { + "epoch": 11.846296296296297, + "grad_norm": 1.592237644596314, + "learning_rate": 1.6509062698758475e-07, + "loss": 0.5993, + "step": 6397 + }, + { + "epoch": 11.848148148148148, + "grad_norm": 1.526142095194442, + "learning_rate": 1.6507924325895377e-07, + "loss": 0.6187, + "step": 6398 + }, + { + "epoch": 11.85, + "grad_norm": 1.7550401459096354, + "learning_rate": 1.6506785806716462e-07, + "loss": 0.603, + "step": 6399 + }, + { + "epoch": 11.851851851851851, + "grad_norm": 1.8471304883597084, + "learning_rate": 1.6505647141247337e-07, + "loss": 0.6445, + "step": 6400 + }, + { + "epoch": 11.853703703703705, + "grad_norm": 2.0399302968915523, + "learning_rate": 1.6504508329513591e-07, + "loss": 0.6329, + "step": 6401 + }, + { + "epoch": 11.855555555555556, + "grad_norm": 1.5654836953785103, + "learning_rate": 1.650336937154083e-07, + "loss": 0.6249, + "step": 6402 + }, + { + "epoch": 11.857407407407408, + "grad_norm": 1.5307159652017457, + "learning_rate": 1.6502230267354661e-07, + "loss": 0.6279, + "step": 6403 + }, + { + "epoch": 11.85925925925926, + "grad_norm": 1.499024291128072, + "learning_rate": 1.6501091016980704e-07, + "loss": 0.6263, + "step": 6404 + }, + { + "epoch": 11.86111111111111, + "grad_norm": 1.496073376617333, + "learning_rate": 1.6499951620444556e-07, + "loss": 0.6503, + "step": 6405 + }, + { + "epoch": 11.862962962962962, + "grad_norm": 1.5666873205917298, + "learning_rate": 1.6498812077771847e-07, + "loss": 0.6003, + "step": 6406 + }, + { + "epoch": 11.864814814814816, + "grad_norm": 1.5694465041331427, + "learning_rate": 1.6497672388988192e-07, + "loss": 0.6126, + "step": 6407 + }, + { + "epoch": 11.866666666666667, + "grad_norm": 1.577833565019944, + "learning_rate": 1.6496532554119212e-07, + "loss": 0.6522, + "step": 6408 + }, + { + "epoch": 11.868518518518519, + "grad_norm": 1.621700573035634, + "learning_rate": 1.6495392573190534e-07, + "loss": 0.6126, + "step": 6409 + }, + { + "epoch": 11.87037037037037, + "grad_norm": 1.5300422087108332, + "learning_rate": 1.649425244622779e-07, + "loss": 0.6288, + "step": 6410 + }, + { + "epoch": 11.872222222222222, + "grad_norm": 1.433119099736639, + "learning_rate": 1.6493112173256618e-07, + "loss": 0.6306, + "step": 6411 + }, + { + "epoch": 11.874074074074073, + "grad_norm": 1.5601472470627655, + "learning_rate": 1.6491971754302645e-07, + "loss": 0.6176, + "step": 6412 + }, + { + "epoch": 11.875925925925927, + "grad_norm": 1.6823507070755834, + "learning_rate": 1.649083118939152e-07, + "loss": 0.6166, + "step": 6413 + }, + { + "epoch": 11.877777777777778, + "grad_norm": 1.7711919611390672, + "learning_rate": 1.6489690478548872e-07, + "loss": 0.6251, + "step": 6414 + }, + { + "epoch": 11.87962962962963, + "grad_norm": 1.654190824821072, + "learning_rate": 1.6488549621800366e-07, + "loss": 0.6132, + "step": 6415 + }, + { + "epoch": 11.881481481481481, + "grad_norm": 1.3733726237488841, + "learning_rate": 1.6487408619171637e-07, + "loss": 0.6402, + "step": 6416 + }, + { + "epoch": 11.883333333333333, + "grad_norm": 1.5511284366995772, + "learning_rate": 1.6486267470688343e-07, + "loss": 0.635, + "step": 6417 + }, + { + "epoch": 11.885185185185184, + "grad_norm": 1.799343521264771, + "learning_rate": 1.648512617637614e-07, + "loss": 0.6296, + "step": 6418 + }, + { + "epoch": 11.887037037037038, + "grad_norm": 1.7945471056273337, + "learning_rate": 1.6483984736260688e-07, + "loss": 0.6819, + "step": 6419 + }, + { + "epoch": 11.88888888888889, + "grad_norm": 1.5383190163406673, + "learning_rate": 1.6482843150367647e-07, + "loss": 0.6097, + "step": 6420 + }, + { + "epoch": 11.89074074074074, + "grad_norm": 1.5212908678172758, + "learning_rate": 1.6481701418722685e-07, + "loss": 0.6487, + "step": 6421 + }, + { + "epoch": 11.892592592592592, + "grad_norm": 1.660658174900972, + "learning_rate": 1.6480559541351472e-07, + "loss": 0.6241, + "step": 6422 + }, + { + "epoch": 11.894444444444444, + "grad_norm": 1.4643803939371764, + "learning_rate": 1.6479417518279682e-07, + "loss": 0.6447, + "step": 6423 + }, + { + "epoch": 11.896296296296295, + "grad_norm": 1.462298842236054, + "learning_rate": 1.6478275349532984e-07, + "loss": 0.6325, + "step": 6424 + }, + { + "epoch": 11.898148148148149, + "grad_norm": 1.444820539070176, + "learning_rate": 1.6477133035137065e-07, + "loss": 0.6125, + "step": 6425 + }, + { + "epoch": 11.9, + "grad_norm": 1.4969345087378343, + "learning_rate": 1.6475990575117603e-07, + "loss": 0.6038, + "step": 6426 + }, + { + "epoch": 11.901851851851852, + "grad_norm": 1.6687835908810338, + "learning_rate": 1.6474847969500282e-07, + "loss": 0.5929, + "step": 6427 + }, + { + "epoch": 11.903703703703703, + "grad_norm": 1.9775388023354206, + "learning_rate": 1.6473705218310796e-07, + "loss": 0.622, + "step": 6428 + }, + { + "epoch": 11.905555555555555, + "grad_norm": 1.484128041964014, + "learning_rate": 1.647256232157483e-07, + "loss": 0.6668, + "step": 6429 + }, + { + "epoch": 11.907407407407408, + "grad_norm": 1.6213755769917255, + "learning_rate": 1.647141927931809e-07, + "loss": 0.641, + "step": 6430 + }, + { + "epoch": 11.90925925925926, + "grad_norm": 1.4009184893324613, + "learning_rate": 1.6470276091566265e-07, + "loss": 0.6553, + "step": 6431 + }, + { + "epoch": 11.911111111111111, + "grad_norm": 1.4716907526467358, + "learning_rate": 1.6469132758345058e-07, + "loss": 0.682, + "step": 6432 + }, + { + "epoch": 11.912962962962963, + "grad_norm": 1.6112091580727244, + "learning_rate": 1.6467989279680182e-07, + "loss": 0.5941, + "step": 6433 + }, + { + "epoch": 11.914814814814815, + "grad_norm": 1.562277566432166, + "learning_rate": 1.6466845655597336e-07, + "loss": 0.6399, + "step": 6434 + }, + { + "epoch": 11.916666666666666, + "grad_norm": 2.1170477161998447, + "learning_rate": 1.646570188612224e-07, + "loss": 0.6729, + "step": 6435 + }, + { + "epoch": 11.918518518518518, + "grad_norm": 1.6563289345287175, + "learning_rate": 1.6464557971280603e-07, + "loss": 0.6026, + "step": 6436 + }, + { + "epoch": 11.920370370370371, + "grad_norm": 1.389748994518346, + "learning_rate": 1.6463413911098145e-07, + "loss": 0.6546, + "step": 6437 + }, + { + "epoch": 11.922222222222222, + "grad_norm": 2.603387602889353, + "learning_rate": 1.646226970560059e-07, + "loss": 0.6188, + "step": 6438 + }, + { + "epoch": 11.924074074074074, + "grad_norm": 1.4762220057057864, + "learning_rate": 1.6461125354813658e-07, + "loss": 0.5928, + "step": 6439 + }, + { + "epoch": 11.925925925925926, + "grad_norm": 1.5134537961311547, + "learning_rate": 1.6459980858763077e-07, + "loss": 0.6077, + "step": 6440 + }, + { + "epoch": 11.927777777777777, + "grad_norm": 1.5297984512052334, + "learning_rate": 1.6458836217474587e-07, + "loss": 0.6218, + "step": 6441 + }, + { + "epoch": 11.92962962962963, + "grad_norm": 1.5873374590114824, + "learning_rate": 1.6457691430973915e-07, + "loss": 0.691, + "step": 6442 + }, + { + "epoch": 11.931481481481482, + "grad_norm": 1.3892032431167953, + "learning_rate": 1.6456546499286802e-07, + "loss": 0.6339, + "step": 6443 + }, + { + "epoch": 11.933333333333334, + "grad_norm": 1.5069824120180466, + "learning_rate": 1.6455401422438983e-07, + "loss": 0.66, + "step": 6444 + }, + { + "epoch": 11.935185185185185, + "grad_norm": 2.3065745219298672, + "learning_rate": 1.6454256200456214e-07, + "loss": 0.6281, + "step": 6445 + }, + { + "epoch": 11.937037037037037, + "grad_norm": 1.6815702486090638, + "learning_rate": 1.6453110833364233e-07, + "loss": 0.6023, + "step": 6446 + }, + { + "epoch": 11.938888888888888, + "grad_norm": 1.4346294580377503, + "learning_rate": 1.6451965321188794e-07, + "loss": 0.5717, + "step": 6447 + }, + { + "epoch": 11.940740740740742, + "grad_norm": 1.470746165912409, + "learning_rate": 1.6450819663955652e-07, + "loss": 0.6327, + "step": 6448 + }, + { + "epoch": 11.942592592592593, + "grad_norm": 1.7116026202587127, + "learning_rate": 1.6449673861690563e-07, + "loss": 0.6122, + "step": 6449 + }, + { + "epoch": 11.944444444444445, + "grad_norm": 1.7770710205614189, + "learning_rate": 1.6448527914419287e-07, + "loss": 0.6364, + "step": 6450 + }, + { + "epoch": 11.946296296296296, + "grad_norm": 2.6465345108893215, + "learning_rate": 1.6447381822167593e-07, + "loss": 0.6497, + "step": 6451 + }, + { + "epoch": 11.948148148148148, + "grad_norm": 1.7417576887351762, + "learning_rate": 1.6446235584961243e-07, + "loss": 0.5936, + "step": 6452 + }, + { + "epoch": 11.95, + "grad_norm": 1.522855779940423, + "learning_rate": 1.644508920282601e-07, + "loss": 0.6318, + "step": 6453 + }, + { + "epoch": 11.951851851851853, + "grad_norm": 1.4689981018970466, + "learning_rate": 1.6443942675787669e-07, + "loss": 0.6245, + "step": 6454 + }, + { + "epoch": 11.953703703703704, + "grad_norm": 2.1994593634690287, + "learning_rate": 1.644279600387199e-07, + "loss": 0.6404, + "step": 6455 + }, + { + "epoch": 11.955555555555556, + "grad_norm": 1.92817434977043, + "learning_rate": 1.644164918710476e-07, + "loss": 0.6261, + "step": 6456 + }, + { + "epoch": 11.957407407407407, + "grad_norm": 1.7593196941670797, + "learning_rate": 1.6440502225511766e-07, + "loss": 0.6002, + "step": 6457 + }, + { + "epoch": 11.959259259259259, + "grad_norm": 1.5181922385188862, + "learning_rate": 1.6439355119118785e-07, + "loss": 0.6328, + "step": 6458 + }, + { + "epoch": 11.96111111111111, + "grad_norm": 1.8758601171031275, + "learning_rate": 1.6438207867951613e-07, + "loss": 0.6035, + "step": 6459 + }, + { + "epoch": 11.962962962962964, + "grad_norm": 1.6212827469874653, + "learning_rate": 1.6437060472036044e-07, + "loss": 0.6333, + "step": 6460 + }, + { + "epoch": 11.964814814814815, + "grad_norm": 1.6884028365172763, + "learning_rate": 1.643591293139787e-07, + "loss": 0.6122, + "step": 6461 + }, + { + "epoch": 11.966666666666667, + "grad_norm": 1.6183811448413439, + "learning_rate": 1.6434765246062892e-07, + "loss": 0.6224, + "step": 6462 + }, + { + "epoch": 11.968518518518518, + "grad_norm": 1.5665496644766947, + "learning_rate": 1.643361741605692e-07, + "loss": 0.6304, + "step": 6463 + }, + { + "epoch": 11.97037037037037, + "grad_norm": 1.648530854681906, + "learning_rate": 1.643246944140575e-07, + "loss": 0.6428, + "step": 6464 + }, + { + "epoch": 11.972222222222221, + "grad_norm": 1.7914338123963771, + "learning_rate": 1.64313213221352e-07, + "loss": 0.5859, + "step": 6465 + }, + { + "epoch": 11.974074074074075, + "grad_norm": 1.514763172107704, + "learning_rate": 1.643017305827108e-07, + "loss": 0.6261, + "step": 6466 + }, + { + "epoch": 11.975925925925926, + "grad_norm": 1.5238718459025398, + "learning_rate": 1.6429024649839206e-07, + "loss": 0.6499, + "step": 6467 + }, + { + "epoch": 11.977777777777778, + "grad_norm": 1.6218285052342947, + "learning_rate": 1.6427876096865392e-07, + "loss": 0.5857, + "step": 6468 + }, + { + "epoch": 11.97962962962963, + "grad_norm": 1.6002142723997892, + "learning_rate": 1.6426727399375468e-07, + "loss": 0.6491, + "step": 6469 + }, + { + "epoch": 11.981481481481481, + "grad_norm": 1.409323510977805, + "learning_rate": 1.642557855739526e-07, + "loss": 0.6302, + "step": 6470 + }, + { + "epoch": 11.983333333333333, + "grad_norm": 1.7083831671987106, + "learning_rate": 1.642442957095059e-07, + "loss": 0.6142, + "step": 6471 + }, + { + "epoch": 11.985185185185186, + "grad_norm": 1.724358928249825, + "learning_rate": 1.64232804400673e-07, + "loss": 0.6108, + "step": 6472 + }, + { + "epoch": 11.987037037037037, + "grad_norm": 1.6703968721146336, + "learning_rate": 1.6422131164771216e-07, + "loss": 0.6083, + "step": 6473 + }, + { + "epoch": 11.988888888888889, + "grad_norm": 1.725340535329378, + "learning_rate": 1.6420981745088182e-07, + "loss": 0.6127, + "step": 6474 + }, + { + "epoch": 11.99074074074074, + "grad_norm": 3.6813504096107987, + "learning_rate": 1.641983218104404e-07, + "loss": 0.6151, + "step": 6475 + }, + { + "epoch": 11.992592592592592, + "grad_norm": 1.804433500069158, + "learning_rate": 1.641868247266464e-07, + "loss": 0.6267, + "step": 6476 + }, + { + "epoch": 11.994444444444444, + "grad_norm": 1.4259147593166637, + "learning_rate": 1.6417532619975813e-07, + "loss": 0.6123, + "step": 6477 + }, + { + "epoch": 11.996296296296297, + "grad_norm": 1.7361123435008383, + "learning_rate": 1.6416382623003431e-07, + "loss": 0.6257, + "step": 6478 + }, + { + "epoch": 11.998148148148148, + "grad_norm": 1.887508638725493, + "learning_rate": 1.641523248177334e-07, + "loss": 0.6214, + "step": 6479 + }, + { + "epoch": 12.0, + "grad_norm": 1.784575495939203, + "learning_rate": 1.64140821963114e-07, + "loss": 0.6292, + "step": 6480 + }, + { + "epoch": 12.001851851851852, + "grad_norm": 2.2787701584976907, + "learning_rate": 1.641293176664347e-07, + "loss": 0.5842, + "step": 6481 + }, + { + "epoch": 12.003703703703703, + "grad_norm": 1.4810894806085335, + "learning_rate": 1.6411781192795418e-07, + "loss": 0.6291, + "step": 6482 + }, + { + "epoch": 12.005555555555556, + "grad_norm": 1.817293272323697, + "learning_rate": 1.641063047479311e-07, + "loss": 0.6458, + "step": 6483 + }, + { + "epoch": 12.007407407407408, + "grad_norm": 1.853385663599283, + "learning_rate": 1.640947961266242e-07, + "loss": 0.6456, + "step": 6484 + }, + { + "epoch": 12.00925925925926, + "grad_norm": 1.6402491965792916, + "learning_rate": 1.6408328606429216e-07, + "loss": 0.6472, + "step": 6485 + }, + { + "epoch": 12.011111111111111, + "grad_norm": 1.6526928363550455, + "learning_rate": 1.6407177456119383e-07, + "loss": 0.6197, + "step": 6486 + }, + { + "epoch": 12.012962962962963, + "grad_norm": 1.6501295977234771, + "learning_rate": 1.64060261617588e-07, + "loss": 0.5992, + "step": 6487 + }, + { + "epoch": 12.014814814814814, + "grad_norm": 1.4693232790439867, + "learning_rate": 1.6404874723373345e-07, + "loss": 0.6108, + "step": 6488 + }, + { + "epoch": 12.016666666666667, + "grad_norm": 1.6787383319659326, + "learning_rate": 1.6403723140988914e-07, + "loss": 0.5942, + "step": 6489 + }, + { + "epoch": 12.018518518518519, + "grad_norm": 1.6162219170357184, + "learning_rate": 1.64025714146314e-07, + "loss": 0.6001, + "step": 6490 + }, + { + "epoch": 12.02037037037037, + "grad_norm": 1.7904313042001472, + "learning_rate": 1.6401419544326682e-07, + "loss": 0.6572, + "step": 6491 + }, + { + "epoch": 12.022222222222222, + "grad_norm": 1.543355254360909, + "learning_rate": 1.6400267530100674e-07, + "loss": 0.6286, + "step": 6492 + }, + { + "epoch": 12.024074074074074, + "grad_norm": 1.6540914056004015, + "learning_rate": 1.6399115371979267e-07, + "loss": 0.6357, + "step": 6493 + }, + { + "epoch": 12.025925925925925, + "grad_norm": 1.510951231354261, + "learning_rate": 1.6397963069988366e-07, + "loss": 0.6626, + "step": 6494 + }, + { + "epoch": 12.027777777777779, + "grad_norm": 1.5335809411608767, + "learning_rate": 1.639681062415388e-07, + "loss": 0.605, + "step": 6495 + }, + { + "epoch": 12.02962962962963, + "grad_norm": 1.7204886316557308, + "learning_rate": 1.6395658034501717e-07, + "loss": 0.6482, + "step": 6496 + }, + { + "epoch": 12.031481481481482, + "grad_norm": 2.009598895747422, + "learning_rate": 1.639450530105779e-07, + "loss": 0.6036, + "step": 6497 + }, + { + "epoch": 12.033333333333333, + "grad_norm": 1.9258096937047589, + "learning_rate": 1.6393352423848015e-07, + "loss": 0.6713, + "step": 6498 + }, + { + "epoch": 12.035185185185185, + "grad_norm": 1.3965586041185545, + "learning_rate": 1.6392199402898317e-07, + "loss": 0.6065, + "step": 6499 + }, + { + "epoch": 12.037037037037036, + "grad_norm": 1.5714223010425215, + "learning_rate": 1.6391046238234614e-07, + "loss": 0.5869, + "step": 6500 + }, + { + "epoch": 12.03888888888889, + "grad_norm": 1.5101361856185822, + "learning_rate": 1.6389892929882834e-07, + "loss": 0.6497, + "step": 6501 + }, + { + "epoch": 12.040740740740741, + "grad_norm": 1.6652365874458868, + "learning_rate": 1.6388739477868908e-07, + "loss": 0.6253, + "step": 6502 + }, + { + "epoch": 12.042592592592593, + "grad_norm": 1.658238720357663, + "learning_rate": 1.638758588221876e-07, + "loss": 0.5914, + "step": 6503 + }, + { + "epoch": 12.044444444444444, + "grad_norm": 1.4565484039910486, + "learning_rate": 1.638643214295834e-07, + "loss": 0.5648, + "step": 6504 + }, + { + "epoch": 12.046296296296296, + "grad_norm": 1.664461885354162, + "learning_rate": 1.638527826011358e-07, + "loss": 0.6587, + "step": 6505 + }, + { + "epoch": 12.048148148148147, + "grad_norm": 1.4336673346446493, + "learning_rate": 1.6384124233710418e-07, + "loss": 0.6345, + "step": 6506 + }, + { + "epoch": 12.05, + "grad_norm": 1.656852925452071, + "learning_rate": 1.6382970063774808e-07, + "loss": 0.6443, + "step": 6507 + }, + { + "epoch": 12.051851851851852, + "grad_norm": 2.148405851764428, + "learning_rate": 1.6381815750332697e-07, + "loss": 0.5999, + "step": 6508 + }, + { + "epoch": 12.053703703703704, + "grad_norm": 1.3938418144482092, + "learning_rate": 1.6380661293410029e-07, + "loss": 0.6023, + "step": 6509 + }, + { + "epoch": 12.055555555555555, + "grad_norm": 1.442808686166186, + "learning_rate": 1.6379506693032773e-07, + "loss": 0.6427, + "step": 6510 + }, + { + "epoch": 12.057407407407407, + "grad_norm": 1.6405884498720538, + "learning_rate": 1.6378351949226874e-07, + "loss": 0.5915, + "step": 6511 + }, + { + "epoch": 12.059259259259258, + "grad_norm": 1.433721313892015, + "learning_rate": 1.63771970620183e-07, + "loss": 0.5985, + "step": 6512 + }, + { + "epoch": 12.061111111111112, + "grad_norm": 1.6734910927793538, + "learning_rate": 1.6376042031433016e-07, + "loss": 0.6541, + "step": 6513 + }, + { + "epoch": 12.062962962962963, + "grad_norm": 1.4287522639243595, + "learning_rate": 1.6374886857496995e-07, + "loss": 0.6426, + "step": 6514 + }, + { + "epoch": 12.064814814814815, + "grad_norm": 1.6083777544648763, + "learning_rate": 1.6373731540236197e-07, + "loss": 0.6514, + "step": 6515 + }, + { + "epoch": 12.066666666666666, + "grad_norm": 1.7425769248276788, + "learning_rate": 1.6372576079676606e-07, + "loss": 0.5949, + "step": 6516 + }, + { + "epoch": 12.068518518518518, + "grad_norm": 1.387202619465093, + "learning_rate": 1.6371420475844201e-07, + "loss": 0.5958, + "step": 6517 + }, + { + "epoch": 12.07037037037037, + "grad_norm": 2.0225468943029377, + "learning_rate": 1.6370264728764954e-07, + "loss": 0.6291, + "step": 6518 + }, + { + "epoch": 12.072222222222223, + "grad_norm": 1.4538056017265102, + "learning_rate": 1.6369108838464855e-07, + "loss": 0.6247, + "step": 6519 + }, + { + "epoch": 12.074074074074074, + "grad_norm": 1.9796249104032133, + "learning_rate": 1.6367952804969895e-07, + "loss": 0.627, + "step": 6520 + }, + { + "epoch": 12.075925925925926, + "grad_norm": 1.5088544623933808, + "learning_rate": 1.6366796628306056e-07, + "loss": 0.6069, + "step": 6521 + }, + { + "epoch": 12.077777777777778, + "grad_norm": 1.866027702548339, + "learning_rate": 1.636564030849934e-07, + "loss": 0.6123, + "step": 6522 + }, + { + "epoch": 12.079629629629629, + "grad_norm": 1.5175240916048642, + "learning_rate": 1.6364483845575738e-07, + "loss": 0.622, + "step": 6523 + }, + { + "epoch": 12.081481481481482, + "grad_norm": 1.5534456220694706, + "learning_rate": 1.6363327239561256e-07, + "loss": 0.6226, + "step": 6524 + }, + { + "epoch": 12.083333333333334, + "grad_norm": 1.8432339561778077, + "learning_rate": 1.6362170490481896e-07, + "loss": 0.5655, + "step": 6525 + }, + { + "epoch": 12.085185185185185, + "grad_norm": 1.6688563884289775, + "learning_rate": 1.6361013598363663e-07, + "loss": 0.6584, + "step": 6526 + }, + { + "epoch": 12.087037037037037, + "grad_norm": 1.7570377878114263, + "learning_rate": 1.6359856563232565e-07, + "loss": 0.6659, + "step": 6527 + }, + { + "epoch": 12.088888888888889, + "grad_norm": 1.5569023024427324, + "learning_rate": 1.6358699385114621e-07, + "loss": 0.6177, + "step": 6528 + }, + { + "epoch": 12.09074074074074, + "grad_norm": 1.5847221148183925, + "learning_rate": 1.6357542064035847e-07, + "loss": 0.6196, + "step": 6529 + }, + { + "epoch": 12.092592592592593, + "grad_norm": 1.6345527100876884, + "learning_rate": 1.6356384600022258e-07, + "loss": 0.6065, + "step": 6530 + }, + { + "epoch": 12.094444444444445, + "grad_norm": 1.6465471033362065, + "learning_rate": 1.635522699309988e-07, + "loss": 0.6248, + "step": 6531 + }, + { + "epoch": 12.096296296296297, + "grad_norm": 1.9406330249831958, + "learning_rate": 1.6354069243294738e-07, + "loss": 0.5819, + "step": 6532 + }, + { + "epoch": 12.098148148148148, + "grad_norm": 1.5492302458312346, + "learning_rate": 1.635291135063286e-07, + "loss": 0.6655, + "step": 6533 + }, + { + "epoch": 12.1, + "grad_norm": 1.8354753626822733, + "learning_rate": 1.6351753315140285e-07, + "loss": 0.6201, + "step": 6534 + }, + { + "epoch": 12.101851851851851, + "grad_norm": 1.5549770914546528, + "learning_rate": 1.6350595136843044e-07, + "loss": 0.6137, + "step": 6535 + }, + { + "epoch": 12.103703703703705, + "grad_norm": 1.5637848081548742, + "learning_rate": 1.6349436815767172e-07, + "loss": 0.6362, + "step": 6536 + }, + { + "epoch": 12.105555555555556, + "grad_norm": 1.6528884910408053, + "learning_rate": 1.6348278351938717e-07, + "loss": 0.6406, + "step": 6537 + }, + { + "epoch": 12.107407407407408, + "grad_norm": 1.9152508319253156, + "learning_rate": 1.6347119745383728e-07, + "loss": 0.6042, + "step": 6538 + }, + { + "epoch": 12.10925925925926, + "grad_norm": 1.5408813579725722, + "learning_rate": 1.6345960996128244e-07, + "loss": 0.6323, + "step": 6539 + }, + { + "epoch": 12.11111111111111, + "grad_norm": 1.5710886017162735, + "learning_rate": 1.6344802104198323e-07, + "loss": 0.6013, + "step": 6540 + }, + { + "epoch": 12.112962962962962, + "grad_norm": 1.6607233817716947, + "learning_rate": 1.6343643069620016e-07, + "loss": 0.6376, + "step": 6541 + }, + { + "epoch": 12.114814814814816, + "grad_norm": 1.4268066080879431, + "learning_rate": 1.6342483892419385e-07, + "loss": 0.6167, + "step": 6542 + }, + { + "epoch": 12.116666666666667, + "grad_norm": 2.06456398263222, + "learning_rate": 1.6341324572622495e-07, + "loss": 0.6271, + "step": 6543 + }, + { + "epoch": 12.118518518518519, + "grad_norm": 1.405837945099825, + "learning_rate": 1.6340165110255399e-07, + "loss": 0.6549, + "step": 6544 + }, + { + "epoch": 12.12037037037037, + "grad_norm": 1.492356949341976, + "learning_rate": 1.6339005505344174e-07, + "loss": 0.5785, + "step": 6545 + }, + { + "epoch": 12.122222222222222, + "grad_norm": 1.4587434473383574, + "learning_rate": 1.633784575791489e-07, + "loss": 0.6479, + "step": 6546 + }, + { + "epoch": 12.124074074074073, + "grad_norm": 2.016253170059147, + "learning_rate": 1.6336685867993621e-07, + "loss": 0.6284, + "step": 6547 + }, + { + "epoch": 12.125925925925927, + "grad_norm": 1.5193277045240394, + "learning_rate": 1.633552583560644e-07, + "loss": 0.6331, + "step": 6548 + }, + { + "epoch": 12.127777777777778, + "grad_norm": 2.521301034698399, + "learning_rate": 1.6334365660779432e-07, + "loss": 0.5935, + "step": 6549 + }, + { + "epoch": 12.12962962962963, + "grad_norm": 1.6305399538375949, + "learning_rate": 1.633320534353868e-07, + "loss": 0.6689, + "step": 6550 + }, + { + "epoch": 12.131481481481481, + "grad_norm": 2.3557694700756, + "learning_rate": 1.6332044883910273e-07, + "loss": 0.5922, + "step": 6551 + }, + { + "epoch": 12.133333333333333, + "grad_norm": 1.4017311422558687, + "learning_rate": 1.63308842819203e-07, + "loss": 0.649, + "step": 6552 + }, + { + "epoch": 12.135185185185184, + "grad_norm": 1.6734928641771962, + "learning_rate": 1.6329723537594852e-07, + "loss": 0.5723, + "step": 6553 + }, + { + "epoch": 12.137037037037038, + "grad_norm": 1.627155463241186, + "learning_rate": 1.6328562650960026e-07, + "loss": 0.6265, + "step": 6554 + }, + { + "epoch": 12.13888888888889, + "grad_norm": 1.5730465046871955, + "learning_rate": 1.6327401622041927e-07, + "loss": 0.6173, + "step": 6555 + }, + { + "epoch": 12.14074074074074, + "grad_norm": 1.563576987788256, + "learning_rate": 1.6326240450866653e-07, + "loss": 0.6453, + "step": 6556 + }, + { + "epoch": 12.142592592592592, + "grad_norm": 1.8314768806280253, + "learning_rate": 1.6325079137460312e-07, + "loss": 0.5755, + "step": 6557 + }, + { + "epoch": 12.144444444444444, + "grad_norm": 1.6788595507566144, + "learning_rate": 1.6323917681849013e-07, + "loss": 0.6029, + "step": 6558 + }, + { + "epoch": 12.146296296296295, + "grad_norm": 1.6988634075470816, + "learning_rate": 1.632275608405887e-07, + "loss": 0.6497, + "step": 6559 + }, + { + "epoch": 12.148148148148149, + "grad_norm": 1.6743549664174686, + "learning_rate": 1.6321594344115996e-07, + "loss": 0.6747, + "step": 6560 + }, + { + "epoch": 12.15, + "grad_norm": 1.61027545479233, + "learning_rate": 1.6320432462046513e-07, + "loss": 0.6127, + "step": 6561 + }, + { + "epoch": 12.151851851851852, + "grad_norm": 1.5173173449115376, + "learning_rate": 1.6319270437876546e-07, + "loss": 0.6166, + "step": 6562 + }, + { + "epoch": 12.153703703703703, + "grad_norm": 1.96836329381583, + "learning_rate": 1.631810827163221e-07, + "loss": 0.6854, + "step": 6563 + }, + { + "epoch": 12.155555555555555, + "grad_norm": 1.9295510891491456, + "learning_rate": 1.6316945963339644e-07, + "loss": 0.5745, + "step": 6564 + }, + { + "epoch": 12.157407407407407, + "grad_norm": 1.5972623403537682, + "learning_rate": 1.6315783513024973e-07, + "loss": 0.6476, + "step": 6565 + }, + { + "epoch": 12.15925925925926, + "grad_norm": 1.9266437878385532, + "learning_rate": 1.631462092071434e-07, + "loss": 0.655, + "step": 6566 + }, + { + "epoch": 12.161111111111111, + "grad_norm": 1.3488496020973324, + "learning_rate": 1.6313458186433878e-07, + "loss": 0.6698, + "step": 6567 + }, + { + "epoch": 12.162962962962963, + "grad_norm": 2.0155003994474363, + "learning_rate": 1.6312295310209726e-07, + "loss": 0.6157, + "step": 6568 + }, + { + "epoch": 12.164814814814815, + "grad_norm": 1.9602772561395967, + "learning_rate": 1.6311132292068033e-07, + "loss": 0.6049, + "step": 6569 + }, + { + "epoch": 12.166666666666666, + "grad_norm": 1.779101121190205, + "learning_rate": 1.6309969132034944e-07, + "loss": 0.57, + "step": 6570 + }, + { + "epoch": 12.168518518518518, + "grad_norm": 1.7474196550150642, + "learning_rate": 1.6308805830136613e-07, + "loss": 0.6368, + "step": 6571 + }, + { + "epoch": 12.170370370370371, + "grad_norm": 1.5636307263675404, + "learning_rate": 1.630764238639919e-07, + "loss": 0.5961, + "step": 6572 + }, + { + "epoch": 12.172222222222222, + "grad_norm": 1.3911596034294955, + "learning_rate": 1.6306478800848837e-07, + "loss": 0.6197, + "step": 6573 + }, + { + "epoch": 12.174074074074074, + "grad_norm": 1.5910468442516266, + "learning_rate": 1.6305315073511712e-07, + "loss": 0.622, + "step": 6574 + }, + { + "epoch": 12.175925925925926, + "grad_norm": 1.6640548708685794, + "learning_rate": 1.6304151204413978e-07, + "loss": 0.5823, + "step": 6575 + }, + { + "epoch": 12.177777777777777, + "grad_norm": 1.8600778641018916, + "learning_rate": 1.6302987193581805e-07, + "loss": 0.6503, + "step": 6576 + }, + { + "epoch": 12.17962962962963, + "grad_norm": 3.238108196975405, + "learning_rate": 1.6301823041041361e-07, + "loss": 0.6503, + "step": 6577 + }, + { + "epoch": 12.181481481481482, + "grad_norm": 1.5606805385472824, + "learning_rate": 1.6300658746818818e-07, + "loss": 0.656, + "step": 6578 + }, + { + "epoch": 12.183333333333334, + "grad_norm": 1.538704399617346, + "learning_rate": 1.6299494310940356e-07, + "loss": 0.6139, + "step": 6579 + }, + { + "epoch": 12.185185185185185, + "grad_norm": 1.4165390889858942, + "learning_rate": 1.629832973343215e-07, + "loss": 0.5996, + "step": 6580 + }, + { + "epoch": 12.187037037037037, + "grad_norm": 1.5930324318358666, + "learning_rate": 1.6297165014320388e-07, + "loss": 0.6355, + "step": 6581 + }, + { + "epoch": 12.188888888888888, + "grad_norm": 1.9998978511237038, + "learning_rate": 1.6296000153631252e-07, + "loss": 0.6438, + "step": 6582 + }, + { + "epoch": 12.190740740740742, + "grad_norm": 1.5156997176815612, + "learning_rate": 1.6294835151390936e-07, + "loss": 0.6188, + "step": 6583 + }, + { + "epoch": 12.192592592592593, + "grad_norm": 1.4144826874759324, + "learning_rate": 1.6293670007625625e-07, + "loss": 0.5847, + "step": 6584 + }, + { + "epoch": 12.194444444444445, + "grad_norm": 1.6889681940427033, + "learning_rate": 1.6292504722361524e-07, + "loss": 0.64, + "step": 6585 + }, + { + "epoch": 12.196296296296296, + "grad_norm": 1.5517575737154559, + "learning_rate": 1.6291339295624825e-07, + "loss": 0.62, + "step": 6586 + }, + { + "epoch": 12.198148148148148, + "grad_norm": 1.8582824280712988, + "learning_rate": 1.6290173727441727e-07, + "loss": 0.6011, + "step": 6587 + }, + { + "epoch": 12.2, + "grad_norm": 1.56630565582127, + "learning_rate": 1.6289008017838445e-07, + "loss": 0.5976, + "step": 6588 + }, + { + "epoch": 12.201851851851853, + "grad_norm": 1.500845821932329, + "learning_rate": 1.628784216684118e-07, + "loss": 0.6252, + "step": 6589 + }, + { + "epoch": 12.203703703703704, + "grad_norm": 1.5890627908192463, + "learning_rate": 1.6286676174476143e-07, + "loss": 0.6445, + "step": 6590 + }, + { + "epoch": 12.205555555555556, + "grad_norm": 1.5182906680373982, + "learning_rate": 1.6285510040769557e-07, + "loss": 0.6042, + "step": 6591 + }, + { + "epoch": 12.207407407407407, + "grad_norm": 1.7853847086911154, + "learning_rate": 1.628434376574763e-07, + "loss": 0.6038, + "step": 6592 + }, + { + "epoch": 12.209259259259259, + "grad_norm": 1.6615417061196436, + "learning_rate": 1.6283177349436587e-07, + "loss": 0.6743, + "step": 6593 + }, + { + "epoch": 12.21111111111111, + "grad_norm": 1.5307108825072282, + "learning_rate": 1.6282010791862654e-07, + "loss": 0.618, + "step": 6594 + }, + { + "epoch": 12.212962962962964, + "grad_norm": 1.6159445296531232, + "learning_rate": 1.6280844093052058e-07, + "loss": 0.5802, + "step": 6595 + }, + { + "epoch": 12.214814814814815, + "grad_norm": 1.5872601922843572, + "learning_rate": 1.6279677253031025e-07, + "loss": 0.6365, + "step": 6596 + }, + { + "epoch": 12.216666666666667, + "grad_norm": 1.7130597138637507, + "learning_rate": 1.6278510271825793e-07, + "loss": 0.6163, + "step": 6597 + }, + { + "epoch": 12.218518518518518, + "grad_norm": 1.5024369543590266, + "learning_rate": 1.62773431494626e-07, + "loss": 0.5931, + "step": 6598 + }, + { + "epoch": 12.22037037037037, + "grad_norm": 1.4031452228124768, + "learning_rate": 1.6276175885967683e-07, + "loss": 0.6454, + "step": 6599 + }, + { + "epoch": 12.222222222222221, + "grad_norm": 1.6618976941807972, + "learning_rate": 1.6275008481367286e-07, + "loss": 0.6908, + "step": 6600 + }, + { + "epoch": 12.224074074074075, + "grad_norm": 1.6554977564411542, + "learning_rate": 1.6273840935687654e-07, + "loss": 0.6406, + "step": 6601 + }, + { + "epoch": 12.225925925925926, + "grad_norm": 1.614720235985635, + "learning_rate": 1.627267324895504e-07, + "loss": 0.5723, + "step": 6602 + }, + { + "epoch": 12.227777777777778, + "grad_norm": 1.6783746757530102, + "learning_rate": 1.6271505421195694e-07, + "loss": 0.5999, + "step": 6603 + }, + { + "epoch": 12.22962962962963, + "grad_norm": 1.5762241754213928, + "learning_rate": 1.6270337452435874e-07, + "loss": 0.6318, + "step": 6604 + }, + { + "epoch": 12.231481481481481, + "grad_norm": 1.5116313846961869, + "learning_rate": 1.6269169342701838e-07, + "loss": 0.6202, + "step": 6605 + }, + { + "epoch": 12.233333333333333, + "grad_norm": 1.699175362466538, + "learning_rate": 1.626800109201985e-07, + "loss": 0.6381, + "step": 6606 + }, + { + "epoch": 12.235185185185186, + "grad_norm": 1.6300570154576297, + "learning_rate": 1.626683270041617e-07, + "loss": 0.6303, + "step": 6607 + }, + { + "epoch": 12.237037037037037, + "grad_norm": 1.6055710022349048, + "learning_rate": 1.6265664167917073e-07, + "loss": 0.6651, + "step": 6608 + }, + { + "epoch": 12.238888888888889, + "grad_norm": 1.8727886695448739, + "learning_rate": 1.6264495494548828e-07, + "loss": 0.5932, + "step": 6609 + }, + { + "epoch": 12.24074074074074, + "grad_norm": 1.4214025258412486, + "learning_rate": 1.626332668033771e-07, + "loss": 0.6026, + "step": 6610 + }, + { + "epoch": 12.242592592592592, + "grad_norm": 1.3390510790588248, + "learning_rate": 1.6262157725309997e-07, + "loss": 0.5731, + "step": 6611 + }, + { + "epoch": 12.244444444444444, + "grad_norm": 1.7581170186667185, + "learning_rate": 1.626098862949197e-07, + "loss": 0.6572, + "step": 6612 + }, + { + "epoch": 12.246296296296297, + "grad_norm": 1.6051541908675027, + "learning_rate": 1.6259819392909913e-07, + "loss": 0.6331, + "step": 6613 + }, + { + "epoch": 12.248148148148148, + "grad_norm": 1.4941806113201141, + "learning_rate": 1.6258650015590116e-07, + "loss": 0.5985, + "step": 6614 + }, + { + "epoch": 12.25, + "grad_norm": 1.435015237958376, + "learning_rate": 1.6257480497558871e-07, + "loss": 0.5999, + "step": 6615 + }, + { + "epoch": 12.251851851851852, + "grad_norm": 1.9553315189194644, + "learning_rate": 1.6256310838842463e-07, + "loss": 0.6355, + "step": 6616 + }, + { + "epoch": 12.253703703703703, + "grad_norm": 1.6553117790453324, + "learning_rate": 1.6255141039467203e-07, + "loss": 0.6082, + "step": 6617 + }, + { + "epoch": 12.255555555555556, + "grad_norm": 1.5272240377855684, + "learning_rate": 1.625397109945938e-07, + "loss": 0.6261, + "step": 6618 + }, + { + "epoch": 12.257407407407408, + "grad_norm": 1.4704925857518294, + "learning_rate": 1.6252801018845302e-07, + "loss": 0.6327, + "step": 6619 + }, + { + "epoch": 12.25925925925926, + "grad_norm": 1.6127325226807936, + "learning_rate": 1.6251630797651273e-07, + "loss": 0.5891, + "step": 6620 + }, + { + "epoch": 12.261111111111111, + "grad_norm": 1.551038443458036, + "learning_rate": 1.6250460435903607e-07, + "loss": 0.6259, + "step": 6621 + }, + { + "epoch": 12.262962962962963, + "grad_norm": 2.3189172429940585, + "learning_rate": 1.624928993362861e-07, + "loss": 0.6409, + "step": 6622 + }, + { + "epoch": 12.264814814814814, + "grad_norm": 1.3565462691550279, + "learning_rate": 1.6248119290852608e-07, + "loss": 0.6293, + "step": 6623 + }, + { + "epoch": 12.266666666666667, + "grad_norm": 1.9321640098027901, + "learning_rate": 1.6246948507601913e-07, + "loss": 0.6464, + "step": 6624 + }, + { + "epoch": 12.268518518518519, + "grad_norm": 1.8439408605762335, + "learning_rate": 1.6245777583902848e-07, + "loss": 0.613, + "step": 6625 + }, + { + "epoch": 12.27037037037037, + "grad_norm": 1.520599900625449, + "learning_rate": 1.624460651978174e-07, + "loss": 0.6197, + "step": 6626 + }, + { + "epoch": 12.272222222222222, + "grad_norm": 1.5565971027623497, + "learning_rate": 1.6243435315264917e-07, + "loss": 0.63, + "step": 6627 + }, + { + "epoch": 12.274074074074074, + "grad_norm": 1.5292274508510055, + "learning_rate": 1.6242263970378711e-07, + "loss": 0.6174, + "step": 6628 + }, + { + "epoch": 12.275925925925925, + "grad_norm": 1.5064823121259525, + "learning_rate": 1.6241092485149456e-07, + "loss": 0.6484, + "step": 6629 + }, + { + "epoch": 12.277777777777779, + "grad_norm": 1.5585272861276653, + "learning_rate": 1.6239920859603494e-07, + "loss": 0.6323, + "step": 6630 + }, + { + "epoch": 12.27962962962963, + "grad_norm": 1.4883952896329649, + "learning_rate": 1.6238749093767163e-07, + "loss": 0.6782, + "step": 6631 + }, + { + "epoch": 12.281481481481482, + "grad_norm": 1.5806561979946274, + "learning_rate": 1.6237577187666808e-07, + "loss": 0.6412, + "step": 6632 + }, + { + "epoch": 12.283333333333333, + "grad_norm": 2.7180298210675953, + "learning_rate": 1.6236405141328777e-07, + "loss": 0.6242, + "step": 6633 + }, + { + "epoch": 12.285185185185185, + "grad_norm": 1.5444227813345974, + "learning_rate": 1.623523295477942e-07, + "loss": 0.5554, + "step": 6634 + }, + { + "epoch": 12.287037037037036, + "grad_norm": 1.4465858218865122, + "learning_rate": 1.6234060628045093e-07, + "loss": 0.621, + "step": 6635 + }, + { + "epoch": 12.28888888888889, + "grad_norm": 1.425017455330324, + "learning_rate": 1.623288816115215e-07, + "loss": 0.6189, + "step": 6636 + }, + { + "epoch": 12.290740740740741, + "grad_norm": 1.512138796582942, + "learning_rate": 1.6231715554126955e-07, + "loss": 0.6125, + "step": 6637 + }, + { + "epoch": 12.292592592592593, + "grad_norm": 1.4588839274034164, + "learning_rate": 1.623054280699587e-07, + "loss": 0.6586, + "step": 6638 + }, + { + "epoch": 12.294444444444444, + "grad_norm": 1.5547541341306146, + "learning_rate": 1.6229369919785255e-07, + "loss": 0.5733, + "step": 6639 + }, + { + "epoch": 12.296296296296296, + "grad_norm": 1.5848618649918327, + "learning_rate": 1.622819689252149e-07, + "loss": 0.6168, + "step": 6640 + }, + { + "epoch": 12.298148148148147, + "grad_norm": 1.529607003296878, + "learning_rate": 1.622702372523094e-07, + "loss": 0.6325, + "step": 6641 + }, + { + "epoch": 12.3, + "grad_norm": 1.5430004695018045, + "learning_rate": 1.6225850417939988e-07, + "loss": 0.6033, + "step": 6642 + }, + { + "epoch": 12.301851851851852, + "grad_norm": 1.4390134576694458, + "learning_rate": 1.622467697067501e-07, + "loss": 0.6173, + "step": 6643 + }, + { + "epoch": 12.303703703703704, + "grad_norm": 1.9117665942689086, + "learning_rate": 1.6223503383462385e-07, + "loss": 0.6169, + "step": 6644 + }, + { + "epoch": 12.305555555555555, + "grad_norm": 1.5362139899252765, + "learning_rate": 1.62223296563285e-07, + "loss": 0.616, + "step": 6645 + }, + { + "epoch": 12.307407407407407, + "grad_norm": 1.5332623328060124, + "learning_rate": 1.6221155789299749e-07, + "loss": 0.631, + "step": 6646 + }, + { + "epoch": 12.309259259259258, + "grad_norm": 1.6702510202179586, + "learning_rate": 1.6219981782402516e-07, + "loss": 0.6605, + "step": 6647 + }, + { + "epoch": 12.311111111111112, + "grad_norm": 1.7122450359221875, + "learning_rate": 1.62188076356632e-07, + "loss": 0.6384, + "step": 6648 + }, + { + "epoch": 12.312962962962963, + "grad_norm": 1.8828853178244909, + "learning_rate": 1.62176333491082e-07, + "loss": 0.568, + "step": 6649 + }, + { + "epoch": 12.314814814814815, + "grad_norm": 1.6105207777805626, + "learning_rate": 1.6216458922763914e-07, + "loss": 0.6364, + "step": 6650 + }, + { + "epoch": 12.316666666666666, + "grad_norm": 1.5398965599915089, + "learning_rate": 1.621528435665675e-07, + "loss": 0.6296, + "step": 6651 + }, + { + "epoch": 12.318518518518518, + "grad_norm": 1.679341353015418, + "learning_rate": 1.6214109650813107e-07, + "loss": 0.5752, + "step": 6652 + }, + { + "epoch": 12.32037037037037, + "grad_norm": 1.6629218897533944, + "learning_rate": 1.6212934805259408e-07, + "loss": 0.6229, + "step": 6653 + }, + { + "epoch": 12.322222222222223, + "grad_norm": 1.5660911809385762, + "learning_rate": 1.6211759820022058e-07, + "loss": 0.6368, + "step": 6654 + }, + { + "epoch": 12.324074074074074, + "grad_norm": 1.4445883548145186, + "learning_rate": 1.6210584695127478e-07, + "loss": 0.6101, + "step": 6655 + }, + { + "epoch": 12.325925925925926, + "grad_norm": 1.4622276754859247, + "learning_rate": 1.6209409430602085e-07, + "loss": 0.6072, + "step": 6656 + }, + { + "epoch": 12.327777777777778, + "grad_norm": 1.8274624429311774, + "learning_rate": 1.6208234026472305e-07, + "loss": 0.6663, + "step": 6657 + }, + { + "epoch": 12.329629629629629, + "grad_norm": 1.596237889931007, + "learning_rate": 1.620705848276456e-07, + "loss": 0.6235, + "step": 6658 + }, + { + "epoch": 12.331481481481482, + "grad_norm": 1.4381326986273786, + "learning_rate": 1.6205882799505284e-07, + "loss": 0.6533, + "step": 6659 + }, + { + "epoch": 12.333333333333334, + "grad_norm": 1.6141048868813446, + "learning_rate": 1.6204706976720907e-07, + "loss": 0.6056, + "step": 6660 + }, + { + "epoch": 12.335185185185185, + "grad_norm": 1.5752245919685965, + "learning_rate": 1.6203531014437867e-07, + "loss": 0.5937, + "step": 6661 + }, + { + "epoch": 12.337037037037037, + "grad_norm": 2.776928552961708, + "learning_rate": 1.6202354912682598e-07, + "loss": 0.5596, + "step": 6662 + }, + { + "epoch": 12.338888888888889, + "grad_norm": 1.4192099048175186, + "learning_rate": 1.620117867148155e-07, + "loss": 0.6136, + "step": 6663 + }, + { + "epoch": 12.34074074074074, + "grad_norm": 1.372271166929822, + "learning_rate": 1.6200002290861158e-07, + "loss": 0.5909, + "step": 6664 + }, + { + "epoch": 12.342592592592593, + "grad_norm": 4.803161356529421, + "learning_rate": 1.619882577084788e-07, + "loss": 0.645, + "step": 6665 + }, + { + "epoch": 12.344444444444445, + "grad_norm": 1.6570008096125624, + "learning_rate": 1.6197649111468162e-07, + "loss": 0.6471, + "step": 6666 + }, + { + "epoch": 12.346296296296297, + "grad_norm": 1.7599154016337184, + "learning_rate": 1.6196472312748458e-07, + "loss": 0.5933, + "step": 6667 + }, + { + "epoch": 12.348148148148148, + "grad_norm": 1.526100165994934, + "learning_rate": 1.619529537471523e-07, + "loss": 0.6111, + "step": 6668 + }, + { + "epoch": 12.35, + "grad_norm": 1.5675438673126456, + "learning_rate": 1.6194118297394935e-07, + "loss": 0.5964, + "step": 6669 + }, + { + "epoch": 12.351851851851851, + "grad_norm": 1.4495462583452638, + "learning_rate": 1.6192941080814034e-07, + "loss": 0.6223, + "step": 6670 + }, + { + "epoch": 12.353703703703705, + "grad_norm": 2.1730774512058155, + "learning_rate": 1.6191763724998998e-07, + "loss": 0.6195, + "step": 6671 + }, + { + "epoch": 12.355555555555556, + "grad_norm": 3.3725143006871763, + "learning_rate": 1.6190586229976302e-07, + "loss": 0.6391, + "step": 6672 + }, + { + "epoch": 12.357407407407408, + "grad_norm": 1.7214815507989247, + "learning_rate": 1.6189408595772406e-07, + "loss": 0.5992, + "step": 6673 + }, + { + "epoch": 12.35925925925926, + "grad_norm": 1.5272373084204767, + "learning_rate": 1.6188230822413802e-07, + "loss": 0.6355, + "step": 6674 + }, + { + "epoch": 12.36111111111111, + "grad_norm": 1.627865729553433, + "learning_rate": 1.618705290992696e-07, + "loss": 0.6049, + "step": 6675 + }, + { + "epoch": 12.362962962962962, + "grad_norm": 1.9066500323363462, + "learning_rate": 1.6185874858338365e-07, + "loss": 0.6042, + "step": 6676 + }, + { + "epoch": 12.364814814814816, + "grad_norm": 2.04115358499272, + "learning_rate": 1.6184696667674504e-07, + "loss": 0.6366, + "step": 6677 + }, + { + "epoch": 12.366666666666667, + "grad_norm": 3.2010776354981854, + "learning_rate": 1.6183518337961862e-07, + "loss": 0.6333, + "step": 6678 + }, + { + "epoch": 12.368518518518519, + "grad_norm": 1.6365931904066346, + "learning_rate": 1.6182339869226933e-07, + "loss": 0.5968, + "step": 6679 + }, + { + "epoch": 12.37037037037037, + "grad_norm": 3.021436372165264, + "learning_rate": 1.6181161261496213e-07, + "loss": 0.621, + "step": 6680 + }, + { + "epoch": 12.372222222222222, + "grad_norm": 1.472734920206644, + "learning_rate": 1.6179982514796202e-07, + "loss": 0.6569, + "step": 6681 + }, + { + "epoch": 12.374074074074073, + "grad_norm": 1.5195528675956615, + "learning_rate": 1.61788036291534e-07, + "loss": 0.6035, + "step": 6682 + }, + { + "epoch": 12.375925925925927, + "grad_norm": 2.1156055599206214, + "learning_rate": 1.617762460459431e-07, + "loss": 0.6221, + "step": 6683 + }, + { + "epoch": 12.377777777777778, + "grad_norm": 2.0329831742732267, + "learning_rate": 1.6176445441145438e-07, + "loss": 0.6308, + "step": 6684 + }, + { + "epoch": 12.37962962962963, + "grad_norm": 1.6021361103864833, + "learning_rate": 1.6175266138833302e-07, + "loss": 0.6194, + "step": 6685 + }, + { + "epoch": 12.381481481481481, + "grad_norm": 1.5650015936914683, + "learning_rate": 1.617408669768441e-07, + "loss": 0.6369, + "step": 6686 + }, + { + "epoch": 12.383333333333333, + "grad_norm": 1.6215876413509036, + "learning_rate": 1.6172907117725279e-07, + "loss": 0.6228, + "step": 6687 + }, + { + "epoch": 12.385185185185184, + "grad_norm": 1.5098351993197505, + "learning_rate": 1.6171727398982428e-07, + "loss": 0.5764, + "step": 6688 + }, + { + "epoch": 12.387037037037038, + "grad_norm": 1.5395020853703976, + "learning_rate": 1.6170547541482388e-07, + "loss": 0.628, + "step": 6689 + }, + { + "epoch": 12.38888888888889, + "grad_norm": 1.5620088815710957, + "learning_rate": 1.616936754525168e-07, + "loss": 0.5611, + "step": 6690 + }, + { + "epoch": 12.39074074074074, + "grad_norm": 1.7235453802719307, + "learning_rate": 1.616818741031683e-07, + "loss": 0.6266, + "step": 6691 + }, + { + "epoch": 12.392592592592592, + "grad_norm": 1.3915572789105382, + "learning_rate": 1.616700713670438e-07, + "loss": 0.6384, + "step": 6692 + }, + { + "epoch": 12.394444444444444, + "grad_norm": 1.7370381306710112, + "learning_rate": 1.6165826724440854e-07, + "loss": 0.6192, + "step": 6693 + }, + { + "epoch": 12.396296296296295, + "grad_norm": 1.6081971050431976, + "learning_rate": 1.6164646173552802e-07, + "loss": 0.6453, + "step": 6694 + }, + { + "epoch": 12.398148148148149, + "grad_norm": 1.7344468339917467, + "learning_rate": 1.616346548406676e-07, + "loss": 0.6368, + "step": 6695 + }, + { + "epoch": 12.4, + "grad_norm": 1.4422640100939879, + "learning_rate": 1.6162284656009274e-07, + "loss": 0.6057, + "step": 6696 + }, + { + "epoch": 12.401851851851852, + "grad_norm": 1.621497058307677, + "learning_rate": 1.616110368940689e-07, + "loss": 0.5783, + "step": 6697 + }, + { + "epoch": 12.403703703703703, + "grad_norm": 1.546223165874739, + "learning_rate": 1.6159922584286164e-07, + "loss": 0.671, + "step": 6698 + }, + { + "epoch": 12.405555555555555, + "grad_norm": 1.5474537248013673, + "learning_rate": 1.615874134067365e-07, + "loss": 0.6123, + "step": 6699 + }, + { + "epoch": 12.407407407407407, + "grad_norm": 1.4787270387375124, + "learning_rate": 1.61575599585959e-07, + "loss": 0.6314, + "step": 6700 + }, + { + "epoch": 12.40925925925926, + "grad_norm": 1.7273091249486154, + "learning_rate": 1.615637843807948e-07, + "loss": 0.6376, + "step": 6701 + }, + { + "epoch": 12.411111111111111, + "grad_norm": 1.8812011119113892, + "learning_rate": 1.6155196779150954e-07, + "loss": 0.5996, + "step": 6702 + }, + { + "epoch": 12.412962962962963, + "grad_norm": 1.619848258974585, + "learning_rate": 1.6154014981836884e-07, + "loss": 0.6333, + "step": 6703 + }, + { + "epoch": 12.414814814814815, + "grad_norm": 1.3868376269066072, + "learning_rate": 1.6152833046163843e-07, + "loss": 0.6217, + "step": 6704 + }, + { + "epoch": 12.416666666666666, + "grad_norm": 1.7234605081825254, + "learning_rate": 1.6151650972158408e-07, + "loss": 0.6541, + "step": 6705 + }, + { + "epoch": 12.418518518518518, + "grad_norm": 1.732866130801652, + "learning_rate": 1.615046875984715e-07, + "loss": 0.6189, + "step": 6706 + }, + { + "epoch": 12.420370370370371, + "grad_norm": 1.39916696619329, + "learning_rate": 1.6149286409256647e-07, + "loss": 0.6182, + "step": 6707 + }, + { + "epoch": 12.422222222222222, + "grad_norm": 1.6613702358771392, + "learning_rate": 1.6148103920413489e-07, + "loss": 0.6727, + "step": 6708 + }, + { + "epoch": 12.424074074074074, + "grad_norm": 1.821863792053127, + "learning_rate": 1.6146921293344254e-07, + "loss": 0.5735, + "step": 6709 + }, + { + "epoch": 12.425925925925926, + "grad_norm": 1.3577622022312832, + "learning_rate": 1.6145738528075535e-07, + "loss": 0.6106, + "step": 6710 + }, + { + "epoch": 12.427777777777777, + "grad_norm": 1.7602412179167821, + "learning_rate": 1.6144555624633918e-07, + "loss": 0.6459, + "step": 6711 + }, + { + "epoch": 12.42962962962963, + "grad_norm": 1.5101847967225224, + "learning_rate": 1.6143372583046006e-07, + "loss": 0.6251, + "step": 6712 + }, + { + "epoch": 12.431481481481482, + "grad_norm": 1.695320221636631, + "learning_rate": 1.6142189403338394e-07, + "loss": 0.6341, + "step": 6713 + }, + { + "epoch": 12.433333333333334, + "grad_norm": 1.6764135747248183, + "learning_rate": 1.614100608553768e-07, + "loss": 0.6065, + "step": 6714 + }, + { + "epoch": 12.435185185185185, + "grad_norm": 1.5180336853503937, + "learning_rate": 1.613982262967047e-07, + "loss": 0.6297, + "step": 6715 + }, + { + "epoch": 12.437037037037037, + "grad_norm": 1.4451097703190439, + "learning_rate": 1.6138639035763375e-07, + "loss": 0.5801, + "step": 6716 + }, + { + "epoch": 12.438888888888888, + "grad_norm": 1.501209780115876, + "learning_rate": 1.6137455303843e-07, + "loss": 0.672, + "step": 6717 + }, + { + "epoch": 12.440740740740742, + "grad_norm": 1.551995743307302, + "learning_rate": 1.6136271433935963e-07, + "loss": 0.638, + "step": 6718 + }, + { + "epoch": 12.442592592592593, + "grad_norm": 1.444861987967557, + "learning_rate": 1.6135087426068876e-07, + "loss": 0.6459, + "step": 6719 + }, + { + "epoch": 12.444444444444445, + "grad_norm": 1.7419807203155202, + "learning_rate": 1.6133903280268362e-07, + "loss": 0.6045, + "step": 6720 + }, + { + "epoch": 12.446296296296296, + "grad_norm": 2.0614981942526236, + "learning_rate": 1.6132718996561044e-07, + "loss": 0.6103, + "step": 6721 + }, + { + "epoch": 12.448148148148148, + "grad_norm": 1.8342747789336824, + "learning_rate": 1.6131534574973543e-07, + "loss": 0.6207, + "step": 6722 + }, + { + "epoch": 12.45, + "grad_norm": 1.6919490628090412, + "learning_rate": 1.6130350015532494e-07, + "loss": 0.6331, + "step": 6723 + }, + { + "epoch": 12.451851851851853, + "grad_norm": 1.7157488714688314, + "learning_rate": 1.612916531826453e-07, + "loss": 0.5967, + "step": 6724 + }, + { + "epoch": 12.453703703703704, + "grad_norm": 1.6300345835805943, + "learning_rate": 1.6127980483196278e-07, + "loss": 0.6013, + "step": 6725 + }, + { + "epoch": 12.455555555555556, + "grad_norm": 1.4957333485476974, + "learning_rate": 1.612679551035438e-07, + "loss": 0.6032, + "step": 6726 + }, + { + "epoch": 12.457407407407407, + "grad_norm": 1.843506681342497, + "learning_rate": 1.6125610399765486e-07, + "loss": 0.6412, + "step": 6727 + }, + { + "epoch": 12.459259259259259, + "grad_norm": 1.6542256609222146, + "learning_rate": 1.6124425151456227e-07, + "loss": 0.6732, + "step": 6728 + }, + { + "epoch": 12.46111111111111, + "grad_norm": 1.438272911441366, + "learning_rate": 1.6123239765453257e-07, + "loss": 0.5925, + "step": 6729 + }, + { + "epoch": 12.462962962962964, + "grad_norm": 1.5756248187825583, + "learning_rate": 1.612205424178323e-07, + "loss": 0.6276, + "step": 6730 + }, + { + "epoch": 12.464814814814815, + "grad_norm": 1.630278639994648, + "learning_rate": 1.6120868580472796e-07, + "loss": 0.5784, + "step": 6731 + }, + { + "epoch": 12.466666666666667, + "grad_norm": 1.527909196796277, + "learning_rate": 1.6119682781548612e-07, + "loss": 0.5841, + "step": 6732 + }, + { + "epoch": 12.468518518518518, + "grad_norm": 1.4289853368852803, + "learning_rate": 1.611849684503734e-07, + "loss": 0.6061, + "step": 6733 + }, + { + "epoch": 12.47037037037037, + "grad_norm": 1.708498419700212, + "learning_rate": 1.6117310770965638e-07, + "loss": 0.6562, + "step": 6734 + }, + { + "epoch": 12.472222222222221, + "grad_norm": 1.8476025550069186, + "learning_rate": 1.6116124559360176e-07, + "loss": 0.6044, + "step": 6735 + }, + { + "epoch": 12.474074074074075, + "grad_norm": 1.577579629598864, + "learning_rate": 1.611493821024762e-07, + "loss": 0.6457, + "step": 6736 + }, + { + "epoch": 12.475925925925926, + "grad_norm": 1.8771734073270754, + "learning_rate": 1.6113751723654652e-07, + "loss": 0.6331, + "step": 6737 + }, + { + "epoch": 12.477777777777778, + "grad_norm": 1.3985466495227243, + "learning_rate": 1.6112565099607936e-07, + "loss": 0.6236, + "step": 6738 + }, + { + "epoch": 12.47962962962963, + "grad_norm": 1.5931850879221627, + "learning_rate": 1.611137833813416e-07, + "loss": 0.6501, + "step": 6739 + }, + { + "epoch": 12.481481481481481, + "grad_norm": 1.4911377057410506, + "learning_rate": 1.6110191439259994e-07, + "loss": 0.623, + "step": 6740 + }, + { + "epoch": 12.483333333333333, + "grad_norm": 1.493446596010548, + "learning_rate": 1.610900440301213e-07, + "loss": 0.6062, + "step": 6741 + }, + { + "epoch": 12.485185185185186, + "grad_norm": 1.8745480233916807, + "learning_rate": 1.610781722941726e-07, + "loss": 0.6358, + "step": 6742 + }, + { + "epoch": 12.487037037037037, + "grad_norm": 1.4628326994055836, + "learning_rate": 1.6106629918502068e-07, + "loss": 0.5937, + "step": 6743 + }, + { + "epoch": 12.488888888888889, + "grad_norm": 1.4370141627877862, + "learning_rate": 1.6105442470293248e-07, + "loss": 0.6698, + "step": 6744 + }, + { + "epoch": 12.49074074074074, + "grad_norm": 1.4019496182120215, + "learning_rate": 1.61042548848175e-07, + "loss": 0.6451, + "step": 6745 + }, + { + "epoch": 12.492592592592592, + "grad_norm": 1.5803883327203812, + "learning_rate": 1.6103067162101523e-07, + "loss": 0.6052, + "step": 6746 + }, + { + "epoch": 12.494444444444444, + "grad_norm": 1.4849732380091656, + "learning_rate": 1.6101879302172024e-07, + "loss": 0.6337, + "step": 6747 + }, + { + "epoch": 12.496296296296297, + "grad_norm": 1.5523537680839534, + "learning_rate": 1.6100691305055705e-07, + "loss": 0.6368, + "step": 6748 + }, + { + "epoch": 12.498148148148148, + "grad_norm": 1.4503353980944524, + "learning_rate": 1.6099503170779275e-07, + "loss": 0.6384, + "step": 6749 + }, + { + "epoch": 12.5, + "grad_norm": 1.336787061169782, + "learning_rate": 1.6098314899369444e-07, + "loss": 0.6313, + "step": 6750 + }, + { + "epoch": 12.501851851851852, + "grad_norm": 1.58137241698652, + "learning_rate": 1.6097126490852936e-07, + "loss": 0.6414, + "step": 6751 + }, + { + "epoch": 12.503703703703703, + "grad_norm": 1.4015886532799962, + "learning_rate": 1.6095937945256463e-07, + "loss": 0.6134, + "step": 6752 + }, + { + "epoch": 12.505555555555556, + "grad_norm": 3.0615025113085435, + "learning_rate": 1.609474926260675e-07, + "loss": 0.6017, + "step": 6753 + }, + { + "epoch": 12.507407407407408, + "grad_norm": 1.785876336755664, + "learning_rate": 1.609356044293052e-07, + "loss": 0.6475, + "step": 6754 + }, + { + "epoch": 12.50925925925926, + "grad_norm": 1.549268553515418, + "learning_rate": 1.6092371486254505e-07, + "loss": 0.6153, + "step": 6755 + }, + { + "epoch": 12.511111111111111, + "grad_norm": 1.5512241377844795, + "learning_rate": 1.6091182392605424e-07, + "loss": 0.6657, + "step": 6756 + }, + { + "epoch": 12.512962962962963, + "grad_norm": 1.919523385233102, + "learning_rate": 1.6089993162010025e-07, + "loss": 0.628, + "step": 6757 + }, + { + "epoch": 12.514814814814814, + "grad_norm": 1.5779930159400157, + "learning_rate": 1.6088803794495039e-07, + "loss": 0.6121, + "step": 6758 + }, + { + "epoch": 12.516666666666667, + "grad_norm": 1.453727811521583, + "learning_rate": 1.6087614290087206e-07, + "loss": 0.6021, + "step": 6759 + }, + { + "epoch": 12.518518518518519, + "grad_norm": 1.606121439166297, + "learning_rate": 1.608642464881327e-07, + "loss": 0.5676, + "step": 6760 + }, + { + "epoch": 12.52037037037037, + "grad_norm": 1.491608123734651, + "learning_rate": 1.608523487069998e-07, + "loss": 0.6244, + "step": 6761 + }, + { + "epoch": 12.522222222222222, + "grad_norm": 1.4895390869520764, + "learning_rate": 1.608404495577408e-07, + "loss": 0.64, + "step": 6762 + }, + { + "epoch": 12.524074074074074, + "grad_norm": 1.576671313126675, + "learning_rate": 1.6082854904062325e-07, + "loss": 0.6252, + "step": 6763 + }, + { + "epoch": 12.525925925925925, + "grad_norm": 1.744803875587018, + "learning_rate": 1.6081664715591472e-07, + "loss": 0.5892, + "step": 6764 + }, + { + "epoch": 12.527777777777779, + "grad_norm": 1.5210954340070728, + "learning_rate": 1.608047439038828e-07, + "loss": 0.6102, + "step": 6765 + }, + { + "epoch": 12.52962962962963, + "grad_norm": 1.6319152672684785, + "learning_rate": 1.6079283928479508e-07, + "loss": 0.5924, + "step": 6766 + }, + { + "epoch": 12.531481481481482, + "grad_norm": 2.8294307494249327, + "learning_rate": 1.607809332989192e-07, + "loss": 0.6504, + "step": 6767 + }, + { + "epoch": 12.533333333333333, + "grad_norm": 1.704731451763355, + "learning_rate": 1.6076902594652288e-07, + "loss": 0.5722, + "step": 6768 + }, + { + "epoch": 12.535185185185185, + "grad_norm": 1.7691983169778456, + "learning_rate": 1.6075711722787385e-07, + "loss": 0.6573, + "step": 6769 + }, + { + "epoch": 12.537037037037036, + "grad_norm": 2.1711031930002163, + "learning_rate": 1.6074520714323977e-07, + "loss": 0.5993, + "step": 6770 + }, + { + "epoch": 12.53888888888889, + "grad_norm": 1.6125800014290208, + "learning_rate": 1.6073329569288843e-07, + "loss": 0.6025, + "step": 6771 + }, + { + "epoch": 12.540740740740741, + "grad_norm": 1.5873208356131967, + "learning_rate": 1.6072138287708767e-07, + "loss": 0.6335, + "step": 6772 + }, + { + "epoch": 12.542592592592593, + "grad_norm": 1.6046972054598476, + "learning_rate": 1.6070946869610532e-07, + "loss": 0.5901, + "step": 6773 + }, + { + "epoch": 12.544444444444444, + "grad_norm": 1.595423085591024, + "learning_rate": 1.6069755315020924e-07, + "loss": 0.6045, + "step": 6774 + }, + { + "epoch": 12.546296296296296, + "grad_norm": 1.9371582764788189, + "learning_rate": 1.606856362396673e-07, + "loss": 0.5927, + "step": 6775 + }, + { + "epoch": 12.548148148148147, + "grad_norm": 1.5694052249405797, + "learning_rate": 1.6067371796474742e-07, + "loss": 0.6323, + "step": 6776 + }, + { + "epoch": 12.55, + "grad_norm": 1.5614783347936312, + "learning_rate": 1.606617983257176e-07, + "loss": 0.585, + "step": 6777 + }, + { + "epoch": 12.551851851851852, + "grad_norm": 4.647419220422835, + "learning_rate": 1.606498773228458e-07, + "loss": 0.5856, + "step": 6778 + }, + { + "epoch": 12.553703703703704, + "grad_norm": 1.9430957089556058, + "learning_rate": 1.606379549564e-07, + "loss": 0.6378, + "step": 6779 + }, + { + "epoch": 12.555555555555555, + "grad_norm": 1.6921605452649477, + "learning_rate": 1.6062603122664831e-07, + "loss": 0.6167, + "step": 6780 + }, + { + "epoch": 12.557407407407407, + "grad_norm": 1.729361586729261, + "learning_rate": 1.6061410613385879e-07, + "loss": 0.6161, + "step": 6781 + }, + { + "epoch": 12.559259259259258, + "grad_norm": 2.160195075360732, + "learning_rate": 1.6060217967829952e-07, + "loss": 0.6505, + "step": 6782 + }, + { + "epoch": 12.561111111111112, + "grad_norm": 1.4684027863187157, + "learning_rate": 1.6059025186023864e-07, + "loss": 0.5937, + "step": 6783 + }, + { + "epoch": 12.562962962962963, + "grad_norm": 1.746339979668932, + "learning_rate": 1.6057832267994437e-07, + "loss": 0.6088, + "step": 6784 + }, + { + "epoch": 12.564814814814815, + "grad_norm": 1.4975468709928144, + "learning_rate": 1.605663921376849e-07, + "loss": 0.5995, + "step": 6785 + }, + { + "epoch": 12.566666666666666, + "grad_norm": 1.7525048146158066, + "learning_rate": 1.6055446023372836e-07, + "loss": 0.623, + "step": 6786 + }, + { + "epoch": 12.568518518518518, + "grad_norm": 1.6800059561658405, + "learning_rate": 1.6054252696834314e-07, + "loss": 0.6074, + "step": 6787 + }, + { + "epoch": 12.57037037037037, + "grad_norm": 1.9485820795576567, + "learning_rate": 1.6053059234179749e-07, + "loss": 0.6501, + "step": 6788 + }, + { + "epoch": 12.572222222222223, + "grad_norm": 1.5573476411654357, + "learning_rate": 1.6051865635435973e-07, + "loss": 0.5981, + "step": 6789 + }, + { + "epoch": 12.574074074074074, + "grad_norm": 1.5966695575987275, + "learning_rate": 1.6050671900629818e-07, + "loss": 0.6331, + "step": 6790 + }, + { + "epoch": 12.575925925925926, + "grad_norm": 2.451439210832922, + "learning_rate": 1.6049478029788127e-07, + "loss": 0.6374, + "step": 6791 + }, + { + "epoch": 12.577777777777778, + "grad_norm": 2.399746586688835, + "learning_rate": 1.604828402293774e-07, + "loss": 0.5894, + "step": 6792 + }, + { + "epoch": 12.579629629629629, + "grad_norm": 1.7952955023936086, + "learning_rate": 1.60470898801055e-07, + "loss": 0.6445, + "step": 6793 + }, + { + "epoch": 12.581481481481482, + "grad_norm": 1.6228837070931483, + "learning_rate": 1.6045895601318258e-07, + "loss": 0.5721, + "step": 6794 + }, + { + "epoch": 12.583333333333334, + "grad_norm": 1.7018523689577818, + "learning_rate": 1.604470118660286e-07, + "loss": 0.6271, + "step": 6795 + }, + { + "epoch": 12.585185185185185, + "grad_norm": 1.379640586503175, + "learning_rate": 1.6043506635986162e-07, + "loss": 0.5684, + "step": 6796 + }, + { + "epoch": 12.587037037037037, + "grad_norm": 1.7931694900504649, + "learning_rate": 1.6042311949495024e-07, + "loss": 0.6121, + "step": 6797 + }, + { + "epoch": 12.588888888888889, + "grad_norm": 1.7854155733545496, + "learning_rate": 1.60411171271563e-07, + "loss": 0.6266, + "step": 6798 + }, + { + "epoch": 12.59074074074074, + "grad_norm": 1.717485265669741, + "learning_rate": 1.6039922168996857e-07, + "loss": 0.6164, + "step": 6799 + }, + { + "epoch": 12.592592592592592, + "grad_norm": 1.7821970231296957, + "learning_rate": 1.6038727075043558e-07, + "loss": 0.5831, + "step": 6800 + }, + { + "epoch": 12.594444444444445, + "grad_norm": 1.5811840551132714, + "learning_rate": 1.6037531845323274e-07, + "loss": 0.6138, + "step": 6801 + }, + { + "epoch": 12.596296296296297, + "grad_norm": 1.411211185081435, + "learning_rate": 1.603633647986288e-07, + "loss": 0.6105, + "step": 6802 + }, + { + "epoch": 12.598148148148148, + "grad_norm": 1.7168552057894608, + "learning_rate": 1.6035140978689242e-07, + "loss": 0.6202, + "step": 6803 + }, + { + "epoch": 12.6, + "grad_norm": 1.8197273072150357, + "learning_rate": 1.6033945341829247e-07, + "loss": 0.6073, + "step": 6804 + }, + { + "epoch": 12.601851851851851, + "grad_norm": 3.783030194013275, + "learning_rate": 1.603274956930977e-07, + "loss": 0.6039, + "step": 6805 + }, + { + "epoch": 12.603703703703705, + "grad_norm": 1.4465897650135064, + "learning_rate": 1.60315536611577e-07, + "loss": 0.5757, + "step": 6806 + }, + { + "epoch": 12.605555555555556, + "grad_norm": 1.7660923193512383, + "learning_rate": 1.603035761739992e-07, + "loss": 0.6213, + "step": 6807 + }, + { + "epoch": 12.607407407407408, + "grad_norm": 1.5832966737920553, + "learning_rate": 1.6029161438063325e-07, + "loss": 0.6123, + "step": 6808 + }, + { + "epoch": 12.60925925925926, + "grad_norm": 1.4625725710762831, + "learning_rate": 1.6027965123174806e-07, + "loss": 0.6183, + "step": 6809 + }, + { + "epoch": 12.61111111111111, + "grad_norm": 1.4750028356899958, + "learning_rate": 1.602676867276126e-07, + "loss": 0.6155, + "step": 6810 + }, + { + "epoch": 12.612962962962962, + "grad_norm": 1.4285935297020942, + "learning_rate": 1.6025572086849583e-07, + "loss": 0.6567, + "step": 6811 + }, + { + "epoch": 12.614814814814816, + "grad_norm": 1.7802253508382944, + "learning_rate": 1.6024375365466683e-07, + "loss": 0.5838, + "step": 6812 + }, + { + "epoch": 12.616666666666667, + "grad_norm": 1.563898316628401, + "learning_rate": 1.6023178508639462e-07, + "loss": 0.6276, + "step": 6813 + }, + { + "epoch": 12.618518518518519, + "grad_norm": 1.6908489757004845, + "learning_rate": 1.6021981516394826e-07, + "loss": 0.598, + "step": 6814 + }, + { + "epoch": 12.62037037037037, + "grad_norm": 1.5568319661955032, + "learning_rate": 1.6020784388759695e-07, + "loss": 0.617, + "step": 6815 + }, + { + "epoch": 12.622222222222222, + "grad_norm": 1.61159548424979, + "learning_rate": 1.6019587125760977e-07, + "loss": 0.5927, + "step": 6816 + }, + { + "epoch": 12.624074074074073, + "grad_norm": 1.4522496756525654, + "learning_rate": 1.6018389727425592e-07, + "loss": 0.6019, + "step": 6817 + }, + { + "epoch": 12.625925925925927, + "grad_norm": 1.6529772869382366, + "learning_rate": 1.601719219378046e-07, + "loss": 0.6439, + "step": 6818 + }, + { + "epoch": 12.627777777777778, + "grad_norm": 1.7132252520891806, + "learning_rate": 1.6015994524852504e-07, + "loss": 0.6447, + "step": 6819 + }, + { + "epoch": 12.62962962962963, + "grad_norm": 1.4998307154610613, + "learning_rate": 1.6014796720668652e-07, + "loss": 0.6528, + "step": 6820 + }, + { + "epoch": 12.631481481481481, + "grad_norm": 1.5692311404371204, + "learning_rate": 1.6013598781255831e-07, + "loss": 0.6305, + "step": 6821 + }, + { + "epoch": 12.633333333333333, + "grad_norm": 1.543445686295679, + "learning_rate": 1.6012400706640983e-07, + "loss": 0.6889, + "step": 6822 + }, + { + "epoch": 12.635185185185184, + "grad_norm": 2.3071664843383792, + "learning_rate": 1.6011202496851034e-07, + "loss": 0.6086, + "step": 6823 + }, + { + "epoch": 12.637037037037038, + "grad_norm": 1.7648127834545924, + "learning_rate": 1.6010004151912926e-07, + "loss": 0.6412, + "step": 6824 + }, + { + "epoch": 12.63888888888889, + "grad_norm": 1.4903289633097423, + "learning_rate": 1.6008805671853605e-07, + "loss": 0.6048, + "step": 6825 + }, + { + "epoch": 12.64074074074074, + "grad_norm": 1.5962944557466139, + "learning_rate": 1.600760705670001e-07, + "loss": 0.649, + "step": 6826 + }, + { + "epoch": 12.642592592592592, + "grad_norm": 1.581592432581292, + "learning_rate": 1.600640830647909e-07, + "loss": 0.6207, + "step": 6827 + }, + { + "epoch": 12.644444444444444, + "grad_norm": 1.8286967803900258, + "learning_rate": 1.6005209421217803e-07, + "loss": 0.6303, + "step": 6828 + }, + { + "epoch": 12.646296296296295, + "grad_norm": 1.3631116369340788, + "learning_rate": 1.6004010400943097e-07, + "loss": 0.6053, + "step": 6829 + }, + { + "epoch": 12.648148148148149, + "grad_norm": 1.4722906278412151, + "learning_rate": 1.6002811245681929e-07, + "loss": 0.6349, + "step": 6830 + }, + { + "epoch": 12.65, + "grad_norm": 1.7000184921463328, + "learning_rate": 1.6001611955461264e-07, + "loss": 0.6113, + "step": 6831 + }, + { + "epoch": 12.651851851851852, + "grad_norm": 1.557111446109822, + "learning_rate": 1.6000412530308055e-07, + "loss": 0.5934, + "step": 6832 + }, + { + "epoch": 12.653703703703703, + "grad_norm": 3.9814872391656166, + "learning_rate": 1.5999212970249282e-07, + "loss": 0.6469, + "step": 6833 + }, + { + "epoch": 12.655555555555555, + "grad_norm": 1.4576137386246166, + "learning_rate": 1.5998013275311905e-07, + "loss": 0.5807, + "step": 6834 + }, + { + "epoch": 12.657407407407408, + "grad_norm": 1.4728520668214509, + "learning_rate": 1.5996813445522902e-07, + "loss": 0.6252, + "step": 6835 + }, + { + "epoch": 12.65925925925926, + "grad_norm": 2.160072341824441, + "learning_rate": 1.5995613480909243e-07, + "loss": 0.6194, + "step": 6836 + }, + { + "epoch": 12.661111111111111, + "grad_norm": 1.7776333713749342, + "learning_rate": 1.5994413381497908e-07, + "loss": 0.6191, + "step": 6837 + }, + { + "epoch": 12.662962962962963, + "grad_norm": 1.4491885027916003, + "learning_rate": 1.5993213147315886e-07, + "loss": 0.637, + "step": 6838 + }, + { + "epoch": 12.664814814814815, + "grad_norm": 1.5962358372956102, + "learning_rate": 1.5992012778390147e-07, + "loss": 0.6044, + "step": 6839 + }, + { + "epoch": 12.666666666666666, + "grad_norm": 1.5968504474145202, + "learning_rate": 1.5990812274747693e-07, + "loss": 0.6383, + "step": 6840 + }, + { + "epoch": 12.668518518518518, + "grad_norm": 1.568961725746722, + "learning_rate": 1.5989611636415504e-07, + "loss": 0.6345, + "step": 6841 + }, + { + "epoch": 12.670370370370371, + "grad_norm": 1.6413495044092195, + "learning_rate": 1.5988410863420578e-07, + "loss": 0.6342, + "step": 6842 + }, + { + "epoch": 12.672222222222222, + "grad_norm": 1.7594545805007782, + "learning_rate": 1.5987209955789913e-07, + "loss": 0.572, + "step": 6843 + }, + { + "epoch": 12.674074074074074, + "grad_norm": 1.7350443329675036, + "learning_rate": 1.5986008913550508e-07, + "loss": 0.6188, + "step": 6844 + }, + { + "epoch": 12.675925925925926, + "grad_norm": 1.7694067535008446, + "learning_rate": 1.5984807736729363e-07, + "loss": 0.581, + "step": 6845 + }, + { + "epoch": 12.677777777777777, + "grad_norm": 1.5073123721987824, + "learning_rate": 1.5983606425353488e-07, + "loss": 0.5741, + "step": 6846 + }, + { + "epoch": 12.67962962962963, + "grad_norm": 2.026385496371346, + "learning_rate": 1.5982404979449889e-07, + "loss": 0.6204, + "step": 6847 + }, + { + "epoch": 12.681481481481482, + "grad_norm": 1.9565160971881237, + "learning_rate": 1.5981203399045576e-07, + "loss": 0.5927, + "step": 6848 + }, + { + "epoch": 12.683333333333334, + "grad_norm": 1.4061477408305016, + "learning_rate": 1.5980001684167567e-07, + "loss": 0.6142, + "step": 6849 + }, + { + "epoch": 12.685185185185185, + "grad_norm": 1.484356137606153, + "learning_rate": 1.597879983484288e-07, + "loss": 0.6054, + "step": 6850 + }, + { + "epoch": 12.687037037037037, + "grad_norm": 1.6292511337993587, + "learning_rate": 1.597759785109853e-07, + "loss": 0.6308, + "step": 6851 + }, + { + "epoch": 12.688888888888888, + "grad_norm": 1.7648019980114067, + "learning_rate": 1.597639573296155e-07, + "loss": 0.5971, + "step": 6852 + }, + { + "epoch": 12.690740740740742, + "grad_norm": 1.5122404590378937, + "learning_rate": 1.5975193480458962e-07, + "loss": 0.6165, + "step": 6853 + }, + { + "epoch": 12.692592592592593, + "grad_norm": 2.1351292203467325, + "learning_rate": 1.5973991093617794e-07, + "loss": 0.61, + "step": 6854 + }, + { + "epoch": 12.694444444444445, + "grad_norm": 1.770293579050696, + "learning_rate": 1.5972788572465081e-07, + "loss": 0.6206, + "step": 6855 + }, + { + "epoch": 12.696296296296296, + "grad_norm": 1.6241130942930582, + "learning_rate": 1.597158591702786e-07, + "loss": 0.6677, + "step": 6856 + }, + { + "epoch": 12.698148148148148, + "grad_norm": 2.2842202003581766, + "learning_rate": 1.597038312733317e-07, + "loss": 0.6571, + "step": 6857 + }, + { + "epoch": 12.7, + "grad_norm": 1.6896403378883804, + "learning_rate": 1.596918020340805e-07, + "loss": 0.6133, + "step": 6858 + }, + { + "epoch": 12.701851851851853, + "grad_norm": 1.3828107158238865, + "learning_rate": 1.596797714527955e-07, + "loss": 0.6183, + "step": 6859 + }, + { + "epoch": 12.703703703703704, + "grad_norm": 1.5329900756319352, + "learning_rate": 1.5966773952974711e-07, + "loss": 0.5826, + "step": 6860 + }, + { + "epoch": 12.705555555555556, + "grad_norm": 1.3821215581887945, + "learning_rate": 1.596557062652059e-07, + "loss": 0.6027, + "step": 6861 + }, + { + "epoch": 12.707407407407407, + "grad_norm": 1.5824028104828178, + "learning_rate": 1.596436716594424e-07, + "loss": 0.6302, + "step": 6862 + }, + { + "epoch": 12.709259259259259, + "grad_norm": 10.540412692913531, + "learning_rate": 1.5963163571272717e-07, + "loss": 0.6208, + "step": 6863 + }, + { + "epoch": 12.71111111111111, + "grad_norm": 1.2948847490202897, + "learning_rate": 1.596195984253308e-07, + "loss": 0.5975, + "step": 6864 + }, + { + "epoch": 12.712962962962964, + "grad_norm": 1.648527501671313, + "learning_rate": 1.5960755979752395e-07, + "loss": 0.5892, + "step": 6865 + }, + { + "epoch": 12.714814814814815, + "grad_norm": 1.5823797338444276, + "learning_rate": 1.5959551982957726e-07, + "loss": 0.6602, + "step": 6866 + }, + { + "epoch": 12.716666666666667, + "grad_norm": 1.5387005759209103, + "learning_rate": 1.595834785217614e-07, + "loss": 0.6281, + "step": 6867 + }, + { + "epoch": 12.718518518518518, + "grad_norm": 1.464430108480258, + "learning_rate": 1.5957143587434715e-07, + "loss": 0.6149, + "step": 6868 + }, + { + "epoch": 12.72037037037037, + "grad_norm": 1.9434408643543009, + "learning_rate": 1.595593918876052e-07, + "loss": 0.5966, + "step": 6869 + }, + { + "epoch": 12.722222222222221, + "grad_norm": 1.8048751824567402, + "learning_rate": 1.5954734656180638e-07, + "loss": 0.622, + "step": 6870 + }, + { + "epoch": 12.724074074074075, + "grad_norm": 1.521188957720398, + "learning_rate": 1.5953529989722148e-07, + "loss": 0.6098, + "step": 6871 + }, + { + "epoch": 12.725925925925926, + "grad_norm": 1.9586869029061609, + "learning_rate": 1.5952325189412134e-07, + "loss": 0.634, + "step": 6872 + }, + { + "epoch": 12.727777777777778, + "grad_norm": 1.6100028935188142, + "learning_rate": 1.5951120255277682e-07, + "loss": 0.609, + "step": 6873 + }, + { + "epoch": 12.72962962962963, + "grad_norm": 1.5161375239048271, + "learning_rate": 1.5949915187345885e-07, + "loss": 0.6724, + "step": 6874 + }, + { + "epoch": 12.731481481481481, + "grad_norm": 1.5647460359540073, + "learning_rate": 1.5948709985643834e-07, + "loss": 0.6354, + "step": 6875 + }, + { + "epoch": 12.733333333333333, + "grad_norm": 1.7624037027376658, + "learning_rate": 1.5947504650198627e-07, + "loss": 0.5979, + "step": 6876 + }, + { + "epoch": 12.735185185185186, + "grad_norm": 1.5096608238512021, + "learning_rate": 1.5946299181037362e-07, + "loss": 0.597, + "step": 6877 + }, + { + "epoch": 12.737037037037037, + "grad_norm": 1.7792576772758397, + "learning_rate": 1.594509357818714e-07, + "loss": 0.6325, + "step": 6878 + }, + { + "epoch": 12.738888888888889, + "grad_norm": 1.5436864718796397, + "learning_rate": 1.5943887841675073e-07, + "loss": 0.6263, + "step": 6879 + }, + { + "epoch": 12.74074074074074, + "grad_norm": 1.8466790091564382, + "learning_rate": 1.5942681971528258e-07, + "loss": 0.6453, + "step": 6880 + }, + { + "epoch": 12.742592592592592, + "grad_norm": 1.4496392228453736, + "learning_rate": 1.5941475967773814e-07, + "loss": 0.616, + "step": 6881 + }, + { + "epoch": 12.744444444444444, + "grad_norm": 1.5716183231308856, + "learning_rate": 1.5940269830438852e-07, + "loss": 0.6161, + "step": 6882 + }, + { + "epoch": 12.746296296296297, + "grad_norm": 1.9364924908466352, + "learning_rate": 1.5939063559550495e-07, + "loss": 0.606, + "step": 6883 + }, + { + "epoch": 12.748148148148148, + "grad_norm": 1.6666008967833845, + "learning_rate": 1.5937857155135857e-07, + "loss": 0.599, + "step": 6884 + }, + { + "epoch": 12.75, + "grad_norm": 2.188154980147139, + "learning_rate": 1.593665061722206e-07, + "loss": 0.6101, + "step": 6885 + }, + { + "epoch": 12.751851851851852, + "grad_norm": 1.711231187628782, + "learning_rate": 1.5935443945836237e-07, + "loss": 0.564, + "step": 6886 + }, + { + "epoch": 12.753703703703703, + "grad_norm": 1.9539432327794675, + "learning_rate": 1.593423714100551e-07, + "loss": 0.6371, + "step": 6887 + }, + { + "epoch": 12.755555555555556, + "grad_norm": 1.4984715345635842, + "learning_rate": 1.593303020275702e-07, + "loss": 0.5845, + "step": 6888 + }, + { + "epoch": 12.757407407407408, + "grad_norm": 1.6281732241446114, + "learning_rate": 1.5931823131117893e-07, + "loss": 0.5992, + "step": 6889 + }, + { + "epoch": 12.75925925925926, + "grad_norm": 1.5732456862049327, + "learning_rate": 1.5930615926115273e-07, + "loss": 0.5998, + "step": 6890 + }, + { + "epoch": 12.761111111111111, + "grad_norm": 1.6336412374430413, + "learning_rate": 1.5929408587776303e-07, + "loss": 0.6292, + "step": 6891 + }, + { + "epoch": 12.762962962962963, + "grad_norm": 1.6595460275293108, + "learning_rate": 1.5928201116128118e-07, + "loss": 0.626, + "step": 6892 + }, + { + "epoch": 12.764814814814814, + "grad_norm": 1.7743089326630253, + "learning_rate": 1.5926993511197876e-07, + "loss": 0.6265, + "step": 6893 + }, + { + "epoch": 12.766666666666667, + "grad_norm": 1.4690682030967814, + "learning_rate": 1.592578577301272e-07, + "loss": 0.5921, + "step": 6894 + }, + { + "epoch": 12.768518518518519, + "grad_norm": 1.9112448566114246, + "learning_rate": 1.5924577901599808e-07, + "loss": 0.6525, + "step": 6895 + }, + { + "epoch": 12.77037037037037, + "grad_norm": 1.5096988302004613, + "learning_rate": 1.5923369896986291e-07, + "loss": 0.6245, + "step": 6896 + }, + { + "epoch": 12.772222222222222, + "grad_norm": 1.4499871108322002, + "learning_rate": 1.5922161759199332e-07, + "loss": 0.6091, + "step": 6897 + }, + { + "epoch": 12.774074074074074, + "grad_norm": 1.3584437905518292, + "learning_rate": 1.5920953488266094e-07, + "loss": 0.6296, + "step": 6898 + }, + { + "epoch": 12.775925925925925, + "grad_norm": 1.4414205456580722, + "learning_rate": 1.5919745084213737e-07, + "loss": 0.5988, + "step": 6899 + }, + { + "epoch": 12.777777777777779, + "grad_norm": 1.7255951279242674, + "learning_rate": 1.5918536547069434e-07, + "loss": 0.6115, + "step": 6900 + }, + { + "epoch": 12.77962962962963, + "grad_norm": 3.674450625944859, + "learning_rate": 1.591732787686036e-07, + "loss": 0.6289, + "step": 6901 + }, + { + "epoch": 12.781481481481482, + "grad_norm": 1.708081372876104, + "learning_rate": 1.5916119073613675e-07, + "loss": 0.636, + "step": 6902 + }, + { + "epoch": 12.783333333333333, + "grad_norm": 1.7121803383373337, + "learning_rate": 1.591491013735657e-07, + "loss": 0.5837, + "step": 6903 + }, + { + "epoch": 12.785185185185185, + "grad_norm": 1.4779633527254814, + "learning_rate": 1.591370106811622e-07, + "loss": 0.5874, + "step": 6904 + }, + { + "epoch": 12.787037037037036, + "grad_norm": 1.8615589615450188, + "learning_rate": 1.5912491865919808e-07, + "loss": 0.6467, + "step": 6905 + }, + { + "epoch": 12.78888888888889, + "grad_norm": 1.7408483278296838, + "learning_rate": 1.5911282530794524e-07, + "loss": 0.6539, + "step": 6906 + }, + { + "epoch": 12.790740740740741, + "grad_norm": 1.7013220753869909, + "learning_rate": 1.5910073062767552e-07, + "loss": 0.6232, + "step": 6907 + }, + { + "epoch": 12.792592592592593, + "grad_norm": 1.7918484485255404, + "learning_rate": 1.5908863461866085e-07, + "loss": 0.6218, + "step": 6908 + }, + { + "epoch": 12.794444444444444, + "grad_norm": 1.5314227850314672, + "learning_rate": 1.5907653728117317e-07, + "loss": 0.6062, + "step": 6909 + }, + { + "epoch": 12.796296296296296, + "grad_norm": 1.7631125019601799, + "learning_rate": 1.5906443861548453e-07, + "loss": 0.5881, + "step": 6910 + }, + { + "epoch": 12.798148148148147, + "grad_norm": 1.4775292759926024, + "learning_rate": 1.5905233862186686e-07, + "loss": 0.5994, + "step": 6911 + }, + { + "epoch": 12.8, + "grad_norm": 1.863183881251582, + "learning_rate": 1.5904023730059226e-07, + "loss": 0.6068, + "step": 6912 + }, + { + "epoch": 12.801851851851852, + "grad_norm": 1.5371736650753063, + "learning_rate": 1.590281346519328e-07, + "loss": 0.6102, + "step": 6913 + }, + { + "epoch": 12.803703703703704, + "grad_norm": 1.733718031950228, + "learning_rate": 1.5901603067616046e-07, + "loss": 0.6045, + "step": 6914 + }, + { + "epoch": 12.805555555555555, + "grad_norm": 1.4975076115540258, + "learning_rate": 1.5900392537354754e-07, + "loss": 0.6265, + "step": 6915 + }, + { + "epoch": 12.807407407407407, + "grad_norm": 1.5923752420447312, + "learning_rate": 1.5899181874436612e-07, + "loss": 0.5812, + "step": 6916 + }, + { + "epoch": 12.809259259259258, + "grad_norm": 1.6211205434507365, + "learning_rate": 1.5897971078888836e-07, + "loss": 0.6326, + "step": 6917 + }, + { + "epoch": 12.811111111111112, + "grad_norm": 1.6139633455756157, + "learning_rate": 1.5896760150738656e-07, + "loss": 0.6246, + "step": 6918 + }, + { + "epoch": 12.812962962962963, + "grad_norm": 1.5812682066273467, + "learning_rate": 1.589554909001329e-07, + "loss": 0.6211, + "step": 6919 + }, + { + "epoch": 12.814814814814815, + "grad_norm": 1.4334726785478022, + "learning_rate": 1.5894337896739968e-07, + "loss": 0.6422, + "step": 6920 + }, + { + "epoch": 12.816666666666666, + "grad_norm": 1.9094142664963845, + "learning_rate": 1.5893126570945922e-07, + "loss": 0.5732, + "step": 6921 + }, + { + "epoch": 12.818518518518518, + "grad_norm": 1.8636920512448556, + "learning_rate": 1.5891915112658385e-07, + "loss": 0.6007, + "step": 6922 + }, + { + "epoch": 12.82037037037037, + "grad_norm": 1.5996726394467193, + "learning_rate": 1.5890703521904593e-07, + "loss": 0.6209, + "step": 6923 + }, + { + "epoch": 12.822222222222223, + "grad_norm": 1.5243014029250295, + "learning_rate": 1.5889491798711791e-07, + "loss": 0.6179, + "step": 6924 + }, + { + "epoch": 12.824074074074074, + "grad_norm": 1.7310053055441255, + "learning_rate": 1.5888279943107213e-07, + "loss": 0.5861, + "step": 6925 + }, + { + "epoch": 12.825925925925926, + "grad_norm": 1.6452361730581457, + "learning_rate": 1.5887067955118111e-07, + "loss": 0.5392, + "step": 6926 + }, + { + "epoch": 12.827777777777778, + "grad_norm": 1.4391954697054465, + "learning_rate": 1.5885855834771734e-07, + "loss": 0.619, + "step": 6927 + }, + { + "epoch": 12.829629629629629, + "grad_norm": 1.9432479194480319, + "learning_rate": 1.5884643582095331e-07, + "loss": 0.5908, + "step": 6928 + }, + { + "epoch": 12.831481481481482, + "grad_norm": 1.4715285715227817, + "learning_rate": 1.5883431197116157e-07, + "loss": 0.5974, + "step": 6929 + }, + { + "epoch": 12.833333333333334, + "grad_norm": 1.5714643873072307, + "learning_rate": 1.5882218679861472e-07, + "loss": 0.6449, + "step": 6930 + }, + { + "epoch": 12.835185185185185, + "grad_norm": 1.647207993509161, + "learning_rate": 1.5881006030358535e-07, + "loss": 0.6473, + "step": 6931 + }, + { + "epoch": 12.837037037037037, + "grad_norm": 1.7201150031593586, + "learning_rate": 1.5879793248634608e-07, + "loss": 0.6268, + "step": 6932 + }, + { + "epoch": 12.838888888888889, + "grad_norm": 1.4418662937024034, + "learning_rate": 1.587858033471696e-07, + "loss": 0.6755, + "step": 6933 + }, + { + "epoch": 12.84074074074074, + "grad_norm": 1.4569044943214895, + "learning_rate": 1.5877367288632862e-07, + "loss": 0.6244, + "step": 6934 + }, + { + "epoch": 12.842592592592592, + "grad_norm": 1.6720215772575935, + "learning_rate": 1.5876154110409585e-07, + "loss": 0.6174, + "step": 6935 + }, + { + "epoch": 12.844444444444445, + "grad_norm": 1.4168324603730078, + "learning_rate": 1.5874940800074402e-07, + "loss": 0.6202, + "step": 6936 + }, + { + "epoch": 12.846296296296297, + "grad_norm": 1.7694777187568598, + "learning_rate": 1.5873727357654593e-07, + "loss": 0.6351, + "step": 6937 + }, + { + "epoch": 12.848148148148148, + "grad_norm": 2.1419415089418496, + "learning_rate": 1.5872513783177442e-07, + "loss": 0.6219, + "step": 6938 + }, + { + "epoch": 12.85, + "grad_norm": 1.8717743149142723, + "learning_rate": 1.5871300076670233e-07, + "loss": 0.6287, + "step": 6939 + }, + { + "epoch": 12.851851851851851, + "grad_norm": 1.943203339902939, + "learning_rate": 1.587008623816025e-07, + "loss": 0.6563, + "step": 6940 + }, + { + "epoch": 12.853703703703705, + "grad_norm": 2.703950936589274, + "learning_rate": 1.5868872267674785e-07, + "loss": 0.605, + "step": 6941 + }, + { + "epoch": 12.855555555555556, + "grad_norm": 1.624191640151769, + "learning_rate": 1.5867658165241133e-07, + "loss": 0.6027, + "step": 6942 + }, + { + "epoch": 12.857407407407408, + "grad_norm": 1.3510287854323508, + "learning_rate": 1.5866443930886588e-07, + "loss": 0.6227, + "step": 6943 + }, + { + "epoch": 12.85925925925926, + "grad_norm": 2.6022799631669016, + "learning_rate": 1.5865229564638452e-07, + "loss": 0.5973, + "step": 6944 + }, + { + "epoch": 12.86111111111111, + "grad_norm": 1.7242501443944458, + "learning_rate": 1.5864015066524026e-07, + "loss": 0.6333, + "step": 6945 + }, + { + "epoch": 12.862962962962962, + "grad_norm": 1.5306756716384318, + "learning_rate": 1.5862800436570613e-07, + "loss": 0.6768, + "step": 6946 + }, + { + "epoch": 12.864814814814816, + "grad_norm": 1.518392861496032, + "learning_rate": 1.5861585674805524e-07, + "loss": 0.6613, + "step": 6947 + }, + { + "epoch": 12.866666666666667, + "grad_norm": 1.5098754002583945, + "learning_rate": 1.586037078125607e-07, + "loss": 0.6401, + "step": 6948 + }, + { + "epoch": 12.868518518518519, + "grad_norm": 1.4303987804425788, + "learning_rate": 1.5859155755949563e-07, + "loss": 0.6113, + "step": 6949 + }, + { + "epoch": 12.87037037037037, + "grad_norm": 1.449785974003684, + "learning_rate": 1.5857940598913323e-07, + "loss": 0.6035, + "step": 6950 + }, + { + "epoch": 12.872222222222222, + "grad_norm": 1.4559852264373372, + "learning_rate": 1.5856725310174667e-07, + "loss": 0.5877, + "step": 6951 + }, + { + "epoch": 12.874074074074073, + "grad_norm": 1.6857319457332511, + "learning_rate": 1.5855509889760917e-07, + "loss": 0.6279, + "step": 6952 + }, + { + "epoch": 12.875925925925927, + "grad_norm": 1.7141514974392464, + "learning_rate": 1.5854294337699406e-07, + "loss": 0.6412, + "step": 6953 + }, + { + "epoch": 12.877777777777778, + "grad_norm": 1.6928035016161531, + "learning_rate": 1.5853078654017456e-07, + "loss": 0.6583, + "step": 6954 + }, + { + "epoch": 12.87962962962963, + "grad_norm": 1.6064487142525163, + "learning_rate": 1.5851862838742397e-07, + "loss": 0.6488, + "step": 6955 + }, + { + "epoch": 12.881481481481481, + "grad_norm": 1.7822770556018663, + "learning_rate": 1.5850646891901575e-07, + "loss": 0.6116, + "step": 6956 + }, + { + "epoch": 12.883333333333333, + "grad_norm": 1.7613025405935119, + "learning_rate": 1.584943081352232e-07, + "loss": 0.6449, + "step": 6957 + }, + { + "epoch": 12.885185185185184, + "grad_norm": 1.580041681251821, + "learning_rate": 1.5848214603631972e-07, + "loss": 0.6504, + "step": 6958 + }, + { + "epoch": 12.887037037037038, + "grad_norm": 1.8036107019762868, + "learning_rate": 1.5846998262257872e-07, + "loss": 0.639, + "step": 6959 + }, + { + "epoch": 12.88888888888889, + "grad_norm": 1.5809629190605328, + "learning_rate": 1.5845781789427375e-07, + "loss": 0.6227, + "step": 6960 + }, + { + "epoch": 12.89074074074074, + "grad_norm": 1.5945560407753643, + "learning_rate": 1.5844565185167825e-07, + "loss": 0.6073, + "step": 6961 + }, + { + "epoch": 12.892592592592592, + "grad_norm": 1.831433857915805, + "learning_rate": 1.584334844950658e-07, + "loss": 0.6301, + "step": 6962 + }, + { + "epoch": 12.894444444444444, + "grad_norm": 1.568157789412509, + "learning_rate": 1.584213158247099e-07, + "loss": 0.5873, + "step": 6963 + }, + { + "epoch": 12.896296296296295, + "grad_norm": 1.4298543076861676, + "learning_rate": 1.5840914584088412e-07, + "loss": 0.5826, + "step": 6964 + }, + { + "epoch": 12.898148148148149, + "grad_norm": 1.6206316375965726, + "learning_rate": 1.5839697454386212e-07, + "loss": 0.6048, + "step": 6965 + }, + { + "epoch": 12.9, + "grad_norm": 1.7415934414137002, + "learning_rate": 1.5838480193391753e-07, + "loss": 0.6275, + "step": 6966 + }, + { + "epoch": 12.901851851851852, + "grad_norm": 1.5390717404474823, + "learning_rate": 1.58372628011324e-07, + "loss": 0.6191, + "step": 6967 + }, + { + "epoch": 12.903703703703703, + "grad_norm": 1.6978985594300364, + "learning_rate": 1.583604527763553e-07, + "loss": 0.5884, + "step": 6968 + }, + { + "epoch": 12.905555555555555, + "grad_norm": 1.4935896724237472, + "learning_rate": 1.583482762292851e-07, + "loss": 0.612, + "step": 6969 + }, + { + "epoch": 12.907407407407408, + "grad_norm": 1.63411268014853, + "learning_rate": 1.5833609837038716e-07, + "loss": 0.6193, + "step": 6970 + }, + { + "epoch": 12.90925925925926, + "grad_norm": 1.7637309920632391, + "learning_rate": 1.583239191999353e-07, + "loss": 0.6573, + "step": 6971 + }, + { + "epoch": 12.911111111111111, + "grad_norm": 1.4453907275111304, + "learning_rate": 1.5831173871820336e-07, + "loss": 0.6431, + "step": 6972 + }, + { + "epoch": 12.912962962962963, + "grad_norm": 1.7008633559195536, + "learning_rate": 1.5829955692546514e-07, + "loss": 0.6554, + "step": 6973 + }, + { + "epoch": 12.914814814814815, + "grad_norm": 1.512307483605181, + "learning_rate": 1.5828737382199452e-07, + "loss": 0.6205, + "step": 6974 + }, + { + "epoch": 12.916666666666666, + "grad_norm": 1.5766727888728036, + "learning_rate": 1.5827518940806543e-07, + "loss": 0.6312, + "step": 6975 + }, + { + "epoch": 12.918518518518518, + "grad_norm": 1.6919612820452719, + "learning_rate": 1.5826300368395184e-07, + "loss": 0.6676, + "step": 6976 + }, + { + "epoch": 12.920370370370371, + "grad_norm": 1.7250342390113371, + "learning_rate": 1.5825081664992767e-07, + "loss": 0.6012, + "step": 6977 + }, + { + "epoch": 12.922222222222222, + "grad_norm": 1.4318583791806843, + "learning_rate": 1.5823862830626694e-07, + "loss": 0.6358, + "step": 6978 + }, + { + "epoch": 12.924074074074074, + "grad_norm": 1.7413914272721611, + "learning_rate": 1.5822643865324365e-07, + "loss": 0.6111, + "step": 6979 + }, + { + "epoch": 12.925925925925926, + "grad_norm": 1.622968580250366, + "learning_rate": 1.5821424769113191e-07, + "loss": 0.6435, + "step": 6980 + }, + { + "epoch": 12.927777777777777, + "grad_norm": 1.5138749289499516, + "learning_rate": 1.5820205542020578e-07, + "loss": 0.5911, + "step": 6981 + }, + { + "epoch": 12.92962962962963, + "grad_norm": 1.643142531046615, + "learning_rate": 1.5818986184073935e-07, + "loss": 0.5724, + "step": 6982 + }, + { + "epoch": 12.931481481481482, + "grad_norm": 1.423707992260151, + "learning_rate": 1.5817766695300678e-07, + "loss": 0.6058, + "step": 6983 + }, + { + "epoch": 12.933333333333334, + "grad_norm": 1.3639529825988501, + "learning_rate": 1.5816547075728226e-07, + "loss": 0.6187, + "step": 6984 + }, + { + "epoch": 12.935185185185185, + "grad_norm": 1.5587202908539697, + "learning_rate": 1.5815327325383998e-07, + "loss": 0.6304, + "step": 6985 + }, + { + "epoch": 12.937037037037037, + "grad_norm": 1.4850261269029026, + "learning_rate": 1.5814107444295415e-07, + "loss": 0.6534, + "step": 6986 + }, + { + "epoch": 12.938888888888888, + "grad_norm": 1.8257773458899504, + "learning_rate": 1.581288743248991e-07, + "loss": 0.5648, + "step": 6987 + }, + { + "epoch": 12.940740740740742, + "grad_norm": 1.7394621738283145, + "learning_rate": 1.5811667289994907e-07, + "loss": 0.6233, + "step": 6988 + }, + { + "epoch": 12.942592592592593, + "grad_norm": 1.6293088979028831, + "learning_rate": 1.5810447016837838e-07, + "loss": 0.6221, + "step": 6989 + }, + { + "epoch": 12.944444444444445, + "grad_norm": 4.663484142314459, + "learning_rate": 1.580922661304614e-07, + "loss": 0.6473, + "step": 6990 + }, + { + "epoch": 12.946296296296296, + "grad_norm": 1.6690144557644095, + "learning_rate": 1.5808006078647248e-07, + "loss": 0.6027, + "step": 6991 + }, + { + "epoch": 12.948148148148148, + "grad_norm": 2.1842778211814156, + "learning_rate": 1.5806785413668606e-07, + "loss": 0.6497, + "step": 6992 + }, + { + "epoch": 12.95, + "grad_norm": 1.4722082470509708, + "learning_rate": 1.5805564618137658e-07, + "loss": 0.6214, + "step": 6993 + }, + { + "epoch": 12.951851851851853, + "grad_norm": 1.6097807020775359, + "learning_rate": 1.5804343692081846e-07, + "loss": 0.6343, + "step": 6994 + }, + { + "epoch": 12.953703703703704, + "grad_norm": 1.4633283074982837, + "learning_rate": 1.580312263552863e-07, + "loss": 0.6109, + "step": 6995 + }, + { + "epoch": 12.955555555555556, + "grad_norm": 1.5217539361475318, + "learning_rate": 1.5801901448505453e-07, + "loss": 0.59, + "step": 6996 + }, + { + "epoch": 12.957407407407407, + "grad_norm": 1.6648615011599652, + "learning_rate": 1.5800680131039772e-07, + "loss": 0.5859, + "step": 6997 + }, + { + "epoch": 12.959259259259259, + "grad_norm": 1.4857559300464391, + "learning_rate": 1.579945868315905e-07, + "loss": 0.6228, + "step": 6998 + }, + { + "epoch": 12.96111111111111, + "grad_norm": 1.6245193304820784, + "learning_rate": 1.5798237104890743e-07, + "loss": 0.6035, + "step": 6999 + }, + { + "epoch": 12.962962962962964, + "grad_norm": 1.7805554382779154, + "learning_rate": 1.579701539626232e-07, + "loss": 0.6077, + "step": 7000 + }, + { + "epoch": 12.964814814814815, + "grad_norm": 1.576147586724914, + "learning_rate": 1.5795793557301244e-07, + "loss": 0.6575, + "step": 7001 + }, + { + "epoch": 12.966666666666667, + "grad_norm": 1.6560448293355672, + "learning_rate": 1.579457158803499e-07, + "loss": 0.5856, + "step": 7002 + }, + { + "epoch": 12.968518518518518, + "grad_norm": 1.5203719283934036, + "learning_rate": 1.579334948849103e-07, + "loss": 0.5934, + "step": 7003 + }, + { + "epoch": 12.97037037037037, + "grad_norm": 1.4615112475366825, + "learning_rate": 1.5792127258696838e-07, + "loss": 0.6516, + "step": 7004 + }, + { + "epoch": 12.972222222222221, + "grad_norm": 1.346287901487352, + "learning_rate": 1.5790904898679892e-07, + "loss": 0.6213, + "step": 7005 + }, + { + "epoch": 12.974074074074075, + "grad_norm": 1.4355437594059939, + "learning_rate": 1.5789682408467677e-07, + "loss": 0.6156, + "step": 7006 + }, + { + "epoch": 12.975925925925926, + "grad_norm": 1.5389726803122858, + "learning_rate": 1.578845978808768e-07, + "loss": 0.6665, + "step": 7007 + }, + { + "epoch": 12.977777777777778, + "grad_norm": 1.6814046833548433, + "learning_rate": 1.5787237037567382e-07, + "loss": 0.6415, + "step": 7008 + }, + { + "epoch": 12.97962962962963, + "grad_norm": 1.609268038728392, + "learning_rate": 1.5786014156934277e-07, + "loss": 0.6169, + "step": 7009 + }, + { + "epoch": 12.981481481481481, + "grad_norm": 1.6100856550206866, + "learning_rate": 1.578479114621586e-07, + "loss": 0.5911, + "step": 7010 + }, + { + "epoch": 12.983333333333333, + "grad_norm": 1.5181380655697292, + "learning_rate": 1.5783568005439629e-07, + "loss": 0.6462, + "step": 7011 + }, + { + "epoch": 12.985185185185186, + "grad_norm": 1.3889241714683453, + "learning_rate": 1.5782344734633078e-07, + "loss": 0.6237, + "step": 7012 + }, + { + "epoch": 12.987037037037037, + "grad_norm": 1.3594050534025974, + "learning_rate": 1.5781121333823716e-07, + "loss": 0.6205, + "step": 7013 + }, + { + "epoch": 12.988888888888889, + "grad_norm": 1.4884411938890487, + "learning_rate": 1.5779897803039045e-07, + "loss": 0.5819, + "step": 7014 + }, + { + "epoch": 12.99074074074074, + "grad_norm": 1.4055462518083262, + "learning_rate": 1.5778674142306572e-07, + "loss": 0.6102, + "step": 7015 + }, + { + "epoch": 12.992592592592592, + "grad_norm": 1.7603573039322413, + "learning_rate": 1.577745035165381e-07, + "loss": 0.6351, + "step": 7016 + }, + { + "epoch": 12.994444444444444, + "grad_norm": 6.30460241917364, + "learning_rate": 1.5776226431108273e-07, + "loss": 0.583, + "step": 7017 + }, + { + "epoch": 12.996296296296297, + "grad_norm": 1.4766214122948749, + "learning_rate": 1.5775002380697474e-07, + "loss": 0.6199, + "step": 7018 + }, + { + "epoch": 12.998148148148148, + "grad_norm": 1.5270464215188726, + "learning_rate": 1.5773778200448944e-07, + "loss": 0.6206, + "step": 7019 + }, + { + "epoch": 13.0, + "grad_norm": 1.6216188368603912, + "learning_rate": 1.5772553890390196e-07, + "loss": 0.597, + "step": 7020 + }, + { + "epoch": 13.001851851851852, + "grad_norm": 1.4958417777720983, + "learning_rate": 1.5771329450548757e-07, + "loss": 0.6209, + "step": 7021 + }, + { + "epoch": 13.003703703703703, + "grad_norm": 1.475676342937982, + "learning_rate": 1.5770104880952157e-07, + "loss": 0.6296, + "step": 7022 + }, + { + "epoch": 13.005555555555556, + "grad_norm": 1.4018004363315908, + "learning_rate": 1.576888018162793e-07, + "loss": 0.6302, + "step": 7023 + }, + { + "epoch": 13.007407407407408, + "grad_norm": 1.9329323468873099, + "learning_rate": 1.5767655352603606e-07, + "loss": 0.6204, + "step": 7024 + }, + { + "epoch": 13.00925925925926, + "grad_norm": 1.9800664296948831, + "learning_rate": 1.5766430393906727e-07, + "loss": 0.6187, + "step": 7025 + }, + { + "epoch": 13.011111111111111, + "grad_norm": 1.382362839353423, + "learning_rate": 1.576520530556483e-07, + "loss": 0.6031, + "step": 7026 + }, + { + "epoch": 13.012962962962963, + "grad_norm": 1.608236489165492, + "learning_rate": 1.5763980087605456e-07, + "loss": 0.6426, + "step": 7027 + }, + { + "epoch": 13.014814814814814, + "grad_norm": 2.225481161842278, + "learning_rate": 1.5762754740056163e-07, + "loss": 0.6479, + "step": 7028 + }, + { + "epoch": 13.016666666666667, + "grad_norm": 1.8838160375784836, + "learning_rate": 1.5761529262944487e-07, + "loss": 0.6572, + "step": 7029 + }, + { + "epoch": 13.018518518518519, + "grad_norm": 1.8351727014629593, + "learning_rate": 1.5760303656297988e-07, + "loss": 0.5973, + "step": 7030 + }, + { + "epoch": 13.02037037037037, + "grad_norm": 1.72451616371401, + "learning_rate": 1.5759077920144216e-07, + "loss": 0.5977, + "step": 7031 + }, + { + "epoch": 13.022222222222222, + "grad_norm": 1.560677502276557, + "learning_rate": 1.575785205451073e-07, + "loss": 0.6225, + "step": 7032 + }, + { + "epoch": 13.024074074074074, + "grad_norm": 1.79634903907718, + "learning_rate": 1.5756626059425093e-07, + "loss": 0.6229, + "step": 7033 + }, + { + "epoch": 13.025925925925925, + "grad_norm": 1.717439029823767, + "learning_rate": 1.5755399934914868e-07, + "loss": 0.6283, + "step": 7034 + }, + { + "epoch": 13.027777777777779, + "grad_norm": 1.542868156702427, + "learning_rate": 1.575417368100762e-07, + "loss": 0.5781, + "step": 7035 + }, + { + "epoch": 13.02962962962963, + "grad_norm": 1.5881375922006693, + "learning_rate": 1.575294729773092e-07, + "loss": 0.5924, + "step": 7036 + }, + { + "epoch": 13.031481481481482, + "grad_norm": 1.8992551398870448, + "learning_rate": 1.575172078511234e-07, + "loss": 0.5847, + "step": 7037 + }, + { + "epoch": 13.033333333333333, + "grad_norm": 1.6845763999849395, + "learning_rate": 1.5750494143179453e-07, + "loss": 0.6345, + "step": 7038 + }, + { + "epoch": 13.035185185185185, + "grad_norm": 1.4250985818152622, + "learning_rate": 1.5749267371959843e-07, + "loss": 0.6317, + "step": 7039 + }, + { + "epoch": 13.037037037037036, + "grad_norm": 1.6941437706055298, + "learning_rate": 1.5748040471481088e-07, + "loss": 0.6216, + "step": 7040 + }, + { + "epoch": 13.03888888888889, + "grad_norm": 1.6374785398221825, + "learning_rate": 1.574681344177077e-07, + "loss": 0.6524, + "step": 7041 + }, + { + "epoch": 13.040740740740741, + "grad_norm": 1.6135624625140734, + "learning_rate": 1.5745586282856475e-07, + "loss": 0.5744, + "step": 7042 + }, + { + "epoch": 13.042592592592593, + "grad_norm": 1.8245704324093361, + "learning_rate": 1.57443589947658e-07, + "loss": 0.6129, + "step": 7043 + }, + { + "epoch": 13.044444444444444, + "grad_norm": 2.6964552210065826, + "learning_rate": 1.5743131577526332e-07, + "loss": 0.6274, + "step": 7044 + }, + { + "epoch": 13.046296296296296, + "grad_norm": 1.46257375695365, + "learning_rate": 1.574190403116567e-07, + "loss": 0.5931, + "step": 7045 + }, + { + "epoch": 13.048148148148147, + "grad_norm": 1.4312721588467456, + "learning_rate": 1.574067635571141e-07, + "loss": 0.6314, + "step": 7046 + }, + { + "epoch": 13.05, + "grad_norm": 1.5314058617642976, + "learning_rate": 1.573944855119115e-07, + "loss": 0.6102, + "step": 7047 + }, + { + "epoch": 13.051851851851852, + "grad_norm": 1.7135392632513267, + "learning_rate": 1.57382206176325e-07, + "loss": 0.6992, + "step": 7048 + }, + { + "epoch": 13.053703703703704, + "grad_norm": 1.8784351213062935, + "learning_rate": 1.5736992555063066e-07, + "loss": 0.627, + "step": 7049 + }, + { + "epoch": 13.055555555555555, + "grad_norm": 2.750254011801841, + "learning_rate": 1.573576436351046e-07, + "loss": 0.5894, + "step": 7050 + }, + { + "epoch": 13.057407407407407, + "grad_norm": 1.5886573150655008, + "learning_rate": 1.5734536043002292e-07, + "loss": 0.6178, + "step": 7051 + }, + { + "epoch": 13.059259259259258, + "grad_norm": 1.66627603545752, + "learning_rate": 1.573330759356618e-07, + "loss": 0.6044, + "step": 7052 + }, + { + "epoch": 13.061111111111112, + "grad_norm": 1.813911209527428, + "learning_rate": 1.573207901522974e-07, + "loss": 0.5918, + "step": 7053 + }, + { + "epoch": 13.062962962962963, + "grad_norm": 1.6198117244874142, + "learning_rate": 1.5730850308020597e-07, + "loss": 0.5649, + "step": 7054 + }, + { + "epoch": 13.064814814814815, + "grad_norm": 1.5884962898388135, + "learning_rate": 1.5729621471966375e-07, + "loss": 0.6669, + "step": 7055 + }, + { + "epoch": 13.066666666666666, + "grad_norm": 1.5741596935430127, + "learning_rate": 1.5728392507094696e-07, + "loss": 0.6259, + "step": 7056 + }, + { + "epoch": 13.068518518518518, + "grad_norm": 1.5108260359930092, + "learning_rate": 1.57271634134332e-07, + "loss": 0.617, + "step": 7057 + }, + { + "epoch": 13.07037037037037, + "grad_norm": 1.7346379844524948, + "learning_rate": 1.5725934191009514e-07, + "loss": 0.6281, + "step": 7058 + }, + { + "epoch": 13.072222222222223, + "grad_norm": 1.890229380921913, + "learning_rate": 1.5724704839851277e-07, + "loss": 0.6647, + "step": 7059 + }, + { + "epoch": 13.074074074074074, + "grad_norm": 1.6543594369475483, + "learning_rate": 1.5723475359986126e-07, + "loss": 0.5998, + "step": 7060 + }, + { + "epoch": 13.075925925925926, + "grad_norm": 1.620814655848228, + "learning_rate": 1.5722245751441708e-07, + "loss": 0.5585, + "step": 7061 + }, + { + "epoch": 13.077777777777778, + "grad_norm": 2.058514693021922, + "learning_rate": 1.5721016014245658e-07, + "loss": 0.5706, + "step": 7062 + }, + { + "epoch": 13.079629629629629, + "grad_norm": 2.0494715950841305, + "learning_rate": 1.5719786148425633e-07, + "loss": 0.6655, + "step": 7063 + }, + { + "epoch": 13.081481481481482, + "grad_norm": 1.705113359979874, + "learning_rate": 1.571855615400928e-07, + "loss": 0.6096, + "step": 7064 + }, + { + "epoch": 13.083333333333334, + "grad_norm": 1.4199487604869294, + "learning_rate": 1.5717326031024254e-07, + "loss": 0.6267, + "step": 7065 + }, + { + "epoch": 13.085185185185185, + "grad_norm": 3.5770625576567316, + "learning_rate": 1.5716095779498213e-07, + "loss": 0.6881, + "step": 7066 + }, + { + "epoch": 13.087037037037037, + "grad_norm": 3.338691735492136, + "learning_rate": 1.571486539945881e-07, + "loss": 0.6055, + "step": 7067 + }, + { + "epoch": 13.088888888888889, + "grad_norm": 1.4570520402825595, + "learning_rate": 1.5713634890933715e-07, + "loss": 0.6193, + "step": 7068 + }, + { + "epoch": 13.09074074074074, + "grad_norm": 1.5439779787829557, + "learning_rate": 1.5712404253950587e-07, + "loss": 0.6189, + "step": 7069 + }, + { + "epoch": 13.092592592592593, + "grad_norm": 1.4409118852189011, + "learning_rate": 1.57111734885371e-07, + "loss": 0.6166, + "step": 7070 + }, + { + "epoch": 13.094444444444445, + "grad_norm": 2.323020085454377, + "learning_rate": 1.5709942594720918e-07, + "loss": 0.5975, + "step": 7071 + }, + { + "epoch": 13.096296296296297, + "grad_norm": 1.7822081002871646, + "learning_rate": 1.570871157252972e-07, + "loss": 0.6218, + "step": 7072 + }, + { + "epoch": 13.098148148148148, + "grad_norm": 1.4554626851309314, + "learning_rate": 1.5707480421991183e-07, + "loss": 0.6137, + "step": 7073 + }, + { + "epoch": 13.1, + "grad_norm": 1.8352518797954986, + "learning_rate": 1.570624914313298e-07, + "loss": 0.6119, + "step": 7074 + }, + { + "epoch": 13.101851851851851, + "grad_norm": 1.5115228936448304, + "learning_rate": 1.5705017735982801e-07, + "loss": 0.6396, + "step": 7075 + }, + { + "epoch": 13.103703703703705, + "grad_norm": 1.5479924722574259, + "learning_rate": 1.570378620056833e-07, + "loss": 0.6371, + "step": 7076 + }, + { + "epoch": 13.105555555555556, + "grad_norm": 1.7283385592692195, + "learning_rate": 1.5702554536917254e-07, + "loss": 0.6128, + "step": 7077 + }, + { + "epoch": 13.107407407407408, + "grad_norm": 1.5073843938268008, + "learning_rate": 1.5701322745057262e-07, + "loss": 0.6133, + "step": 7078 + }, + { + "epoch": 13.10925925925926, + "grad_norm": 2.171924463991556, + "learning_rate": 1.5700090825016051e-07, + "loss": 0.6039, + "step": 7079 + }, + { + "epoch": 13.11111111111111, + "grad_norm": 1.4858684453635949, + "learning_rate": 1.5698858776821318e-07, + "loss": 0.6126, + "step": 7080 + }, + { + "epoch": 13.112962962962962, + "grad_norm": 1.5847552179941182, + "learning_rate": 1.569762660050076e-07, + "loss": 0.6034, + "step": 7081 + }, + { + "epoch": 13.114814814814816, + "grad_norm": 1.556310345714512, + "learning_rate": 1.5696394296082084e-07, + "loss": 0.5952, + "step": 7082 + }, + { + "epoch": 13.116666666666667, + "grad_norm": 1.4876415424146252, + "learning_rate": 1.5695161863592992e-07, + "loss": 0.6369, + "step": 7083 + }, + { + "epoch": 13.118518518518519, + "grad_norm": 1.8342770872925007, + "learning_rate": 1.5693929303061193e-07, + "loss": 0.6302, + "step": 7084 + }, + { + "epoch": 13.12037037037037, + "grad_norm": 1.4306705325041016, + "learning_rate": 1.5692696614514401e-07, + "loss": 0.616, + "step": 7085 + }, + { + "epoch": 13.122222222222222, + "grad_norm": 1.4972129471710074, + "learning_rate": 1.5691463797980325e-07, + "loss": 0.592, + "step": 7086 + }, + { + "epoch": 13.124074074074073, + "grad_norm": 1.714298186085463, + "learning_rate": 1.5690230853486686e-07, + "loss": 0.5892, + "step": 7087 + }, + { + "epoch": 13.125925925925927, + "grad_norm": 1.6004699226733163, + "learning_rate": 1.56889977810612e-07, + "loss": 0.6446, + "step": 7088 + }, + { + "epoch": 13.127777777777778, + "grad_norm": 1.6330351856888123, + "learning_rate": 1.5687764580731596e-07, + "loss": 0.6599, + "step": 7089 + }, + { + "epoch": 13.12962962962963, + "grad_norm": 1.650666070105923, + "learning_rate": 1.5686531252525597e-07, + "loss": 0.594, + "step": 7090 + }, + { + "epoch": 13.131481481481481, + "grad_norm": 1.8705760231052602, + "learning_rate": 1.5685297796470933e-07, + "loss": 0.6452, + "step": 7091 + }, + { + "epoch": 13.133333333333333, + "grad_norm": 1.7710449592802031, + "learning_rate": 1.568406421259533e-07, + "loss": 0.626, + "step": 7092 + }, + { + "epoch": 13.135185185185184, + "grad_norm": 1.6900422810429627, + "learning_rate": 1.5682830500926527e-07, + "loss": 0.6404, + "step": 7093 + }, + { + "epoch": 13.137037037037038, + "grad_norm": 2.1407381259962954, + "learning_rate": 1.568159666149226e-07, + "loss": 0.6206, + "step": 7094 + }, + { + "epoch": 13.13888888888889, + "grad_norm": 2.4730224236064835, + "learning_rate": 1.5680362694320266e-07, + "loss": 0.6267, + "step": 7095 + }, + { + "epoch": 13.14074074074074, + "grad_norm": 1.4709880177963799, + "learning_rate": 1.5679128599438295e-07, + "loss": 0.613, + "step": 7096 + }, + { + "epoch": 13.142592592592592, + "grad_norm": 1.4658075830738226, + "learning_rate": 1.5677894376874088e-07, + "loss": 0.5994, + "step": 7097 + }, + { + "epoch": 13.144444444444444, + "grad_norm": 1.5557579315422567, + "learning_rate": 1.5676660026655393e-07, + "loss": 0.5647, + "step": 7098 + }, + { + "epoch": 13.146296296296295, + "grad_norm": 1.4104751968510527, + "learning_rate": 1.5675425548809964e-07, + "loss": 0.5821, + "step": 7099 + }, + { + "epoch": 13.148148148148149, + "grad_norm": 2.4936061108386025, + "learning_rate": 1.5674190943365554e-07, + "loss": 0.6075, + "step": 7100 + }, + { + "epoch": 13.15, + "grad_norm": 1.5967834745204277, + "learning_rate": 1.5672956210349922e-07, + "loss": 0.6263, + "step": 7101 + }, + { + "epoch": 13.151851851851852, + "grad_norm": 1.4593158302998328, + "learning_rate": 1.5671721349790827e-07, + "loss": 0.6097, + "step": 7102 + }, + { + "epoch": 13.153703703703703, + "grad_norm": 1.9576601989944968, + "learning_rate": 1.567048636171603e-07, + "loss": 0.6016, + "step": 7103 + }, + { + "epoch": 13.155555555555555, + "grad_norm": 1.5141568945366077, + "learning_rate": 1.56692512461533e-07, + "loss": 0.5833, + "step": 7104 + }, + { + "epoch": 13.157407407407407, + "grad_norm": 1.8817099256151053, + "learning_rate": 1.5668016003130403e-07, + "loss": 0.5668, + "step": 7105 + }, + { + "epoch": 13.15925925925926, + "grad_norm": 1.5539965335364627, + "learning_rate": 1.5666780632675116e-07, + "loss": 0.6258, + "step": 7106 + }, + { + "epoch": 13.161111111111111, + "grad_norm": 1.6876452793846377, + "learning_rate": 1.5665545134815205e-07, + "loss": 0.6041, + "step": 7107 + }, + { + "epoch": 13.162962962962963, + "grad_norm": 1.94022983646296, + "learning_rate": 1.5664309509578456e-07, + "loss": 0.6058, + "step": 7108 + }, + { + "epoch": 13.164814814814815, + "grad_norm": 1.548248407237479, + "learning_rate": 1.5663073756992643e-07, + "loss": 0.6167, + "step": 7109 + }, + { + "epoch": 13.166666666666666, + "grad_norm": 1.6352882254801526, + "learning_rate": 1.566183787708555e-07, + "loss": 0.6307, + "step": 7110 + }, + { + "epoch": 13.168518518518518, + "grad_norm": 2.2855313650822566, + "learning_rate": 1.5660601869884967e-07, + "loss": 0.6218, + "step": 7111 + }, + { + "epoch": 13.170370370370371, + "grad_norm": 1.3847395082687353, + "learning_rate": 1.5659365735418678e-07, + "loss": 0.6219, + "step": 7112 + }, + { + "epoch": 13.172222222222222, + "grad_norm": 1.6927535566155099, + "learning_rate": 1.5658129473714477e-07, + "loss": 0.6325, + "step": 7113 + }, + { + "epoch": 13.174074074074074, + "grad_norm": 1.7956571728487167, + "learning_rate": 1.565689308480016e-07, + "loss": 0.5942, + "step": 7114 + }, + { + "epoch": 13.175925925925926, + "grad_norm": 1.7541327153587984, + "learning_rate": 1.5655656568703523e-07, + "loss": 0.6522, + "step": 7115 + }, + { + "epoch": 13.177777777777777, + "grad_norm": 1.4885519507175027, + "learning_rate": 1.565441992545236e-07, + "loss": 0.5946, + "step": 7116 + }, + { + "epoch": 13.17962962962963, + "grad_norm": 1.5915189221090702, + "learning_rate": 1.5653183155074486e-07, + "loss": 0.6077, + "step": 7117 + }, + { + "epoch": 13.181481481481482, + "grad_norm": 2.502650475644679, + "learning_rate": 1.5651946257597696e-07, + "loss": 0.6433, + "step": 7118 + }, + { + "epoch": 13.183333333333334, + "grad_norm": 1.376787567560288, + "learning_rate": 1.5650709233049807e-07, + "loss": 0.6266, + "step": 7119 + }, + { + "epoch": 13.185185185185185, + "grad_norm": 1.518433654511087, + "learning_rate": 1.5649472081458628e-07, + "loss": 0.632, + "step": 7120 + }, + { + "epoch": 13.187037037037037, + "grad_norm": 1.9058944420754775, + "learning_rate": 1.564823480285197e-07, + "loss": 0.5861, + "step": 7121 + }, + { + "epoch": 13.188888888888888, + "grad_norm": 1.5878063196337242, + "learning_rate": 1.5646997397257655e-07, + "loss": 0.6206, + "step": 7122 + }, + { + "epoch": 13.190740740740742, + "grad_norm": 1.403100950652496, + "learning_rate": 1.5645759864703503e-07, + "loss": 0.6282, + "step": 7123 + }, + { + "epoch": 13.192592592592593, + "grad_norm": 1.4179022264877932, + "learning_rate": 1.5644522205217335e-07, + "loss": 0.6348, + "step": 7124 + }, + { + "epoch": 13.194444444444445, + "grad_norm": 1.490665425752897, + "learning_rate": 1.5643284418826975e-07, + "loss": 0.6042, + "step": 7125 + }, + { + "epoch": 13.196296296296296, + "grad_norm": 1.5270277794993898, + "learning_rate": 1.5642046505560256e-07, + "loss": 0.5997, + "step": 7126 + }, + { + "epoch": 13.198148148148148, + "grad_norm": 1.494866735653359, + "learning_rate": 1.5640808465445007e-07, + "loss": 0.569, + "step": 7127 + }, + { + "epoch": 13.2, + "grad_norm": 1.7168862780206393, + "learning_rate": 1.5639570298509063e-07, + "loss": 0.6054, + "step": 7128 + }, + { + "epoch": 13.201851851851853, + "grad_norm": 2.8516847300923267, + "learning_rate": 1.5638332004780262e-07, + "loss": 0.6486, + "step": 7129 + }, + { + "epoch": 13.203703703703704, + "grad_norm": 1.5282436234024466, + "learning_rate": 1.5637093584286448e-07, + "loss": 0.5545, + "step": 7130 + }, + { + "epoch": 13.205555555555556, + "grad_norm": 1.4862224379288491, + "learning_rate": 1.5635855037055459e-07, + "loss": 0.588, + "step": 7131 + }, + { + "epoch": 13.207407407407407, + "grad_norm": 1.4233588105279789, + "learning_rate": 1.5634616363115138e-07, + "loss": 0.6393, + "step": 7132 + }, + { + "epoch": 13.209259259259259, + "grad_norm": 1.8378596363917523, + "learning_rate": 1.5633377562493342e-07, + "loss": 0.6219, + "step": 7133 + }, + { + "epoch": 13.21111111111111, + "grad_norm": 1.5898260549190242, + "learning_rate": 1.5632138635217916e-07, + "loss": 0.6304, + "step": 7134 + }, + { + "epoch": 13.212962962962964, + "grad_norm": 2.012889751994183, + "learning_rate": 1.5630899581316718e-07, + "loss": 0.6446, + "step": 7135 + }, + { + "epoch": 13.214814814814815, + "grad_norm": 3.2732604266456153, + "learning_rate": 1.5629660400817604e-07, + "loss": 0.5904, + "step": 7136 + }, + { + "epoch": 13.216666666666667, + "grad_norm": 1.5106587828841185, + "learning_rate": 1.562842109374843e-07, + "loss": 0.643, + "step": 7137 + }, + { + "epoch": 13.218518518518518, + "grad_norm": 2.0836126144573575, + "learning_rate": 1.562718166013707e-07, + "loss": 0.5976, + "step": 7138 + }, + { + "epoch": 13.22037037037037, + "grad_norm": 1.5558278498467972, + "learning_rate": 1.562594210001138e-07, + "loss": 0.631, + "step": 7139 + }, + { + "epoch": 13.222222222222221, + "grad_norm": 1.7223859995467241, + "learning_rate": 1.562470241339923e-07, + "loss": 0.6146, + "step": 7140 + }, + { + "epoch": 13.224074074074075, + "grad_norm": 1.5080277195184704, + "learning_rate": 1.5623462600328497e-07, + "loss": 0.5845, + "step": 7141 + }, + { + "epoch": 13.225925925925926, + "grad_norm": 1.5007770968310816, + "learning_rate": 1.5622222660827049e-07, + "loss": 0.6359, + "step": 7142 + }, + { + "epoch": 13.227777777777778, + "grad_norm": 1.4711528896239094, + "learning_rate": 1.5620982594922768e-07, + "loss": 0.6069, + "step": 7143 + }, + { + "epoch": 13.22962962962963, + "grad_norm": 1.4682487332492835, + "learning_rate": 1.561974240264353e-07, + "loss": 0.6015, + "step": 7144 + }, + { + "epoch": 13.231481481481481, + "grad_norm": 1.5438506406875174, + "learning_rate": 1.5618502084017218e-07, + "loss": 0.6155, + "step": 7145 + }, + { + "epoch": 13.233333333333333, + "grad_norm": 1.8011109053017722, + "learning_rate": 1.5617261639071723e-07, + "loss": 0.6285, + "step": 7146 + }, + { + "epoch": 13.235185185185186, + "grad_norm": 1.5024381589216425, + "learning_rate": 1.5616021067834928e-07, + "loss": 0.5624, + "step": 7147 + }, + { + "epoch": 13.237037037037037, + "grad_norm": 1.6527681664251175, + "learning_rate": 1.5614780370334728e-07, + "loss": 0.5806, + "step": 7148 + }, + { + "epoch": 13.238888888888889, + "grad_norm": 1.4644161808606149, + "learning_rate": 1.5613539546599014e-07, + "loss": 0.566, + "step": 7149 + }, + { + "epoch": 13.24074074074074, + "grad_norm": 1.9885807005798977, + "learning_rate": 1.561229859665569e-07, + "loss": 0.6165, + "step": 7150 + }, + { + "epoch": 13.242592592592592, + "grad_norm": 1.7490679004657346, + "learning_rate": 1.5611057520532643e-07, + "loss": 0.5784, + "step": 7151 + }, + { + "epoch": 13.244444444444444, + "grad_norm": 1.974566314147183, + "learning_rate": 1.5609816318257788e-07, + "loss": 0.6377, + "step": 7152 + }, + { + "epoch": 13.246296296296297, + "grad_norm": 1.844730433219054, + "learning_rate": 1.5608574989859026e-07, + "loss": 0.6257, + "step": 7153 + }, + { + "epoch": 13.248148148148148, + "grad_norm": 2.085746332903787, + "learning_rate": 1.5607333535364266e-07, + "loss": 0.6051, + "step": 7154 + }, + { + "epoch": 13.25, + "grad_norm": 1.6632466478469985, + "learning_rate": 1.5606091954801418e-07, + "loss": 0.6094, + "step": 7155 + }, + { + "epoch": 13.251851851851852, + "grad_norm": 1.4712284219704022, + "learning_rate": 1.5604850248198395e-07, + "loss": 0.6055, + "step": 7156 + }, + { + "epoch": 13.253703703703703, + "grad_norm": 1.496918293587102, + "learning_rate": 1.560360841558312e-07, + "loss": 0.6155, + "step": 7157 + }, + { + "epoch": 13.255555555555556, + "grad_norm": 1.548264825894146, + "learning_rate": 1.5602366456983506e-07, + "loss": 0.6222, + "step": 7158 + }, + { + "epoch": 13.257407407407408, + "grad_norm": 1.3363685501441196, + "learning_rate": 1.5601124372427478e-07, + "loss": 0.6055, + "step": 7159 + }, + { + "epoch": 13.25925925925926, + "grad_norm": 1.4565240443840588, + "learning_rate": 1.5599882161942964e-07, + "loss": 0.6502, + "step": 7160 + }, + { + "epoch": 13.261111111111111, + "grad_norm": 1.6176817858151131, + "learning_rate": 1.5598639825557887e-07, + "loss": 0.6142, + "step": 7161 + }, + { + "epoch": 13.262962962962963, + "grad_norm": 1.5527666367366177, + "learning_rate": 1.5597397363300183e-07, + "loss": 0.5862, + "step": 7162 + }, + { + "epoch": 13.264814814814814, + "grad_norm": 1.513823629471141, + "learning_rate": 1.5596154775197783e-07, + "loss": 0.6125, + "step": 7163 + }, + { + "epoch": 13.266666666666667, + "grad_norm": 2.3678619192155796, + "learning_rate": 1.5594912061278625e-07, + "loss": 0.6055, + "step": 7164 + }, + { + "epoch": 13.268518518518519, + "grad_norm": 1.6877413243806767, + "learning_rate": 1.5593669221570647e-07, + "loss": 0.6099, + "step": 7165 + }, + { + "epoch": 13.27037037037037, + "grad_norm": 1.5670393451826474, + "learning_rate": 1.5592426256101794e-07, + "loss": 0.5953, + "step": 7166 + }, + { + "epoch": 13.272222222222222, + "grad_norm": 1.637019713254645, + "learning_rate": 1.559118316490001e-07, + "loss": 0.6245, + "step": 7167 + }, + { + "epoch": 13.274074074074074, + "grad_norm": 1.4936109091676781, + "learning_rate": 1.5589939947993244e-07, + "loss": 0.6231, + "step": 7168 + }, + { + "epoch": 13.275925925925925, + "grad_norm": 1.516841577638341, + "learning_rate": 1.5588696605409444e-07, + "loss": 0.6186, + "step": 7169 + }, + { + "epoch": 13.277777777777779, + "grad_norm": 1.5601332465833169, + "learning_rate": 1.5587453137176568e-07, + "loss": 0.6262, + "step": 7170 + }, + { + "epoch": 13.27962962962963, + "grad_norm": 2.136025791024716, + "learning_rate": 1.558620954332257e-07, + "loss": 0.5759, + "step": 7171 + }, + { + "epoch": 13.281481481481482, + "grad_norm": 1.5916842080098519, + "learning_rate": 1.5584965823875406e-07, + "loss": 0.6032, + "step": 7172 + }, + { + "epoch": 13.283333333333333, + "grad_norm": 1.5822974119647308, + "learning_rate": 1.5583721978863044e-07, + "loss": 0.6148, + "step": 7173 + }, + { + "epoch": 13.285185185185185, + "grad_norm": 1.3879963852549086, + "learning_rate": 1.558247800831345e-07, + "loss": 0.6115, + "step": 7174 + }, + { + "epoch": 13.287037037037036, + "grad_norm": 1.7031413341957036, + "learning_rate": 1.5581233912254582e-07, + "loss": 0.6403, + "step": 7175 + }, + { + "epoch": 13.28888888888889, + "grad_norm": 1.4872108702002669, + "learning_rate": 1.557998969071442e-07, + "loss": 0.5937, + "step": 7176 + }, + { + "epoch": 13.290740740740741, + "grad_norm": 1.5827488685868027, + "learning_rate": 1.557874534372094e-07, + "loss": 0.6261, + "step": 7177 + }, + { + "epoch": 13.292592592592593, + "grad_norm": 1.4314220413011183, + "learning_rate": 1.5577500871302104e-07, + "loss": 0.5681, + "step": 7178 + }, + { + "epoch": 13.294444444444444, + "grad_norm": 1.5904591622898008, + "learning_rate": 1.5576256273485907e-07, + "loss": 0.5818, + "step": 7179 + }, + { + "epoch": 13.296296296296296, + "grad_norm": 2.2518748080039392, + "learning_rate": 1.557501155030032e-07, + "loss": 0.6015, + "step": 7180 + }, + { + "epoch": 13.298148148148147, + "grad_norm": 2.1293974731108634, + "learning_rate": 1.5573766701773335e-07, + "loss": 0.6023, + "step": 7181 + }, + { + "epoch": 13.3, + "grad_norm": 1.4576436345808617, + "learning_rate": 1.5572521727932936e-07, + "loss": 0.6093, + "step": 7182 + }, + { + "epoch": 13.301851851851852, + "grad_norm": 2.805747573234161, + "learning_rate": 1.5571276628807113e-07, + "loss": 0.615, + "step": 7183 + }, + { + "epoch": 13.303703703703704, + "grad_norm": 1.5033622270393323, + "learning_rate": 1.5570031404423857e-07, + "loss": 0.5731, + "step": 7184 + }, + { + "epoch": 13.305555555555555, + "grad_norm": 1.688349133988544, + "learning_rate": 1.5568786054811172e-07, + "loss": 0.6441, + "step": 7185 + }, + { + "epoch": 13.307407407407407, + "grad_norm": 1.476820335603725, + "learning_rate": 1.5567540579997048e-07, + "loss": 0.6229, + "step": 7186 + }, + { + "epoch": 13.309259259259258, + "grad_norm": 1.7225069094992602, + "learning_rate": 1.5566294980009494e-07, + "loss": 0.6151, + "step": 7187 + }, + { + "epoch": 13.311111111111112, + "grad_norm": 2.1389717721659394, + "learning_rate": 1.5565049254876507e-07, + "loss": 0.6297, + "step": 7188 + }, + { + "epoch": 13.312962962962963, + "grad_norm": 1.6711063435386788, + "learning_rate": 1.5563803404626105e-07, + "loss": 0.6661, + "step": 7189 + }, + { + "epoch": 13.314814814814815, + "grad_norm": 1.7705777056635927, + "learning_rate": 1.5562557429286284e-07, + "loss": 0.6033, + "step": 7190 + }, + { + "epoch": 13.316666666666666, + "grad_norm": 1.667030791659813, + "learning_rate": 1.556131132888507e-07, + "loss": 0.6139, + "step": 7191 + }, + { + "epoch": 13.318518518518518, + "grad_norm": 1.5127777248251448, + "learning_rate": 1.556006510345047e-07, + "loss": 0.6473, + "step": 7192 + }, + { + "epoch": 13.32037037037037, + "grad_norm": 1.6200711571689175, + "learning_rate": 1.5558818753010503e-07, + "loss": 0.6363, + "step": 7193 + }, + { + "epoch": 13.322222222222223, + "grad_norm": 1.8210418813816627, + "learning_rate": 1.5557572277593196e-07, + "loss": 0.5975, + "step": 7194 + }, + { + "epoch": 13.324074074074074, + "grad_norm": 1.6835784244307808, + "learning_rate": 1.5556325677226572e-07, + "loss": 0.6308, + "step": 7195 + }, + { + "epoch": 13.325925925925926, + "grad_norm": 1.6793993339238522, + "learning_rate": 1.555507895193865e-07, + "loss": 0.6399, + "step": 7196 + }, + { + "epoch": 13.327777777777778, + "grad_norm": 1.4709646184328942, + "learning_rate": 1.5553832101757474e-07, + "loss": 0.5732, + "step": 7197 + }, + { + "epoch": 13.329629629629629, + "grad_norm": 1.6320202024501513, + "learning_rate": 1.5552585126711065e-07, + "loss": 0.6282, + "step": 7198 + }, + { + "epoch": 13.331481481481482, + "grad_norm": 1.6373435069553492, + "learning_rate": 1.5551338026827458e-07, + "loss": 0.5896, + "step": 7199 + }, + { + "epoch": 13.333333333333334, + "grad_norm": 1.4650827802238808, + "learning_rate": 1.55500908021347e-07, + "loss": 0.6318, + "step": 7200 + }, + { + "epoch": 13.335185185185185, + "grad_norm": 1.5513335421585195, + "learning_rate": 1.5548843452660823e-07, + "loss": 0.6044, + "step": 7201 + }, + { + "epoch": 13.337037037037037, + "grad_norm": 1.5297668390045105, + "learning_rate": 1.554759597843388e-07, + "loss": 0.5951, + "step": 7202 + }, + { + "epoch": 13.338888888888889, + "grad_norm": 1.9096775338258725, + "learning_rate": 1.5546348379481906e-07, + "loss": 0.6609, + "step": 7203 + }, + { + "epoch": 13.34074074074074, + "grad_norm": 3.274762102893177, + "learning_rate": 1.554510065583296e-07, + "loss": 0.6278, + "step": 7204 + }, + { + "epoch": 13.342592592592593, + "grad_norm": 1.4427923949228452, + "learning_rate": 1.554385280751509e-07, + "loss": 0.626, + "step": 7205 + }, + { + "epoch": 13.344444444444445, + "grad_norm": 1.6533534661881462, + "learning_rate": 1.5542604834556352e-07, + "loss": 0.6243, + "step": 7206 + }, + { + "epoch": 13.346296296296297, + "grad_norm": 1.592967389770644, + "learning_rate": 1.5541356736984805e-07, + "loss": 0.6168, + "step": 7207 + }, + { + "epoch": 13.348148148148148, + "grad_norm": 1.5639555245975274, + "learning_rate": 1.5540108514828508e-07, + "loss": 0.6152, + "step": 7208 + }, + { + "epoch": 13.35, + "grad_norm": 1.853515960216099, + "learning_rate": 1.5538860168115525e-07, + "loss": 0.6179, + "step": 7209 + }, + { + "epoch": 13.351851851851851, + "grad_norm": 1.4444427657327827, + "learning_rate": 1.5537611696873922e-07, + "loss": 0.6189, + "step": 7210 + }, + { + "epoch": 13.353703703703705, + "grad_norm": 1.5211776580755323, + "learning_rate": 1.5536363101131768e-07, + "loss": 0.6341, + "step": 7211 + }, + { + "epoch": 13.355555555555556, + "grad_norm": 2.048897242291352, + "learning_rate": 1.5535114380917136e-07, + "loss": 0.5821, + "step": 7212 + }, + { + "epoch": 13.357407407407408, + "grad_norm": 1.8078325145819816, + "learning_rate": 1.5533865536258098e-07, + "loss": 0.5753, + "step": 7213 + }, + { + "epoch": 13.35925925925926, + "grad_norm": 4.399871436313289, + "learning_rate": 1.5532616567182735e-07, + "loss": 0.6235, + "step": 7214 + }, + { + "epoch": 13.36111111111111, + "grad_norm": 1.4027593518509, + "learning_rate": 1.5531367473719122e-07, + "loss": 0.6343, + "step": 7215 + }, + { + "epoch": 13.362962962962962, + "grad_norm": 1.5334102322901992, + "learning_rate": 1.553011825589535e-07, + "loss": 0.5927, + "step": 7216 + }, + { + "epoch": 13.364814814814816, + "grad_norm": 2.885473030185685, + "learning_rate": 1.5528868913739496e-07, + "loss": 0.6401, + "step": 7217 + }, + { + "epoch": 13.366666666666667, + "grad_norm": 1.6874424494807094, + "learning_rate": 1.5527619447279654e-07, + "loss": 0.6482, + "step": 7218 + }, + { + "epoch": 13.368518518518519, + "grad_norm": 1.676149172295434, + "learning_rate": 1.5526369856543912e-07, + "loss": 0.6299, + "step": 7219 + }, + { + "epoch": 13.37037037037037, + "grad_norm": 1.5605094292624055, + "learning_rate": 1.5525120141560368e-07, + "loss": 0.5801, + "step": 7220 + }, + { + "epoch": 13.372222222222222, + "grad_norm": 1.7229455207529392, + "learning_rate": 1.552387030235712e-07, + "loss": 0.613, + "step": 7221 + }, + { + "epoch": 13.374074074074073, + "grad_norm": 2.5048857770039175, + "learning_rate": 1.552262033896226e-07, + "loss": 0.6439, + "step": 7222 + }, + { + "epoch": 13.375925925925927, + "grad_norm": 1.5525856630085684, + "learning_rate": 1.5521370251403897e-07, + "loss": 0.6456, + "step": 7223 + }, + { + "epoch": 13.377777777777778, + "grad_norm": 1.450106358722777, + "learning_rate": 1.5520120039710135e-07, + "loss": 0.6366, + "step": 7224 + }, + { + "epoch": 13.37962962962963, + "grad_norm": 1.5866116806712873, + "learning_rate": 1.5518869703909085e-07, + "loss": 0.6531, + "step": 7225 + }, + { + "epoch": 13.381481481481481, + "grad_norm": 1.39321956578684, + "learning_rate": 1.5517619244028851e-07, + "loss": 0.6266, + "step": 7226 + }, + { + "epoch": 13.383333333333333, + "grad_norm": 1.7032694536662418, + "learning_rate": 1.5516368660097554e-07, + "loss": 0.6059, + "step": 7227 + }, + { + "epoch": 13.385185185185184, + "grad_norm": 1.719515595469275, + "learning_rate": 1.5515117952143306e-07, + "loss": 0.6041, + "step": 7228 + }, + { + "epoch": 13.387037037037038, + "grad_norm": 1.4816846150279102, + "learning_rate": 1.5513867120194228e-07, + "loss": 0.603, + "step": 7229 + }, + { + "epoch": 13.38888888888889, + "grad_norm": 1.6810960457715094, + "learning_rate": 1.551261616427844e-07, + "loss": 0.5998, + "step": 7230 + }, + { + "epoch": 13.39074074074074, + "grad_norm": 1.460274833542817, + "learning_rate": 1.551136508442407e-07, + "loss": 0.5875, + "step": 7231 + }, + { + "epoch": 13.392592592592592, + "grad_norm": 1.5976815313291926, + "learning_rate": 1.5510113880659245e-07, + "loss": 0.6369, + "step": 7232 + }, + { + "epoch": 13.394444444444444, + "grad_norm": 1.6637518168805452, + "learning_rate": 1.5508862553012095e-07, + "loss": 0.6443, + "step": 7233 + }, + { + "epoch": 13.396296296296295, + "grad_norm": 1.584093813551479, + "learning_rate": 1.550761110151075e-07, + "loss": 0.5938, + "step": 7234 + }, + { + "epoch": 13.398148148148149, + "grad_norm": 1.3902828841954127, + "learning_rate": 1.550635952618335e-07, + "loss": 0.5905, + "step": 7235 + }, + { + "epoch": 13.4, + "grad_norm": 1.6112465683726291, + "learning_rate": 1.5505107827058036e-07, + "loss": 0.611, + "step": 7236 + }, + { + "epoch": 13.401851851851852, + "grad_norm": 1.5427502555257506, + "learning_rate": 1.5503856004162943e-07, + "loss": 0.6195, + "step": 7237 + }, + { + "epoch": 13.403703703703703, + "grad_norm": 2.1613888322822903, + "learning_rate": 1.5502604057526218e-07, + "loss": 0.5967, + "step": 7238 + }, + { + "epoch": 13.405555555555555, + "grad_norm": 1.4482183321569075, + "learning_rate": 1.5501351987176013e-07, + "loss": 0.6236, + "step": 7239 + }, + { + "epoch": 13.407407407407407, + "grad_norm": 1.7071839678327017, + "learning_rate": 1.5500099793140472e-07, + "loss": 0.6317, + "step": 7240 + }, + { + "epoch": 13.40925925925926, + "grad_norm": 1.4328849031745656, + "learning_rate": 1.5498847475447748e-07, + "loss": 0.556, + "step": 7241 + }, + { + "epoch": 13.411111111111111, + "grad_norm": 2.0680950725964804, + "learning_rate": 1.5497595034126e-07, + "loss": 0.6157, + "step": 7242 + }, + { + "epoch": 13.412962962962963, + "grad_norm": 1.4340663397466076, + "learning_rate": 1.5496342469203382e-07, + "loss": 0.5666, + "step": 7243 + }, + { + "epoch": 13.414814814814815, + "grad_norm": 2.2323254468394285, + "learning_rate": 1.549508978070806e-07, + "loss": 0.5856, + "step": 7244 + }, + { + "epoch": 13.416666666666666, + "grad_norm": 1.5738682735443401, + "learning_rate": 1.5493836968668196e-07, + "loss": 0.6544, + "step": 7245 + }, + { + "epoch": 13.418518518518518, + "grad_norm": 3.4473329389301184, + "learning_rate": 1.5492584033111952e-07, + "loss": 0.5675, + "step": 7246 + }, + { + "epoch": 13.420370370370371, + "grad_norm": 1.7431074658088808, + "learning_rate": 1.5491330974067506e-07, + "loss": 0.6204, + "step": 7247 + }, + { + "epoch": 13.422222222222222, + "grad_norm": 1.7102770591626848, + "learning_rate": 1.549007779156302e-07, + "loss": 0.6341, + "step": 7248 + }, + { + "epoch": 13.424074074074074, + "grad_norm": 1.4716501837398548, + "learning_rate": 1.548882448562668e-07, + "loss": 0.5965, + "step": 7249 + }, + { + "epoch": 13.425925925925926, + "grad_norm": 1.6122510352119186, + "learning_rate": 1.5487571056286655e-07, + "loss": 0.6262, + "step": 7250 + }, + { + "epoch": 13.427777777777777, + "grad_norm": 1.5630625688884, + "learning_rate": 1.5486317503571128e-07, + "loss": 0.6029, + "step": 7251 + }, + { + "epoch": 13.42962962962963, + "grad_norm": 1.7753991485483485, + "learning_rate": 1.5485063827508284e-07, + "loss": 0.6419, + "step": 7252 + }, + { + "epoch": 13.431481481481482, + "grad_norm": 1.5111736578440107, + "learning_rate": 1.548381002812631e-07, + "loss": 0.6219, + "step": 7253 + }, + { + "epoch": 13.433333333333334, + "grad_norm": 1.4277885822816156, + "learning_rate": 1.548255610545339e-07, + "loss": 0.5958, + "step": 7254 + }, + { + "epoch": 13.435185185185185, + "grad_norm": 1.659906211316063, + "learning_rate": 1.548130205951772e-07, + "loss": 0.6223, + "step": 7255 + }, + { + "epoch": 13.437037037037037, + "grad_norm": 1.4701232226134355, + "learning_rate": 1.548004789034749e-07, + "loss": 0.6083, + "step": 7256 + }, + { + "epoch": 13.438888888888888, + "grad_norm": 2.099838260800555, + "learning_rate": 1.5478793597970904e-07, + "loss": 0.6277, + "step": 7257 + }, + { + "epoch": 13.440740740740742, + "grad_norm": 1.397297940851872, + "learning_rate": 1.5477539182416155e-07, + "loss": 0.6081, + "step": 7258 + }, + { + "epoch": 13.442592592592593, + "grad_norm": 1.6824252374009836, + "learning_rate": 1.547628464371145e-07, + "loss": 0.6071, + "step": 7259 + }, + { + "epoch": 13.444444444444445, + "grad_norm": 1.6801908021288223, + "learning_rate": 1.5475029981884992e-07, + "loss": 0.6226, + "step": 7260 + }, + { + "epoch": 13.446296296296296, + "grad_norm": 1.693111036716262, + "learning_rate": 1.5473775196964988e-07, + "loss": 0.649, + "step": 7261 + }, + { + "epoch": 13.448148148148148, + "grad_norm": 1.5285946865132818, + "learning_rate": 1.5472520288979654e-07, + "loss": 0.6693, + "step": 7262 + }, + { + "epoch": 13.45, + "grad_norm": 1.6776202118672494, + "learning_rate": 1.54712652579572e-07, + "loss": 0.5978, + "step": 7263 + }, + { + "epoch": 13.451851851851853, + "grad_norm": 1.4903361181878623, + "learning_rate": 1.5470010103925841e-07, + "loss": 0.554, + "step": 7264 + }, + { + "epoch": 13.453703703703704, + "grad_norm": 1.6270620847033141, + "learning_rate": 1.54687548269138e-07, + "loss": 0.6114, + "step": 7265 + }, + { + "epoch": 13.455555555555556, + "grad_norm": 1.450077681166889, + "learning_rate": 1.5467499426949301e-07, + "loss": 0.6432, + "step": 7266 + }, + { + "epoch": 13.457407407407407, + "grad_norm": 2.0873861628103962, + "learning_rate": 1.546624390406056e-07, + "loss": 0.6107, + "step": 7267 + }, + { + "epoch": 13.459259259259259, + "grad_norm": 2.0368006397081495, + "learning_rate": 1.5464988258275816e-07, + "loss": 0.6325, + "step": 7268 + }, + { + "epoch": 13.46111111111111, + "grad_norm": 1.5276026367458988, + "learning_rate": 1.546373248962329e-07, + "loss": 0.5599, + "step": 7269 + }, + { + "epoch": 13.462962962962964, + "grad_norm": 1.9163286826330306, + "learning_rate": 1.5462476598131215e-07, + "loss": 0.6354, + "step": 7270 + }, + { + "epoch": 13.464814814814815, + "grad_norm": 2.8722769750608483, + "learning_rate": 1.5461220583827834e-07, + "loss": 0.6556, + "step": 7271 + }, + { + "epoch": 13.466666666666667, + "grad_norm": 1.6272131929280682, + "learning_rate": 1.545996444674138e-07, + "loss": 0.6361, + "step": 7272 + }, + { + "epoch": 13.468518518518518, + "grad_norm": 3.1862731656053076, + "learning_rate": 1.5458708186900097e-07, + "loss": 0.6238, + "step": 7273 + }, + { + "epoch": 13.47037037037037, + "grad_norm": 1.5230550313486333, + "learning_rate": 1.545745180433223e-07, + "loss": 0.6548, + "step": 7274 + }, + { + "epoch": 13.472222222222221, + "grad_norm": 1.5796975329947691, + "learning_rate": 1.5456195299066024e-07, + "loss": 0.6761, + "step": 7275 + }, + { + "epoch": 13.474074074074075, + "grad_norm": 2.2177988597927794, + "learning_rate": 1.5454938671129725e-07, + "loss": 0.5899, + "step": 7276 + }, + { + "epoch": 13.475925925925926, + "grad_norm": 2.9647813476820613, + "learning_rate": 1.5453681920551593e-07, + "loss": 0.5911, + "step": 7277 + }, + { + "epoch": 13.477777777777778, + "grad_norm": 3.747270398907113, + "learning_rate": 1.545242504735988e-07, + "loss": 0.5929, + "step": 7278 + }, + { + "epoch": 13.47962962962963, + "grad_norm": 1.701062681988641, + "learning_rate": 1.545116805158284e-07, + "loss": 0.6342, + "step": 7279 + }, + { + "epoch": 13.481481481481481, + "grad_norm": 1.5296037841240349, + "learning_rate": 1.544991093324874e-07, + "loss": 0.5982, + "step": 7280 + }, + { + "epoch": 13.483333333333333, + "grad_norm": 1.5355962732251203, + "learning_rate": 1.544865369238584e-07, + "loss": 0.5723, + "step": 7281 + }, + { + "epoch": 13.485185185185186, + "grad_norm": 1.4924029441769386, + "learning_rate": 1.5447396329022404e-07, + "loss": 0.5913, + "step": 7282 + }, + { + "epoch": 13.487037037037037, + "grad_norm": 1.7686834754944267, + "learning_rate": 1.5446138843186708e-07, + "loss": 0.6184, + "step": 7283 + }, + { + "epoch": 13.488888888888889, + "grad_norm": 1.609445053178199, + "learning_rate": 1.5444881234907015e-07, + "loss": 0.5756, + "step": 7284 + }, + { + "epoch": 13.49074074074074, + "grad_norm": 1.511023451679778, + "learning_rate": 1.5443623504211604e-07, + "loss": 0.6127, + "step": 7285 + }, + { + "epoch": 13.492592592592592, + "grad_norm": 1.5958671644925277, + "learning_rate": 1.5442365651128757e-07, + "loss": 0.6159, + "step": 7286 + }, + { + "epoch": 13.494444444444444, + "grad_norm": 1.6427870429851574, + "learning_rate": 1.5441107675686746e-07, + "loss": 0.5427, + "step": 7287 + }, + { + "epoch": 13.496296296296297, + "grad_norm": 1.48078786290128, + "learning_rate": 1.5439849577913854e-07, + "loss": 0.6036, + "step": 7288 + }, + { + "epoch": 13.498148148148148, + "grad_norm": 1.536288241700882, + "learning_rate": 1.5438591357838373e-07, + "loss": 0.5611, + "step": 7289 + }, + { + "epoch": 13.5, + "grad_norm": 3.2795814027610666, + "learning_rate": 1.5437333015488586e-07, + "loss": 0.6404, + "step": 7290 + }, + { + "epoch": 13.501851851851852, + "grad_norm": 2.4877275128378393, + "learning_rate": 1.5436074550892782e-07, + "loss": 0.5842, + "step": 7291 + }, + { + "epoch": 13.503703703703703, + "grad_norm": 1.8444215746675836, + "learning_rate": 1.5434815964079265e-07, + "loss": 0.6054, + "step": 7292 + }, + { + "epoch": 13.505555555555556, + "grad_norm": 1.7101993425511546, + "learning_rate": 1.5433557255076316e-07, + "loss": 0.5957, + "step": 7293 + }, + { + "epoch": 13.507407407407408, + "grad_norm": 1.9236807851678925, + "learning_rate": 1.5432298423912248e-07, + "loss": 0.5625, + "step": 7294 + }, + { + "epoch": 13.50925925925926, + "grad_norm": 1.6001900064404895, + "learning_rate": 1.5431039470615356e-07, + "loss": 0.6268, + "step": 7295 + }, + { + "epoch": 13.511111111111111, + "grad_norm": 1.5960329218471694, + "learning_rate": 1.5429780395213946e-07, + "loss": 0.6261, + "step": 7296 + }, + { + "epoch": 13.512962962962963, + "grad_norm": 1.7449133974113504, + "learning_rate": 1.5428521197736327e-07, + "loss": 0.6013, + "step": 7297 + }, + { + "epoch": 13.514814814814814, + "grad_norm": 1.8771426929945274, + "learning_rate": 1.542726187821081e-07, + "loss": 0.6271, + "step": 7298 + }, + { + "epoch": 13.516666666666667, + "grad_norm": 1.663045731213215, + "learning_rate": 1.5426002436665702e-07, + "loss": 0.6337, + "step": 7299 + }, + { + "epoch": 13.518518518518519, + "grad_norm": 1.658408113420754, + "learning_rate": 1.542474287312932e-07, + "loss": 0.596, + "step": 7300 + }, + { + "epoch": 13.52037037037037, + "grad_norm": 1.5682888823924663, + "learning_rate": 1.5423483187629992e-07, + "loss": 0.6405, + "step": 7301 + }, + { + "epoch": 13.522222222222222, + "grad_norm": 1.765702104168939, + "learning_rate": 1.5422223380196027e-07, + "loss": 0.6186, + "step": 7302 + }, + { + "epoch": 13.524074074074074, + "grad_norm": 1.8488512607310224, + "learning_rate": 1.5420963450855752e-07, + "loss": 0.62, + "step": 7303 + }, + { + "epoch": 13.525925925925925, + "grad_norm": 1.8507928254643464, + "learning_rate": 1.54197033996375e-07, + "loss": 0.6267, + "step": 7304 + }, + { + "epoch": 13.527777777777779, + "grad_norm": 1.5225311310517957, + "learning_rate": 1.5418443226569595e-07, + "loss": 0.6084, + "step": 7305 + }, + { + "epoch": 13.52962962962963, + "grad_norm": 1.5031336554184112, + "learning_rate": 1.541718293168037e-07, + "loss": 0.5904, + "step": 7306 + }, + { + "epoch": 13.531481481481482, + "grad_norm": 1.678870570414537, + "learning_rate": 1.5415922514998158e-07, + "loss": 0.5988, + "step": 7307 + }, + { + "epoch": 13.533333333333333, + "grad_norm": 1.8478465513422078, + "learning_rate": 1.54146619765513e-07, + "loss": 0.6099, + "step": 7308 + }, + { + "epoch": 13.535185185185185, + "grad_norm": 1.789387149025911, + "learning_rate": 1.5413401316368132e-07, + "loss": 0.6125, + "step": 7309 + }, + { + "epoch": 13.537037037037036, + "grad_norm": 1.6405450700681674, + "learning_rate": 1.5412140534477004e-07, + "loss": 0.6049, + "step": 7310 + }, + { + "epoch": 13.53888888888889, + "grad_norm": 1.5386246610526035, + "learning_rate": 1.5410879630906257e-07, + "loss": 0.6097, + "step": 7311 + }, + { + "epoch": 13.540740740740741, + "grad_norm": 1.9177223826017549, + "learning_rate": 1.540961860568424e-07, + "loss": 0.6289, + "step": 7312 + }, + { + "epoch": 13.542592592592593, + "grad_norm": 1.526017841772827, + "learning_rate": 1.54083574588393e-07, + "loss": 0.6126, + "step": 7313 + }, + { + "epoch": 13.544444444444444, + "grad_norm": 1.4722972119743998, + "learning_rate": 1.5407096190399803e-07, + "loss": 0.6409, + "step": 7314 + }, + { + "epoch": 13.546296296296296, + "grad_norm": 1.6244677154763651, + "learning_rate": 1.540583480039409e-07, + "loss": 0.5764, + "step": 7315 + }, + { + "epoch": 13.548148148148147, + "grad_norm": 1.4614190575251744, + "learning_rate": 1.5404573288850533e-07, + "loss": 0.5818, + "step": 7316 + }, + { + "epoch": 13.55, + "grad_norm": 1.4984861578826052, + "learning_rate": 1.5403311655797492e-07, + "loss": 0.5604, + "step": 7317 + }, + { + "epoch": 13.551851851851852, + "grad_norm": 1.610429523590397, + "learning_rate": 1.5402049901263324e-07, + "loss": 0.5921, + "step": 7318 + }, + { + "epoch": 13.553703703703704, + "grad_norm": 1.8484256165244102, + "learning_rate": 1.5400788025276408e-07, + "loss": 0.6097, + "step": 7319 + }, + { + "epoch": 13.555555555555555, + "grad_norm": 1.5898008302380773, + "learning_rate": 1.5399526027865106e-07, + "loss": 0.6159, + "step": 7320 + }, + { + "epoch": 13.557407407407407, + "grad_norm": 1.4882874953799088, + "learning_rate": 1.5398263909057797e-07, + "loss": 0.6075, + "step": 7321 + }, + { + "epoch": 13.559259259259258, + "grad_norm": 1.5062444618629536, + "learning_rate": 1.5397001668882853e-07, + "loss": 0.6304, + "step": 7322 + }, + { + "epoch": 13.561111111111112, + "grad_norm": 1.608475389720989, + "learning_rate": 1.5395739307368652e-07, + "loss": 0.6446, + "step": 7323 + }, + { + "epoch": 13.562962962962963, + "grad_norm": 1.4871846207653443, + "learning_rate": 1.5394476824543574e-07, + "loss": 0.6002, + "step": 7324 + }, + { + "epoch": 13.564814814814815, + "grad_norm": 1.4487657844756618, + "learning_rate": 1.5393214220436008e-07, + "loss": 0.5768, + "step": 7325 + }, + { + "epoch": 13.566666666666666, + "grad_norm": 1.7145655159524333, + "learning_rate": 1.539195149507434e-07, + "loss": 0.6152, + "step": 7326 + }, + { + "epoch": 13.568518518518518, + "grad_norm": 1.4793640287813234, + "learning_rate": 1.5390688648486957e-07, + "loss": 0.5893, + "step": 7327 + }, + { + "epoch": 13.57037037037037, + "grad_norm": 1.4231671694241494, + "learning_rate": 1.5389425680702254e-07, + "loss": 0.5981, + "step": 7328 + }, + { + "epoch": 13.572222222222223, + "grad_norm": 1.644217236502398, + "learning_rate": 1.5388162591748623e-07, + "loss": 0.5845, + "step": 7329 + }, + { + "epoch": 13.574074074074074, + "grad_norm": 1.8000892300667184, + "learning_rate": 1.5386899381654462e-07, + "loss": 0.5989, + "step": 7330 + }, + { + "epoch": 13.575925925925926, + "grad_norm": 1.5449991231500702, + "learning_rate": 1.538563605044817e-07, + "loss": 0.5856, + "step": 7331 + }, + { + "epoch": 13.577777777777778, + "grad_norm": 1.5575671892729914, + "learning_rate": 1.5384372598158158e-07, + "loss": 0.5955, + "step": 7332 + }, + { + "epoch": 13.579629629629629, + "grad_norm": 1.569823369996743, + "learning_rate": 1.5383109024812822e-07, + "loss": 0.5563, + "step": 7333 + }, + { + "epoch": 13.581481481481482, + "grad_norm": 1.521514141677872, + "learning_rate": 1.5381845330440578e-07, + "loss": 0.5996, + "step": 7334 + }, + { + "epoch": 13.583333333333334, + "grad_norm": 1.6437900401362477, + "learning_rate": 1.5380581515069832e-07, + "loss": 0.6201, + "step": 7335 + }, + { + "epoch": 13.585185185185185, + "grad_norm": 2.001926645392715, + "learning_rate": 1.5379317578729003e-07, + "loss": 0.628, + "step": 7336 + }, + { + "epoch": 13.587037037037037, + "grad_norm": 1.8155289947708695, + "learning_rate": 1.5378053521446504e-07, + "loss": 0.6114, + "step": 7337 + }, + { + "epoch": 13.588888888888889, + "grad_norm": 2.8460597973518165, + "learning_rate": 1.537678934325075e-07, + "loss": 0.62, + "step": 7338 + }, + { + "epoch": 13.59074074074074, + "grad_norm": 1.4605211043987703, + "learning_rate": 1.537552504417017e-07, + "loss": 0.6138, + "step": 7339 + }, + { + "epoch": 13.592592592592592, + "grad_norm": 1.613214358650215, + "learning_rate": 1.5374260624233192e-07, + "loss": 0.6319, + "step": 7340 + }, + { + "epoch": 13.594444444444445, + "grad_norm": 1.765575642499274, + "learning_rate": 1.537299608346824e-07, + "loss": 0.5596, + "step": 7341 + }, + { + "epoch": 13.596296296296297, + "grad_norm": 1.4628720372800443, + "learning_rate": 1.5371731421903737e-07, + "loss": 0.652, + "step": 7342 + }, + { + "epoch": 13.598148148148148, + "grad_norm": 1.6429603751209803, + "learning_rate": 1.5370466639568127e-07, + "loss": 0.609, + "step": 7343 + }, + { + "epoch": 13.6, + "grad_norm": 1.9259401560372884, + "learning_rate": 1.5369201736489839e-07, + "loss": 0.6622, + "step": 7344 + }, + { + "epoch": 13.601851851851851, + "grad_norm": 1.5081132064301976, + "learning_rate": 1.5367936712697311e-07, + "loss": 0.5951, + "step": 7345 + }, + { + "epoch": 13.603703703703705, + "grad_norm": 1.9195505522029188, + "learning_rate": 1.536667156821899e-07, + "loss": 0.6105, + "step": 7346 + }, + { + "epoch": 13.605555555555556, + "grad_norm": 1.7798404802093304, + "learning_rate": 1.5365406303083316e-07, + "loss": 0.6236, + "step": 7347 + }, + { + "epoch": 13.607407407407408, + "grad_norm": 1.5916803097408845, + "learning_rate": 1.5364140917318734e-07, + "loss": 0.6189, + "step": 7348 + }, + { + "epoch": 13.60925925925926, + "grad_norm": 1.9182538034717207, + "learning_rate": 1.5362875410953696e-07, + "loss": 0.6027, + "step": 7349 + }, + { + "epoch": 13.61111111111111, + "grad_norm": 1.6311275093183528, + "learning_rate": 1.5361609784016655e-07, + "loss": 0.6236, + "step": 7350 + }, + { + "epoch": 13.612962962962962, + "grad_norm": 2.1395493644448034, + "learning_rate": 1.5360344036536063e-07, + "loss": 0.6166, + "step": 7351 + }, + { + "epoch": 13.614814814814816, + "grad_norm": 1.748014452546716, + "learning_rate": 1.535907816854038e-07, + "loss": 0.6055, + "step": 7352 + }, + { + "epoch": 13.616666666666667, + "grad_norm": 1.5338843474064465, + "learning_rate": 1.5357812180058064e-07, + "loss": 0.6107, + "step": 7353 + }, + { + "epoch": 13.618518518518519, + "grad_norm": 1.8157125463606256, + "learning_rate": 1.5356546071117578e-07, + "loss": 0.6071, + "step": 7354 + }, + { + "epoch": 13.62037037037037, + "grad_norm": 1.8375164802228954, + "learning_rate": 1.535527984174739e-07, + "loss": 0.608, + "step": 7355 + }, + { + "epoch": 13.622222222222222, + "grad_norm": 6.514712478612919, + "learning_rate": 1.5354013491975964e-07, + "loss": 0.577, + "step": 7356 + }, + { + "epoch": 13.624074074074073, + "grad_norm": 1.8417580603042927, + "learning_rate": 1.5352747021831775e-07, + "loss": 0.6347, + "step": 7357 + }, + { + "epoch": 13.625925925925927, + "grad_norm": 1.544444473697421, + "learning_rate": 1.5351480431343296e-07, + "loss": 0.6283, + "step": 7358 + }, + { + "epoch": 13.627777777777778, + "grad_norm": 1.5083167596088436, + "learning_rate": 1.5350213720539e-07, + "loss": 0.6048, + "step": 7359 + }, + { + "epoch": 13.62962962962963, + "grad_norm": 2.098424536431358, + "learning_rate": 1.5348946889447368e-07, + "loss": 0.5886, + "step": 7360 + }, + { + "epoch": 13.631481481481481, + "grad_norm": 1.31449862759255, + "learning_rate": 1.5347679938096887e-07, + "loss": 0.5811, + "step": 7361 + }, + { + "epoch": 13.633333333333333, + "grad_norm": 1.5898861162618418, + "learning_rate": 1.534641286651603e-07, + "loss": 0.5947, + "step": 7362 + }, + { + "epoch": 13.635185185185184, + "grad_norm": 6.156323347452639, + "learning_rate": 1.5345145674733296e-07, + "loss": 0.6359, + "step": 7363 + }, + { + "epoch": 13.637037037037038, + "grad_norm": 1.4335975194366528, + "learning_rate": 1.534387836277717e-07, + "loss": 0.5937, + "step": 7364 + }, + { + "epoch": 13.63888888888889, + "grad_norm": 1.4793057527138505, + "learning_rate": 1.5342610930676144e-07, + "loss": 0.6375, + "step": 7365 + }, + { + "epoch": 13.64074074074074, + "grad_norm": 1.543080566756526, + "learning_rate": 1.5341343378458714e-07, + "loss": 0.612, + "step": 7366 + }, + { + "epoch": 13.642592592592592, + "grad_norm": 1.5461909039071664, + "learning_rate": 1.534007570615338e-07, + "loss": 0.6266, + "step": 7367 + }, + { + "epoch": 13.644444444444444, + "grad_norm": 1.552960230418376, + "learning_rate": 1.5338807913788637e-07, + "loss": 0.6147, + "step": 7368 + }, + { + "epoch": 13.646296296296295, + "grad_norm": 1.5874116701323173, + "learning_rate": 1.533754000139299e-07, + "loss": 0.6015, + "step": 7369 + }, + { + "epoch": 13.648148148148149, + "grad_norm": 1.4788500930859279, + "learning_rate": 1.5336271968994953e-07, + "loss": 0.6026, + "step": 7370 + }, + { + "epoch": 13.65, + "grad_norm": 1.8610380865636993, + "learning_rate": 1.5335003816623025e-07, + "loss": 0.6267, + "step": 7371 + }, + { + "epoch": 13.651851851851852, + "grad_norm": 1.6122250355233885, + "learning_rate": 1.5333735544305724e-07, + "loss": 0.6289, + "step": 7372 + }, + { + "epoch": 13.653703703703703, + "grad_norm": 1.634396649438595, + "learning_rate": 1.5332467152071562e-07, + "loss": 0.6382, + "step": 7373 + }, + { + "epoch": 13.655555555555555, + "grad_norm": 1.4913035533512933, + "learning_rate": 1.5331198639949054e-07, + "loss": 0.5998, + "step": 7374 + }, + { + "epoch": 13.657407407407408, + "grad_norm": 1.7665538456455654, + "learning_rate": 1.5329930007966724e-07, + "loss": 0.6186, + "step": 7375 + }, + { + "epoch": 13.65925925925926, + "grad_norm": 1.4900024599502697, + "learning_rate": 1.532866125615309e-07, + "loss": 0.5899, + "step": 7376 + }, + { + "epoch": 13.661111111111111, + "grad_norm": 2.6166246127599644, + "learning_rate": 1.5327392384536677e-07, + "loss": 0.6437, + "step": 7377 + }, + { + "epoch": 13.662962962962963, + "grad_norm": 1.9815153624534798, + "learning_rate": 1.5326123393146015e-07, + "loss": 0.6009, + "step": 7378 + }, + { + "epoch": 13.664814814814815, + "grad_norm": 1.654381623895425, + "learning_rate": 1.5324854282009635e-07, + "loss": 0.6323, + "step": 7379 + }, + { + "epoch": 13.666666666666666, + "grad_norm": 1.4508373524002751, + "learning_rate": 1.5323585051156067e-07, + "loss": 0.5848, + "step": 7380 + }, + { + "epoch": 13.668518518518518, + "grad_norm": 1.4805706596039312, + "learning_rate": 1.532231570061385e-07, + "loss": 0.5719, + "step": 7381 + }, + { + "epoch": 13.670370370370371, + "grad_norm": 1.4998068656259442, + "learning_rate": 1.5321046230411518e-07, + "loss": 0.6457, + "step": 7382 + }, + { + "epoch": 13.672222222222222, + "grad_norm": 1.7749691130908583, + "learning_rate": 1.5319776640577615e-07, + "loss": 0.6141, + "step": 7383 + }, + { + "epoch": 13.674074074074074, + "grad_norm": 1.6413897860270015, + "learning_rate": 1.531850693114069e-07, + "loss": 0.5974, + "step": 7384 + }, + { + "epoch": 13.675925925925926, + "grad_norm": 1.5885803883419858, + "learning_rate": 1.5317237102129278e-07, + "loss": 0.6212, + "step": 7385 + }, + { + "epoch": 13.677777777777777, + "grad_norm": 1.7864568282209805, + "learning_rate": 1.5315967153571938e-07, + "loss": 0.6203, + "step": 7386 + }, + { + "epoch": 13.67962962962963, + "grad_norm": 1.4952710468927601, + "learning_rate": 1.531469708549722e-07, + "loss": 0.5791, + "step": 7387 + }, + { + "epoch": 13.681481481481482, + "grad_norm": 1.6540614503238513, + "learning_rate": 1.5313426897933676e-07, + "loss": 0.632, + "step": 7388 + }, + { + "epoch": 13.683333333333334, + "grad_norm": 1.530135127565535, + "learning_rate": 1.5312156590909862e-07, + "loss": 0.6766, + "step": 7389 + }, + { + "epoch": 13.685185185185185, + "grad_norm": 2.724539342665472, + "learning_rate": 1.5310886164454342e-07, + "loss": 0.6259, + "step": 7390 + }, + { + "epoch": 13.687037037037037, + "grad_norm": 1.5831600943349637, + "learning_rate": 1.5309615618595676e-07, + "loss": 0.6185, + "step": 7391 + }, + { + "epoch": 13.688888888888888, + "grad_norm": 1.7304900947870607, + "learning_rate": 1.530834495336243e-07, + "loss": 0.6375, + "step": 7392 + }, + { + "epoch": 13.690740740740742, + "grad_norm": 1.5969775695476667, + "learning_rate": 1.5307074168783172e-07, + "loss": 0.6067, + "step": 7393 + }, + { + "epoch": 13.692592592592593, + "grad_norm": 2.12885004052106, + "learning_rate": 1.5305803264886478e-07, + "loss": 0.6166, + "step": 7394 + }, + { + "epoch": 13.694444444444445, + "grad_norm": 1.479468479344743, + "learning_rate": 1.530453224170091e-07, + "loss": 0.6492, + "step": 7395 + }, + { + "epoch": 13.696296296296296, + "grad_norm": 1.415858567106282, + "learning_rate": 1.5303261099255051e-07, + "loss": 0.6009, + "step": 7396 + }, + { + "epoch": 13.698148148148148, + "grad_norm": 1.713486469796353, + "learning_rate": 1.5301989837577482e-07, + "loss": 0.62, + "step": 7397 + }, + { + "epoch": 13.7, + "grad_norm": 1.4126780018601661, + "learning_rate": 1.5300718456696777e-07, + "loss": 0.5976, + "step": 7398 + }, + { + "epoch": 13.701851851851853, + "grad_norm": 2.963230108176807, + "learning_rate": 1.5299446956641528e-07, + "loss": 0.6037, + "step": 7399 + }, + { + "epoch": 13.703703703703704, + "grad_norm": 1.6065410080413833, + "learning_rate": 1.529817533744032e-07, + "loss": 0.5605, + "step": 7400 + }, + { + "epoch": 13.705555555555556, + "grad_norm": 1.4501392385522893, + "learning_rate": 1.5296903599121738e-07, + "loss": 0.5731, + "step": 7401 + }, + { + "epoch": 13.707407407407407, + "grad_norm": 1.6026306288183525, + "learning_rate": 1.5295631741714376e-07, + "loss": 0.6419, + "step": 7402 + }, + { + "epoch": 13.709259259259259, + "grad_norm": 3.99042240217245, + "learning_rate": 1.5294359765246833e-07, + "loss": 0.5675, + "step": 7403 + }, + { + "epoch": 13.71111111111111, + "grad_norm": 1.753696972944809, + "learning_rate": 1.52930876697477e-07, + "loss": 0.648, + "step": 7404 + }, + { + "epoch": 13.712962962962964, + "grad_norm": 1.6167464511805714, + "learning_rate": 1.5291815455245582e-07, + "loss": 0.6131, + "step": 7405 + }, + { + "epoch": 13.714814814814815, + "grad_norm": 1.5039602910889693, + "learning_rate": 1.529054312176908e-07, + "loss": 0.6076, + "step": 7406 + }, + { + "epoch": 13.716666666666667, + "grad_norm": 1.6105651325131158, + "learning_rate": 1.5289270669346798e-07, + "loss": 0.5775, + "step": 7407 + }, + { + "epoch": 13.718518518518518, + "grad_norm": 2.269332215107279, + "learning_rate": 1.5287998098007348e-07, + "loss": 0.6132, + "step": 7408 + }, + { + "epoch": 13.72037037037037, + "grad_norm": 1.6379395157746772, + "learning_rate": 1.528672540777934e-07, + "loss": 0.6155, + "step": 7409 + }, + { + "epoch": 13.722222222222221, + "grad_norm": 1.696520219701363, + "learning_rate": 1.5285452598691385e-07, + "loss": 0.6205, + "step": 7410 + }, + { + "epoch": 13.724074074074075, + "grad_norm": 1.6390249447210787, + "learning_rate": 1.52841796707721e-07, + "loss": 0.5892, + "step": 7411 + }, + { + "epoch": 13.725925925925926, + "grad_norm": 1.4353779472714796, + "learning_rate": 1.5282906624050106e-07, + "loss": 0.6021, + "step": 7412 + }, + { + "epoch": 13.727777777777778, + "grad_norm": 1.665041093117239, + "learning_rate": 1.528163345855402e-07, + "loss": 0.6219, + "step": 7413 + }, + { + "epoch": 13.72962962962963, + "grad_norm": 1.8361238860153073, + "learning_rate": 1.528036017431247e-07, + "loss": 0.6302, + "step": 7414 + }, + { + "epoch": 13.731481481481481, + "grad_norm": 2.444227357715716, + "learning_rate": 1.5279086771354085e-07, + "loss": 0.6384, + "step": 7415 + }, + { + "epoch": 13.733333333333333, + "grad_norm": 1.676112674264642, + "learning_rate": 1.5277813249707485e-07, + "loss": 0.6277, + "step": 7416 + }, + { + "epoch": 13.735185185185186, + "grad_norm": 1.5411296763924103, + "learning_rate": 1.5276539609401313e-07, + "loss": 0.5846, + "step": 7417 + }, + { + "epoch": 13.737037037037037, + "grad_norm": 1.8206556729405725, + "learning_rate": 1.52752658504642e-07, + "loss": 0.6279, + "step": 7418 + }, + { + "epoch": 13.738888888888889, + "grad_norm": 1.5270173900009647, + "learning_rate": 1.5273991972924785e-07, + "loss": 0.5975, + "step": 7419 + }, + { + "epoch": 13.74074074074074, + "grad_norm": 1.4001710954295337, + "learning_rate": 1.5272717976811706e-07, + "loss": 0.5899, + "step": 7420 + }, + { + "epoch": 13.742592592592592, + "grad_norm": 1.7603814253515966, + "learning_rate": 1.5271443862153602e-07, + "loss": 0.6483, + "step": 7421 + }, + { + "epoch": 13.744444444444444, + "grad_norm": 2.231649106584098, + "learning_rate": 1.5270169628979125e-07, + "loss": 0.6435, + "step": 7422 + }, + { + "epoch": 13.746296296296297, + "grad_norm": 1.9902801010244227, + "learning_rate": 1.5268895277316927e-07, + "loss": 0.6372, + "step": 7423 + }, + { + "epoch": 13.748148148148148, + "grad_norm": 2.219839439468968, + "learning_rate": 1.5267620807195648e-07, + "loss": 0.6347, + "step": 7424 + }, + { + "epoch": 13.75, + "grad_norm": 1.6064749472489948, + "learning_rate": 1.5266346218643947e-07, + "loss": 0.6216, + "step": 7425 + }, + { + "epoch": 13.751851851851852, + "grad_norm": 2.9189469580457255, + "learning_rate": 1.526507151169048e-07, + "loss": 0.633, + "step": 7426 + }, + { + "epoch": 13.753703703703703, + "grad_norm": 1.8347096468245494, + "learning_rate": 1.5263796686363906e-07, + "loss": 0.6008, + "step": 7427 + }, + { + "epoch": 13.755555555555556, + "grad_norm": 1.7388257856616471, + "learning_rate": 1.5262521742692882e-07, + "loss": 0.6186, + "step": 7428 + }, + { + "epoch": 13.757407407407408, + "grad_norm": 1.7701552838125272, + "learning_rate": 1.5261246680706081e-07, + "loss": 0.6224, + "step": 7429 + }, + { + "epoch": 13.75925925925926, + "grad_norm": 1.4783343364839439, + "learning_rate": 1.5259971500432164e-07, + "loss": 0.5875, + "step": 7430 + }, + { + "epoch": 13.761111111111111, + "grad_norm": 1.727381410170526, + "learning_rate": 1.52586962018998e-07, + "loss": 0.5658, + "step": 7431 + }, + { + "epoch": 13.762962962962963, + "grad_norm": 1.6390493812494877, + "learning_rate": 1.5257420785137667e-07, + "loss": 0.6398, + "step": 7432 + }, + { + "epoch": 13.764814814814814, + "grad_norm": 1.4442205600902545, + "learning_rate": 1.5256145250174432e-07, + "loss": 0.596, + "step": 7433 + }, + { + "epoch": 13.766666666666667, + "grad_norm": 1.7294103959350773, + "learning_rate": 1.525486959703878e-07, + "loss": 0.6363, + "step": 7434 + }, + { + "epoch": 13.768518518518519, + "grad_norm": 1.4786264646766074, + "learning_rate": 1.5253593825759389e-07, + "loss": 0.6227, + "step": 7435 + }, + { + "epoch": 13.77037037037037, + "grad_norm": 1.7818247306411379, + "learning_rate": 1.5252317936364936e-07, + "loss": 0.6169, + "step": 7436 + }, + { + "epoch": 13.772222222222222, + "grad_norm": 1.8125349478392643, + "learning_rate": 1.5251041928884109e-07, + "loss": 0.5846, + "step": 7437 + }, + { + "epoch": 13.774074074074074, + "grad_norm": 1.9276764408965896, + "learning_rate": 1.52497658033456e-07, + "loss": 0.6637, + "step": 7438 + }, + { + "epoch": 13.775925925925925, + "grad_norm": 1.4501373463490161, + "learning_rate": 1.52484895597781e-07, + "loss": 0.6232, + "step": 7439 + }, + { + "epoch": 13.777777777777779, + "grad_norm": 13.210332842084, + "learning_rate": 1.5247213198210298e-07, + "loss": 0.5964, + "step": 7440 + }, + { + "epoch": 13.77962962962963, + "grad_norm": 1.4691993631638716, + "learning_rate": 1.5245936718670894e-07, + "loss": 0.6464, + "step": 7441 + }, + { + "epoch": 13.781481481481482, + "grad_norm": 1.7313381439660658, + "learning_rate": 1.5244660121188582e-07, + "loss": 0.5708, + "step": 7442 + }, + { + "epoch": 13.783333333333333, + "grad_norm": 1.8955162085848853, + "learning_rate": 1.5243383405792065e-07, + "loss": 0.5971, + "step": 7443 + }, + { + "epoch": 13.785185185185185, + "grad_norm": 1.3887040752912538, + "learning_rate": 1.5242106572510053e-07, + "loss": 0.5704, + "step": 7444 + }, + { + "epoch": 13.787037037037036, + "grad_norm": 1.4437963360413026, + "learning_rate": 1.5240829621371246e-07, + "loss": 0.6285, + "step": 7445 + }, + { + "epoch": 13.78888888888889, + "grad_norm": 1.6516054223999488, + "learning_rate": 1.5239552552404352e-07, + "loss": 0.5954, + "step": 7446 + }, + { + "epoch": 13.790740740740741, + "grad_norm": 1.6698616190381441, + "learning_rate": 1.5238275365638087e-07, + "loss": 0.6293, + "step": 7447 + }, + { + "epoch": 13.792592592592593, + "grad_norm": 1.5485082839958755, + "learning_rate": 1.523699806110117e-07, + "loss": 0.6373, + "step": 7448 + }, + { + "epoch": 13.794444444444444, + "grad_norm": 1.6018300173580908, + "learning_rate": 1.5235720638822307e-07, + "loss": 0.6118, + "step": 7449 + }, + { + "epoch": 13.796296296296296, + "grad_norm": 1.5638209961825218, + "learning_rate": 1.5234443098830227e-07, + "loss": 0.6082, + "step": 7450 + }, + { + "epoch": 13.798148148148147, + "grad_norm": 1.6205340752199529, + "learning_rate": 1.523316544115365e-07, + "loss": 0.6201, + "step": 7451 + }, + { + "epoch": 13.8, + "grad_norm": 2.270837949949198, + "learning_rate": 1.52318876658213e-07, + "loss": 0.6118, + "step": 7452 + }, + { + "epoch": 13.801851851851852, + "grad_norm": 1.5336222349376536, + "learning_rate": 1.5230609772861906e-07, + "loss": 0.6208, + "step": 7453 + }, + { + "epoch": 13.803703703703704, + "grad_norm": 1.6090759455501418, + "learning_rate": 1.5229331762304198e-07, + "loss": 0.676, + "step": 7454 + }, + { + "epoch": 13.805555555555555, + "grad_norm": 1.5691235604657463, + "learning_rate": 1.5228053634176908e-07, + "loss": 0.6195, + "step": 7455 + }, + { + "epoch": 13.807407407407407, + "grad_norm": 1.6087985340238542, + "learning_rate": 1.5226775388508776e-07, + "loss": 0.5408, + "step": 7456 + }, + { + "epoch": 13.809259259259258, + "grad_norm": 1.7289181264702418, + "learning_rate": 1.5225497025328538e-07, + "loss": 0.6325, + "step": 7457 + }, + { + "epoch": 13.811111111111112, + "grad_norm": 1.616759707457118, + "learning_rate": 1.5224218544664932e-07, + "loss": 0.5784, + "step": 7458 + }, + { + "epoch": 13.812962962962963, + "grad_norm": 2.358364085658103, + "learning_rate": 1.5222939946546705e-07, + "loss": 0.5743, + "step": 7459 + }, + { + "epoch": 13.814814814814815, + "grad_norm": 1.624200904225749, + "learning_rate": 1.5221661231002604e-07, + "loss": 0.6122, + "step": 7460 + }, + { + "epoch": 13.816666666666666, + "grad_norm": 2.0094410410225074, + "learning_rate": 1.5220382398061372e-07, + "loss": 0.5968, + "step": 7461 + }, + { + "epoch": 13.818518518518518, + "grad_norm": 1.7524849572883625, + "learning_rate": 1.521910344775177e-07, + "loss": 0.6084, + "step": 7462 + }, + { + "epoch": 13.82037037037037, + "grad_norm": 2.2907665283520746, + "learning_rate": 1.5217824380102546e-07, + "loss": 0.6396, + "step": 7463 + }, + { + "epoch": 13.822222222222223, + "grad_norm": 1.6997170285552388, + "learning_rate": 1.521654519514246e-07, + "loss": 0.5967, + "step": 7464 + }, + { + "epoch": 13.824074074074074, + "grad_norm": 1.4812071035840988, + "learning_rate": 1.5215265892900268e-07, + "loss": 0.6304, + "step": 7465 + }, + { + "epoch": 13.825925925925926, + "grad_norm": 1.3337623288502478, + "learning_rate": 1.5213986473404738e-07, + "loss": 0.6512, + "step": 7466 + }, + { + "epoch": 13.827777777777778, + "grad_norm": 1.774889839335562, + "learning_rate": 1.5212706936684625e-07, + "loss": 0.5953, + "step": 7467 + }, + { + "epoch": 13.829629629629629, + "grad_norm": 1.5128671129439468, + "learning_rate": 1.5211427282768707e-07, + "loss": 0.6017, + "step": 7468 + }, + { + "epoch": 13.831481481481482, + "grad_norm": 1.8758854791859465, + "learning_rate": 1.5210147511685748e-07, + "loss": 0.6377, + "step": 7469 + }, + { + "epoch": 13.833333333333334, + "grad_norm": 1.776259658725903, + "learning_rate": 1.5208867623464525e-07, + "loss": 0.6105, + "step": 7470 + }, + { + "epoch": 13.835185185185185, + "grad_norm": 1.6788360374008477, + "learning_rate": 1.520758761813381e-07, + "loss": 0.6365, + "step": 7471 + }, + { + "epoch": 13.837037037037037, + "grad_norm": 1.7986688118321117, + "learning_rate": 1.5206307495722376e-07, + "loss": 0.6342, + "step": 7472 + }, + { + "epoch": 13.838888888888889, + "grad_norm": 1.7672589686278617, + "learning_rate": 1.5205027256259015e-07, + "loss": 0.6109, + "step": 7473 + }, + { + "epoch": 13.84074074074074, + "grad_norm": 1.5984206854799181, + "learning_rate": 1.5203746899772506e-07, + "loss": 0.6271, + "step": 7474 + }, + { + "epoch": 13.842592592592592, + "grad_norm": 1.6617294882748468, + "learning_rate": 1.5202466426291629e-07, + "loss": 0.6443, + "step": 7475 + }, + { + "epoch": 13.844444444444445, + "grad_norm": 1.777347321358636, + "learning_rate": 1.520118583584518e-07, + "loss": 0.6037, + "step": 7476 + }, + { + "epoch": 13.846296296296297, + "grad_norm": 3.0112949931328545, + "learning_rate": 1.5199905128461946e-07, + "loss": 0.6764, + "step": 7477 + }, + { + "epoch": 13.848148148148148, + "grad_norm": 1.3697416622569698, + "learning_rate": 1.5198624304170724e-07, + "loss": 0.624, + "step": 7478 + }, + { + "epoch": 13.85, + "grad_norm": 1.6242529378445814, + "learning_rate": 1.5197343363000306e-07, + "loss": 0.6214, + "step": 7479 + }, + { + "epoch": 13.851851851851851, + "grad_norm": 1.8263633319167236, + "learning_rate": 1.5196062304979495e-07, + "loss": 0.6323, + "step": 7480 + }, + { + "epoch": 13.853703703703705, + "grad_norm": 1.9313327263554292, + "learning_rate": 1.519478113013709e-07, + "loss": 0.5819, + "step": 7481 + }, + { + "epoch": 13.855555555555556, + "grad_norm": 1.8538232588938657, + "learning_rate": 1.5193499838501894e-07, + "loss": 0.6124, + "step": 7482 + }, + { + "epoch": 13.857407407407408, + "grad_norm": 1.5878132919931682, + "learning_rate": 1.519221843010272e-07, + "loss": 0.5779, + "step": 7483 + }, + { + "epoch": 13.85925925925926, + "grad_norm": 1.7082702356973152, + "learning_rate": 1.5190936904968373e-07, + "loss": 0.6115, + "step": 7484 + }, + { + "epoch": 13.86111111111111, + "grad_norm": 1.44388484363116, + "learning_rate": 1.5189655263127665e-07, + "loss": 0.5855, + "step": 7485 + }, + { + "epoch": 13.862962962962962, + "grad_norm": 1.9176393634695494, + "learning_rate": 1.5188373504609416e-07, + "loss": 0.64, + "step": 7486 + }, + { + "epoch": 13.864814814814816, + "grad_norm": 1.7849555872534677, + "learning_rate": 1.5187091629442435e-07, + "loss": 0.5627, + "step": 7487 + }, + { + "epoch": 13.866666666666667, + "grad_norm": 2.241748322442699, + "learning_rate": 1.5185809637655547e-07, + "loss": 0.5627, + "step": 7488 + }, + { + "epoch": 13.868518518518519, + "grad_norm": 1.658344388672207, + "learning_rate": 1.5184527529277573e-07, + "loss": 0.5979, + "step": 7489 + }, + { + "epoch": 13.87037037037037, + "grad_norm": 1.6379350935913863, + "learning_rate": 1.518324530433734e-07, + "loss": 0.6274, + "step": 7490 + }, + { + "epoch": 13.872222222222222, + "grad_norm": 1.6469432022062998, + "learning_rate": 1.5181962962863675e-07, + "loss": 0.6199, + "step": 7491 + }, + { + "epoch": 13.874074074074073, + "grad_norm": 1.6831439133032498, + "learning_rate": 1.518068050488541e-07, + "loss": 0.6441, + "step": 7492 + }, + { + "epoch": 13.875925925925927, + "grad_norm": 2.4944938096229006, + "learning_rate": 1.5179397930431376e-07, + "loss": 0.6658, + "step": 7493 + }, + { + "epoch": 13.877777777777778, + "grad_norm": 1.6180551369726002, + "learning_rate": 1.5178115239530408e-07, + "loss": 0.6256, + "step": 7494 + }, + { + "epoch": 13.87962962962963, + "grad_norm": 1.6910039192989743, + "learning_rate": 1.5176832432211347e-07, + "loss": 0.6018, + "step": 7495 + }, + { + "epoch": 13.881481481481481, + "grad_norm": 1.90187540617632, + "learning_rate": 1.5175549508503035e-07, + "loss": 0.6609, + "step": 7496 + }, + { + "epoch": 13.883333333333333, + "grad_norm": 1.52060504381097, + "learning_rate": 1.517426646843431e-07, + "loss": 0.644, + "step": 7497 + }, + { + "epoch": 13.885185185185184, + "grad_norm": 1.4153839951882483, + "learning_rate": 1.5172983312034022e-07, + "loss": 0.5626, + "step": 7498 + }, + { + "epoch": 13.887037037037038, + "grad_norm": 1.919721454186169, + "learning_rate": 1.517170003933102e-07, + "loss": 0.61, + "step": 7499 + }, + { + "epoch": 13.88888888888889, + "grad_norm": 1.5001673423864423, + "learning_rate": 1.5170416650354154e-07, + "loss": 0.6032, + "step": 7500 + }, + { + "epoch": 13.89074074074074, + "grad_norm": 1.5437771002134788, + "learning_rate": 1.5169133145132283e-07, + "loss": 0.613, + "step": 7501 + }, + { + "epoch": 13.892592592592592, + "grad_norm": 1.547238307456619, + "learning_rate": 1.5167849523694258e-07, + "loss": 0.6299, + "step": 7502 + }, + { + "epoch": 13.894444444444444, + "grad_norm": 2.0772894368064967, + "learning_rate": 1.516656578606894e-07, + "loss": 0.6165, + "step": 7503 + }, + { + "epoch": 13.896296296296295, + "grad_norm": 1.423394506398061, + "learning_rate": 1.516528193228519e-07, + "loss": 0.6214, + "step": 7504 + }, + { + "epoch": 13.898148148148149, + "grad_norm": 1.6038144541634802, + "learning_rate": 1.516399796237187e-07, + "loss": 0.6162, + "step": 7505 + }, + { + "epoch": 13.9, + "grad_norm": 1.6410251630798371, + "learning_rate": 1.5162713876357859e-07, + "loss": 0.6452, + "step": 7506 + }, + { + "epoch": 13.901851851851852, + "grad_norm": 1.5118939799219775, + "learning_rate": 1.5161429674272014e-07, + "loss": 0.5967, + "step": 7507 + }, + { + "epoch": 13.903703703703703, + "grad_norm": 2.072781119916379, + "learning_rate": 1.516014535614321e-07, + "loss": 0.6068, + "step": 7508 + }, + { + "epoch": 13.905555555555555, + "grad_norm": 1.6248162068287648, + "learning_rate": 1.5158860922000326e-07, + "loss": 0.5914, + "step": 7509 + }, + { + "epoch": 13.907407407407408, + "grad_norm": 1.6968485150986787, + "learning_rate": 1.5157576371872235e-07, + "loss": 0.5756, + "step": 7510 + }, + { + "epoch": 13.90925925925926, + "grad_norm": 1.6217502390994523, + "learning_rate": 1.515629170578782e-07, + "loss": 0.6423, + "step": 7511 + }, + { + "epoch": 13.911111111111111, + "grad_norm": 1.761199447734665, + "learning_rate": 1.5155006923775965e-07, + "loss": 0.6122, + "step": 7512 + }, + { + "epoch": 13.912962962962963, + "grad_norm": 1.3563400575727296, + "learning_rate": 1.515372202586555e-07, + "loss": 0.6462, + "step": 7513 + }, + { + "epoch": 13.914814814814815, + "grad_norm": 1.4803800264250055, + "learning_rate": 1.5152437012085468e-07, + "loss": 0.5768, + "step": 7514 + }, + { + "epoch": 13.916666666666666, + "grad_norm": 1.537054205882389, + "learning_rate": 1.515115188246461e-07, + "loss": 0.6204, + "step": 7515 + }, + { + "epoch": 13.918518518518518, + "grad_norm": 1.435735383814482, + "learning_rate": 1.514986663703187e-07, + "loss": 0.6501, + "step": 7516 + }, + { + "epoch": 13.920370370370371, + "grad_norm": 1.5522668155834092, + "learning_rate": 1.5148581275816134e-07, + "loss": 0.6057, + "step": 7517 + }, + { + "epoch": 13.922222222222222, + "grad_norm": 1.4938712213309915, + "learning_rate": 1.5147295798846312e-07, + "loss": 0.601, + "step": 7518 + }, + { + "epoch": 13.924074074074074, + "grad_norm": 1.6221738162386912, + "learning_rate": 1.51460102061513e-07, + "loss": 0.5997, + "step": 7519 + }, + { + "epoch": 13.925925925925926, + "grad_norm": 1.6207622347607011, + "learning_rate": 1.514472449776e-07, + "loss": 0.6107, + "step": 7520 + }, + { + "epoch": 13.927777777777777, + "grad_norm": 1.7690674963714104, + "learning_rate": 1.5143438673701324e-07, + "loss": 0.5976, + "step": 7521 + }, + { + "epoch": 13.92962962962963, + "grad_norm": 1.7220497798666956, + "learning_rate": 1.5142152734004177e-07, + "loss": 0.6027, + "step": 7522 + }, + { + "epoch": 13.931481481481482, + "grad_norm": 1.3988593283737722, + "learning_rate": 1.514086667869747e-07, + "loss": 0.6133, + "step": 7523 + }, + { + "epoch": 13.933333333333334, + "grad_norm": 1.7188763586541154, + "learning_rate": 1.5139580507810117e-07, + "loss": 0.5559, + "step": 7524 + }, + { + "epoch": 13.935185185185185, + "grad_norm": 1.589229315802639, + "learning_rate": 1.5138294221371036e-07, + "loss": 0.5956, + "step": 7525 + }, + { + "epoch": 13.937037037037037, + "grad_norm": 2.445605083038169, + "learning_rate": 1.513700781940914e-07, + "loss": 0.666, + "step": 7526 + }, + { + "epoch": 13.938888888888888, + "grad_norm": 1.8101960082229966, + "learning_rate": 1.5135721301953367e-07, + "loss": 0.601, + "step": 7527 + }, + { + "epoch": 13.940740740740742, + "grad_norm": 2.3588938450912442, + "learning_rate": 1.5134434669032623e-07, + "loss": 0.6456, + "step": 7528 + }, + { + "epoch": 13.942592592592593, + "grad_norm": 1.432804026013408, + "learning_rate": 1.5133147920675842e-07, + "loss": 0.5891, + "step": 7529 + }, + { + "epoch": 13.944444444444445, + "grad_norm": 1.475675975937916, + "learning_rate": 1.5131861056911958e-07, + "loss": 0.6639, + "step": 7530 + }, + { + "epoch": 13.946296296296296, + "grad_norm": 1.7043737475500285, + "learning_rate": 1.5130574077769901e-07, + "loss": 0.6023, + "step": 7531 + }, + { + "epoch": 13.948148148148148, + "grad_norm": 1.8090890352571212, + "learning_rate": 1.51292869832786e-07, + "loss": 0.6031, + "step": 7532 + }, + { + "epoch": 13.95, + "grad_norm": 1.5429994691003408, + "learning_rate": 1.5127999773467e-07, + "loss": 0.56, + "step": 7533 + }, + { + "epoch": 13.951851851851853, + "grad_norm": 1.5946418815525665, + "learning_rate": 1.5126712448364035e-07, + "loss": 0.5918, + "step": 7534 + }, + { + "epoch": 13.953703703703704, + "grad_norm": 1.9922957525974152, + "learning_rate": 1.512542500799865e-07, + "loss": 0.5742, + "step": 7535 + }, + { + "epoch": 13.955555555555556, + "grad_norm": 1.7302611749906858, + "learning_rate": 1.5124137452399793e-07, + "loss": 0.5735, + "step": 7536 + }, + { + "epoch": 13.957407407407407, + "grad_norm": 1.6184961319236364, + "learning_rate": 1.512284978159641e-07, + "loss": 0.645, + "step": 7537 + }, + { + "epoch": 13.959259259259259, + "grad_norm": 2.7841777943945534, + "learning_rate": 1.5121561995617446e-07, + "loss": 0.5999, + "step": 7538 + }, + { + "epoch": 13.96111111111111, + "grad_norm": 1.6213083384182314, + "learning_rate": 1.5120274094491863e-07, + "loss": 0.6352, + "step": 7539 + }, + { + "epoch": 13.962962962962964, + "grad_norm": 1.7370440385827413, + "learning_rate": 1.511898607824861e-07, + "loss": 0.5674, + "step": 7540 + }, + { + "epoch": 13.964814814814815, + "grad_norm": 1.6061926243213158, + "learning_rate": 1.5117697946916645e-07, + "loss": 0.6464, + "step": 7541 + }, + { + "epoch": 13.966666666666667, + "grad_norm": 1.4386439004730254, + "learning_rate": 1.5116409700524933e-07, + "loss": 0.6309, + "step": 7542 + }, + { + "epoch": 13.968518518518518, + "grad_norm": 1.5727221414471406, + "learning_rate": 1.5115121339102435e-07, + "loss": 0.5621, + "step": 7543 + }, + { + "epoch": 13.97037037037037, + "grad_norm": 1.7831249372749607, + "learning_rate": 1.5113832862678113e-07, + "loss": 0.6112, + "step": 7544 + }, + { + "epoch": 13.972222222222221, + "grad_norm": 1.6562551890252326, + "learning_rate": 1.5112544271280942e-07, + "loss": 0.623, + "step": 7545 + }, + { + "epoch": 13.974074074074075, + "grad_norm": 1.486421417830827, + "learning_rate": 1.5111255564939894e-07, + "loss": 0.5853, + "step": 7546 + }, + { + "epoch": 13.975925925925926, + "grad_norm": 1.814560222068078, + "learning_rate": 1.5109966743683936e-07, + "loss": 0.6129, + "step": 7547 + }, + { + "epoch": 13.977777777777778, + "grad_norm": 1.3952354730022678, + "learning_rate": 1.5108677807542045e-07, + "loss": 0.5786, + "step": 7548 + }, + { + "epoch": 13.97962962962963, + "grad_norm": 1.7886332037068213, + "learning_rate": 1.5107388756543203e-07, + "loss": 0.6267, + "step": 7549 + }, + { + "epoch": 13.981481481481481, + "grad_norm": 1.672169837339197, + "learning_rate": 1.510609959071639e-07, + "loss": 0.6354, + "step": 7550 + }, + { + "epoch": 13.983333333333333, + "grad_norm": 3.231602464363265, + "learning_rate": 1.5104810310090592e-07, + "loss": 0.6085, + "step": 7551 + }, + { + "epoch": 13.985185185185186, + "grad_norm": 1.582589941045024, + "learning_rate": 1.5103520914694793e-07, + "loss": 0.6216, + "step": 7552 + }, + { + "epoch": 13.987037037037037, + "grad_norm": 1.488304524830707, + "learning_rate": 1.5102231404557984e-07, + "loss": 0.595, + "step": 7553 + }, + { + "epoch": 13.988888888888889, + "grad_norm": 1.4600270865327551, + "learning_rate": 1.5100941779709154e-07, + "loss": 0.6377, + "step": 7554 + }, + { + "epoch": 13.99074074074074, + "grad_norm": 1.5290059930363933, + "learning_rate": 1.50996520401773e-07, + "loss": 0.5955, + "step": 7555 + }, + { + "epoch": 13.992592592592592, + "grad_norm": 1.4632364788713286, + "learning_rate": 1.5098362185991417e-07, + "loss": 0.6243, + "step": 7556 + }, + { + "epoch": 13.994444444444444, + "grad_norm": 1.5916757089413898, + "learning_rate": 1.5097072217180508e-07, + "loss": 0.6132, + "step": 7557 + }, + { + "epoch": 13.996296296296297, + "grad_norm": 1.9864804342332922, + "learning_rate": 1.5095782133773568e-07, + "loss": 0.6081, + "step": 7558 + }, + { + "epoch": 13.998148148148148, + "grad_norm": 2.192119728323222, + "learning_rate": 1.5094491935799604e-07, + "loss": 0.6164, + "step": 7559 + }, + { + "epoch": 14.0, + "grad_norm": 1.6209309242937744, + "learning_rate": 1.509320162328763e-07, + "loss": 0.6022, + "step": 7560 + }, + { + "epoch": 14.001851851851852, + "grad_norm": 2.604758127966735, + "learning_rate": 1.5091911196266648e-07, + "loss": 0.5719, + "step": 7561 + }, + { + "epoch": 14.003703703703703, + "grad_norm": 1.75994585102352, + "learning_rate": 1.5090620654765674e-07, + "loss": 0.5956, + "step": 7562 + }, + { + "epoch": 14.005555555555556, + "grad_norm": 2.827194029502754, + "learning_rate": 1.508932999881372e-07, + "loss": 0.6064, + "step": 7563 + }, + { + "epoch": 14.007407407407408, + "grad_norm": 1.9463303038473025, + "learning_rate": 1.5088039228439807e-07, + "loss": 0.6247, + "step": 7564 + }, + { + "epoch": 14.00925925925926, + "grad_norm": 1.50148938554801, + "learning_rate": 1.5086748343672953e-07, + "loss": 0.6056, + "step": 7565 + }, + { + "epoch": 14.011111111111111, + "grad_norm": 1.5116653951358134, + "learning_rate": 1.508545734454218e-07, + "loss": 0.639, + "step": 7566 + }, + { + "epoch": 14.012962962962963, + "grad_norm": 1.4749593468434457, + "learning_rate": 1.5084166231076515e-07, + "loss": 0.6337, + "step": 7567 + }, + { + "epoch": 14.014814814814814, + "grad_norm": 9.499893469555719, + "learning_rate": 1.5082875003304985e-07, + "loss": 0.5658, + "step": 7568 + }, + { + "epoch": 14.016666666666667, + "grad_norm": 1.5937056086985182, + "learning_rate": 1.5081583661256619e-07, + "loss": 0.6391, + "step": 7569 + }, + { + "epoch": 14.018518518518519, + "grad_norm": 1.5027774272678838, + "learning_rate": 1.5080292204960448e-07, + "loss": 0.6541, + "step": 7570 + }, + { + "epoch": 14.02037037037037, + "grad_norm": 1.5736168915432671, + "learning_rate": 1.5079000634445514e-07, + "loss": 0.5793, + "step": 7571 + }, + { + "epoch": 14.022222222222222, + "grad_norm": 2.466699452266482, + "learning_rate": 1.5077708949740854e-07, + "loss": 0.6091, + "step": 7572 + }, + { + "epoch": 14.024074074074074, + "grad_norm": 1.4343678001463427, + "learning_rate": 1.50764171508755e-07, + "loss": 0.6364, + "step": 7573 + }, + { + "epoch": 14.025925925925925, + "grad_norm": 1.9198521254907177, + "learning_rate": 1.5075125237878505e-07, + "loss": 0.5933, + "step": 7574 + }, + { + "epoch": 14.027777777777779, + "grad_norm": 1.5941251095107105, + "learning_rate": 1.5073833210778907e-07, + "loss": 0.5754, + "step": 7575 + }, + { + "epoch": 14.02962962962963, + "grad_norm": 2.0927889274354974, + "learning_rate": 1.5072541069605764e-07, + "loss": 0.6054, + "step": 7576 + }, + { + "epoch": 14.031481481481482, + "grad_norm": 2.1084662846595412, + "learning_rate": 1.5071248814388118e-07, + "loss": 0.6371, + "step": 7577 + }, + { + "epoch": 14.033333333333333, + "grad_norm": 1.598864436970099, + "learning_rate": 1.5069956445155025e-07, + "loss": 0.6429, + "step": 7578 + }, + { + "epoch": 14.035185185185185, + "grad_norm": 2.225775838516352, + "learning_rate": 1.506866396193554e-07, + "loss": 0.5802, + "step": 7579 + }, + { + "epoch": 14.037037037037036, + "grad_norm": 1.4592249300620121, + "learning_rate": 1.5067371364758724e-07, + "loss": 0.5609, + "step": 7580 + }, + { + "epoch": 14.03888888888889, + "grad_norm": 2.2113102484368277, + "learning_rate": 1.506607865365364e-07, + "loss": 0.6126, + "step": 7581 + }, + { + "epoch": 14.040740740740741, + "grad_norm": 1.5418281542151708, + "learning_rate": 1.506478582864934e-07, + "loss": 0.6794, + "step": 7582 + }, + { + "epoch": 14.042592592592593, + "grad_norm": 1.7288794564584509, + "learning_rate": 1.506349288977491e-07, + "loss": 0.6055, + "step": 7583 + }, + { + "epoch": 14.044444444444444, + "grad_norm": 1.9271581446159896, + "learning_rate": 1.5062199837059404e-07, + "loss": 0.5879, + "step": 7584 + }, + { + "epoch": 14.046296296296296, + "grad_norm": 1.6954936895958168, + "learning_rate": 1.5060906670531893e-07, + "loss": 0.6039, + "step": 7585 + }, + { + "epoch": 14.048148148148147, + "grad_norm": 1.4981864965486094, + "learning_rate": 1.5059613390221456e-07, + "loss": 0.5947, + "step": 7586 + }, + { + "epoch": 14.05, + "grad_norm": 1.5304071770474366, + "learning_rate": 1.505831999615717e-07, + "loss": 0.5875, + "step": 7587 + }, + { + "epoch": 14.051851851851852, + "grad_norm": 1.7913263733850884, + "learning_rate": 1.505702648836811e-07, + "loss": 0.6424, + "step": 7588 + }, + { + "epoch": 14.053703703703704, + "grad_norm": 1.6768149640514025, + "learning_rate": 1.5055732866883361e-07, + "loss": 0.6334, + "step": 7589 + }, + { + "epoch": 14.055555555555555, + "grad_norm": 1.441874454108305, + "learning_rate": 1.5054439131732008e-07, + "loss": 0.5887, + "step": 7590 + }, + { + "epoch": 14.057407407407407, + "grad_norm": 1.473412943784779, + "learning_rate": 1.5053145282943135e-07, + "loss": 0.6247, + "step": 7591 + }, + { + "epoch": 14.059259259259258, + "grad_norm": 1.5871441973868543, + "learning_rate": 1.5051851320545833e-07, + "loss": 0.6483, + "step": 7592 + }, + { + "epoch": 14.061111111111112, + "grad_norm": 1.5657553313386878, + "learning_rate": 1.505055724456919e-07, + "loss": 0.6223, + "step": 7593 + }, + { + "epoch": 14.062962962962963, + "grad_norm": 1.6483418326868546, + "learning_rate": 1.50492630550423e-07, + "loss": 0.5793, + "step": 7594 + }, + { + "epoch": 14.064814814814815, + "grad_norm": 6.132313332796144, + "learning_rate": 1.504796875199427e-07, + "loss": 0.6096, + "step": 7595 + }, + { + "epoch": 14.066666666666666, + "grad_norm": 1.6499612497563398, + "learning_rate": 1.5046674335454188e-07, + "loss": 0.6386, + "step": 7596 + }, + { + "epoch": 14.068518518518518, + "grad_norm": 1.5551339861687337, + "learning_rate": 1.504537980545116e-07, + "loss": 0.5956, + "step": 7597 + }, + { + "epoch": 14.07037037037037, + "grad_norm": 1.7670669510990304, + "learning_rate": 1.5044085162014295e-07, + "loss": 0.6107, + "step": 7598 + }, + { + "epoch": 14.072222222222223, + "grad_norm": 1.7605663739524313, + "learning_rate": 1.5042790405172695e-07, + "loss": 0.602, + "step": 7599 + }, + { + "epoch": 14.074074074074074, + "grad_norm": 1.6594233043733004, + "learning_rate": 1.5041495534955465e-07, + "loss": 0.6264, + "step": 7600 + }, + { + "epoch": 14.075925925925926, + "grad_norm": 1.4956276920432126, + "learning_rate": 1.5040200551391727e-07, + "loss": 0.6471, + "step": 7601 + }, + { + "epoch": 14.077777777777778, + "grad_norm": 1.4932250653635204, + "learning_rate": 1.5038905454510596e-07, + "loss": 0.599, + "step": 7602 + }, + { + "epoch": 14.079629629629629, + "grad_norm": 1.7689607643767626, + "learning_rate": 1.5037610244341176e-07, + "loss": 0.5956, + "step": 7603 + }, + { + "epoch": 14.081481481481482, + "grad_norm": 1.436908409289378, + "learning_rate": 1.50363149209126e-07, + "loss": 0.6033, + "step": 7604 + }, + { + "epoch": 14.083333333333334, + "grad_norm": 1.8453985541699853, + "learning_rate": 1.503501948425399e-07, + "loss": 0.6214, + "step": 7605 + }, + { + "epoch": 14.085185185185185, + "grad_norm": 1.4393099057386045, + "learning_rate": 1.5033723934394464e-07, + "loss": 0.5947, + "step": 7606 + }, + { + "epoch": 14.087037037037037, + "grad_norm": 1.6488557182242323, + "learning_rate": 1.503242827136315e-07, + "loss": 0.6292, + "step": 7607 + }, + { + "epoch": 14.088888888888889, + "grad_norm": 1.7530432214019205, + "learning_rate": 1.5031132495189185e-07, + "loss": 0.6218, + "step": 7608 + }, + { + "epoch": 14.09074074074074, + "grad_norm": 2.5422581918690805, + "learning_rate": 1.502983660590169e-07, + "loss": 0.6567, + "step": 7609 + }, + { + "epoch": 14.092592592592593, + "grad_norm": 1.8493467810826807, + "learning_rate": 1.5028540603529815e-07, + "loss": 0.6185, + "step": 7610 + }, + { + "epoch": 14.094444444444445, + "grad_norm": 1.4943640633340753, + "learning_rate": 1.5027244488102687e-07, + "loss": 0.6216, + "step": 7611 + }, + { + "epoch": 14.096296296296297, + "grad_norm": 1.8210036901950928, + "learning_rate": 1.502594825964945e-07, + "loss": 0.5988, + "step": 7612 + }, + { + "epoch": 14.098148148148148, + "grad_norm": 1.555931715786282, + "learning_rate": 1.5024651918199242e-07, + "loss": 0.6004, + "step": 7613 + }, + { + "epoch": 14.1, + "grad_norm": 1.7247720858698312, + "learning_rate": 1.502335546378122e-07, + "loss": 0.6045, + "step": 7614 + }, + { + "epoch": 14.101851851851851, + "grad_norm": 1.4646727458798356, + "learning_rate": 1.5022058896424517e-07, + "loss": 0.621, + "step": 7615 + }, + { + "epoch": 14.103703703703705, + "grad_norm": 1.4953265761874894, + "learning_rate": 1.5020762216158294e-07, + "loss": 0.6064, + "step": 7616 + }, + { + "epoch": 14.105555555555556, + "grad_norm": 1.5197029256162708, + "learning_rate": 1.5019465423011698e-07, + "loss": 0.6288, + "step": 7617 + }, + { + "epoch": 14.107407407407408, + "grad_norm": 1.6198709381597964, + "learning_rate": 1.5018168517013886e-07, + "loss": 0.6414, + "step": 7618 + }, + { + "epoch": 14.10925925925926, + "grad_norm": 1.5377842106867432, + "learning_rate": 1.5016871498194017e-07, + "loss": 0.6465, + "step": 7619 + }, + { + "epoch": 14.11111111111111, + "grad_norm": 2.04785401176347, + "learning_rate": 1.5015574366581253e-07, + "loss": 0.6183, + "step": 7620 + }, + { + "epoch": 14.112962962962962, + "grad_norm": 1.5187260781190668, + "learning_rate": 1.5014277122204756e-07, + "loss": 0.6024, + "step": 7621 + }, + { + "epoch": 14.114814814814816, + "grad_norm": 1.7489821246163195, + "learning_rate": 1.501297976509369e-07, + "loss": 0.6031, + "step": 7622 + }, + { + "epoch": 14.116666666666667, + "grad_norm": 2.487700463693705, + "learning_rate": 1.501168229527722e-07, + "loss": 0.6167, + "step": 7623 + }, + { + "epoch": 14.118518518518519, + "grad_norm": 1.614565649548583, + "learning_rate": 1.5010384712784523e-07, + "loss": 0.6251, + "step": 7624 + }, + { + "epoch": 14.12037037037037, + "grad_norm": 1.6249839690387244, + "learning_rate": 1.500908701764477e-07, + "loss": 0.6033, + "step": 7625 + }, + { + "epoch": 14.122222222222222, + "grad_norm": 1.9488993266066592, + "learning_rate": 1.5007789209887136e-07, + "loss": 0.6536, + "step": 7626 + }, + { + "epoch": 14.124074074074073, + "grad_norm": 1.5632818709300507, + "learning_rate": 1.50064912895408e-07, + "loss": 0.5731, + "step": 7627 + }, + { + "epoch": 14.125925925925927, + "grad_norm": 1.5443205092160375, + "learning_rate": 1.500519325663494e-07, + "loss": 0.5885, + "step": 7628 + }, + { + "epoch": 14.127777777777778, + "grad_norm": 1.9428723937062082, + "learning_rate": 1.500389511119875e-07, + "loss": 0.5547, + "step": 7629 + }, + { + "epoch": 14.12962962962963, + "grad_norm": 1.520061206649567, + "learning_rate": 1.5002596853261399e-07, + "loss": 0.5838, + "step": 7630 + }, + { + "epoch": 14.131481481481481, + "grad_norm": 1.7347340981983514, + "learning_rate": 1.5001298482852086e-07, + "loss": 0.5738, + "step": 7631 + }, + { + "epoch": 14.133333333333333, + "grad_norm": 3.684925826045263, + "learning_rate": 1.5e-07, + "loss": 0.6324, + "step": 7632 + }, + { + "epoch": 14.135185185185184, + "grad_norm": 2.434272090130225, + "learning_rate": 1.4998701404734332e-07, + "loss": 0.5837, + "step": 7633 + }, + { + "epoch": 14.137037037037038, + "grad_norm": 1.492139275235571, + "learning_rate": 1.4997402697084286e-07, + "loss": 0.5885, + "step": 7634 + }, + { + "epoch": 14.13888888888889, + "grad_norm": 1.5201663881417409, + "learning_rate": 1.499610387707905e-07, + "loss": 0.6163, + "step": 7635 + }, + { + "epoch": 14.14074074074074, + "grad_norm": 1.3930291664301588, + "learning_rate": 1.499480494474783e-07, + "loss": 0.5987, + "step": 7636 + }, + { + "epoch": 14.142592592592592, + "grad_norm": 1.7987670500985442, + "learning_rate": 1.4993505900119831e-07, + "loss": 0.6126, + "step": 7637 + }, + { + "epoch": 14.144444444444444, + "grad_norm": 1.597447549399388, + "learning_rate": 1.4992206743224257e-07, + "loss": 0.5895, + "step": 7638 + }, + { + "epoch": 14.146296296296295, + "grad_norm": 1.5927572520024187, + "learning_rate": 1.4990907474090313e-07, + "loss": 0.6414, + "step": 7639 + }, + { + "epoch": 14.148148148148149, + "grad_norm": 1.5592852943566924, + "learning_rate": 1.4989608092747218e-07, + "loss": 0.6204, + "step": 7640 + }, + { + "epoch": 14.15, + "grad_norm": 2.1378556212453197, + "learning_rate": 1.498830859922418e-07, + "loss": 0.5974, + "step": 7641 + }, + { + "epoch": 14.151851851851852, + "grad_norm": 1.9946481489879193, + "learning_rate": 1.4987008993550417e-07, + "loss": 0.6215, + "step": 7642 + }, + { + "epoch": 14.153703703703703, + "grad_norm": 1.6028189793660668, + "learning_rate": 1.4985709275755147e-07, + "loss": 0.6377, + "step": 7643 + }, + { + "epoch": 14.155555555555555, + "grad_norm": 1.4314696429646245, + "learning_rate": 1.4984409445867594e-07, + "loss": 0.6332, + "step": 7644 + }, + { + "epoch": 14.157407407407407, + "grad_norm": 1.561400966355092, + "learning_rate": 1.4983109503916976e-07, + "loss": 0.6147, + "step": 7645 + }, + { + "epoch": 14.15925925925926, + "grad_norm": 1.6507683150564743, + "learning_rate": 1.4981809449932523e-07, + "loss": 0.6264, + "step": 7646 + }, + { + "epoch": 14.161111111111111, + "grad_norm": 1.515116871958776, + "learning_rate": 1.4980509283943464e-07, + "loss": 0.6147, + "step": 7647 + }, + { + "epoch": 14.162962962962963, + "grad_norm": 1.6215723547405965, + "learning_rate": 1.4979209005979029e-07, + "loss": 0.6058, + "step": 7648 + }, + { + "epoch": 14.164814814814815, + "grad_norm": 1.733154360280619, + "learning_rate": 1.497790861606845e-07, + "loss": 0.6051, + "step": 7649 + }, + { + "epoch": 14.166666666666666, + "grad_norm": 1.4438849647600247, + "learning_rate": 1.4976608114240969e-07, + "loss": 0.6082, + "step": 7650 + }, + { + "epoch": 14.168518518518518, + "grad_norm": 1.7901110483185685, + "learning_rate": 1.497530750052582e-07, + "loss": 0.6275, + "step": 7651 + }, + { + "epoch": 14.170370370370371, + "grad_norm": 1.5358185602227856, + "learning_rate": 1.4974006774952246e-07, + "loss": 0.6168, + "step": 7652 + }, + { + "epoch": 14.172222222222222, + "grad_norm": 1.8131998431280767, + "learning_rate": 1.497270593754949e-07, + "loss": 0.6791, + "step": 7653 + }, + { + "epoch": 14.174074074074074, + "grad_norm": 1.7169095142064215, + "learning_rate": 1.49714049883468e-07, + "loss": 0.6135, + "step": 7654 + }, + { + "epoch": 14.175925925925926, + "grad_norm": 1.4135901127366672, + "learning_rate": 1.4970103927373422e-07, + "loss": 0.5671, + "step": 7655 + }, + { + "epoch": 14.177777777777777, + "grad_norm": 1.756800089014376, + "learning_rate": 1.4968802754658613e-07, + "loss": 0.6248, + "step": 7656 + }, + { + "epoch": 14.17962962962963, + "grad_norm": 1.5923061937883074, + "learning_rate": 1.4967501470231617e-07, + "loss": 0.6283, + "step": 7657 + }, + { + "epoch": 14.181481481481482, + "grad_norm": 2.833276024868875, + "learning_rate": 1.4966200074121697e-07, + "loss": 0.5987, + "step": 7658 + }, + { + "epoch": 14.183333333333334, + "grad_norm": 1.7511325232950894, + "learning_rate": 1.4964898566358114e-07, + "loss": 0.5746, + "step": 7659 + }, + { + "epoch": 14.185185185185185, + "grad_norm": 1.4061610290079498, + "learning_rate": 1.4963596946970126e-07, + "loss": 0.6166, + "step": 7660 + }, + { + "epoch": 14.187037037037037, + "grad_norm": 1.8981733129510918, + "learning_rate": 1.4962295215987003e-07, + "loss": 0.6297, + "step": 7661 + }, + { + "epoch": 14.188888888888888, + "grad_norm": 1.6293501022419705, + "learning_rate": 1.4960993373438e-07, + "loss": 0.5681, + "step": 7662 + }, + { + "epoch": 14.190740740740742, + "grad_norm": 1.8347079229215582, + "learning_rate": 1.495969141935239e-07, + "loss": 0.6091, + "step": 7663 + }, + { + "epoch": 14.192592592592593, + "grad_norm": 1.5014665760350658, + "learning_rate": 1.495838935375945e-07, + "loss": 0.6041, + "step": 7664 + }, + { + "epoch": 14.194444444444445, + "grad_norm": 1.8858033223748825, + "learning_rate": 1.495708717668845e-07, + "loss": 0.6614, + "step": 7665 + }, + { + "epoch": 14.196296296296296, + "grad_norm": 6.74372532141603, + "learning_rate": 1.495578488816867e-07, + "loss": 0.6533, + "step": 7666 + }, + { + "epoch": 14.198148148148148, + "grad_norm": 1.409799488030474, + "learning_rate": 1.4954482488229385e-07, + "loss": 0.5977, + "step": 7667 + }, + { + "epoch": 14.2, + "grad_norm": 2.085750659052171, + "learning_rate": 1.4953179976899877e-07, + "loss": 0.6156, + "step": 7668 + }, + { + "epoch": 14.201851851851853, + "grad_norm": 1.6764833276514248, + "learning_rate": 1.4951877354209428e-07, + "loss": 0.6273, + "step": 7669 + }, + { + "epoch": 14.203703703703704, + "grad_norm": 1.83934575450301, + "learning_rate": 1.4950574620187328e-07, + "loss": 0.5682, + "step": 7670 + }, + { + "epoch": 14.205555555555556, + "grad_norm": 1.4636627657960086, + "learning_rate": 1.4949271774862866e-07, + "loss": 0.6106, + "step": 7671 + }, + { + "epoch": 14.207407407407407, + "grad_norm": 1.8500699615407037, + "learning_rate": 1.494796881826533e-07, + "loss": 0.5924, + "step": 7672 + }, + { + "epoch": 14.209259259259259, + "grad_norm": 1.561324444249611, + "learning_rate": 1.494666575042402e-07, + "loss": 0.562, + "step": 7673 + }, + { + "epoch": 14.21111111111111, + "grad_norm": 1.4859167818717407, + "learning_rate": 1.4945362571368227e-07, + "loss": 0.62, + "step": 7674 + }, + { + "epoch": 14.212962962962964, + "grad_norm": 1.5950125881315613, + "learning_rate": 1.4944059281127254e-07, + "loss": 0.6099, + "step": 7675 + }, + { + "epoch": 14.214814814814815, + "grad_norm": 1.4518940386902717, + "learning_rate": 1.49427558797304e-07, + "loss": 0.6653, + "step": 7676 + }, + { + "epoch": 14.216666666666667, + "grad_norm": 1.8477784907743204, + "learning_rate": 1.4941452367206967e-07, + "loss": 0.5994, + "step": 7677 + }, + { + "epoch": 14.218518518518518, + "grad_norm": 1.4786489051890814, + "learning_rate": 1.4940148743586266e-07, + "loss": 0.6161, + "step": 7678 + }, + { + "epoch": 14.22037037037037, + "grad_norm": 1.7293487452444025, + "learning_rate": 1.4938845008897603e-07, + "loss": 0.6041, + "step": 7679 + }, + { + "epoch": 14.222222222222221, + "grad_norm": 1.5791788974353476, + "learning_rate": 1.493754116317029e-07, + "loss": 0.5896, + "step": 7680 + }, + { + "epoch": 14.224074074074075, + "grad_norm": 1.6171428519136037, + "learning_rate": 1.4936237206433645e-07, + "loss": 0.5953, + "step": 7681 + }, + { + "epoch": 14.225925925925926, + "grad_norm": 1.940828537753481, + "learning_rate": 1.493493313871698e-07, + "loss": 0.6346, + "step": 7682 + }, + { + "epoch": 14.227777777777778, + "grad_norm": 1.6527453259714904, + "learning_rate": 1.4933628960049612e-07, + "loss": 0.6111, + "step": 7683 + }, + { + "epoch": 14.22962962962963, + "grad_norm": 1.5722399279352006, + "learning_rate": 1.4932324670460866e-07, + "loss": 0.6197, + "step": 7684 + }, + { + "epoch": 14.231481481481481, + "grad_norm": 1.554248635050263, + "learning_rate": 1.4931020269980067e-07, + "loss": 0.5845, + "step": 7685 + }, + { + "epoch": 14.233333333333333, + "grad_norm": 1.5771759511127588, + "learning_rate": 1.492971575863654e-07, + "loss": 0.6477, + "step": 7686 + }, + { + "epoch": 14.235185185185186, + "grad_norm": 1.4677889257671481, + "learning_rate": 1.4928411136459608e-07, + "loss": 0.6339, + "step": 7687 + }, + { + "epoch": 14.237037037037037, + "grad_norm": 2.478884288984545, + "learning_rate": 1.4927106403478614e-07, + "loss": 0.5941, + "step": 7688 + }, + { + "epoch": 14.238888888888889, + "grad_norm": 2.1931546077170383, + "learning_rate": 1.4925801559722888e-07, + "loss": 0.6825, + "step": 7689 + }, + { + "epoch": 14.24074074074074, + "grad_norm": 1.6656675026051615, + "learning_rate": 1.4924496605221762e-07, + "loss": 0.5905, + "step": 7690 + }, + { + "epoch": 14.242592592592592, + "grad_norm": 1.4647836954121627, + "learning_rate": 1.4923191540004578e-07, + "loss": 0.6287, + "step": 7691 + }, + { + "epoch": 14.244444444444444, + "grad_norm": 1.5100096469472872, + "learning_rate": 1.4921886364100675e-07, + "loss": 0.6045, + "step": 7692 + }, + { + "epoch": 14.246296296296297, + "grad_norm": 1.7365110390741048, + "learning_rate": 1.4920581077539402e-07, + "loss": 0.6124, + "step": 7693 + }, + { + "epoch": 14.248148148148148, + "grad_norm": 1.5762062344180878, + "learning_rate": 1.49192756803501e-07, + "loss": 0.6256, + "step": 7694 + }, + { + "epoch": 14.25, + "grad_norm": 1.4857106470645267, + "learning_rate": 1.491797017256212e-07, + "loss": 0.6127, + "step": 7695 + }, + { + "epoch": 14.251851851851852, + "grad_norm": 1.5902289624051167, + "learning_rate": 1.4916664554204816e-07, + "loss": 0.5948, + "step": 7696 + }, + { + "epoch": 14.253703703703703, + "grad_norm": 2.7200043247983645, + "learning_rate": 1.491535882530754e-07, + "loss": 0.6158, + "step": 7697 + }, + { + "epoch": 14.255555555555556, + "grad_norm": 1.5649669819434802, + "learning_rate": 1.4914052985899645e-07, + "loss": 0.6054, + "step": 7698 + }, + { + "epoch": 14.257407407407408, + "grad_norm": 1.432775402247199, + "learning_rate": 1.4912747036010493e-07, + "loss": 0.6169, + "step": 7699 + }, + { + "epoch": 14.25925925925926, + "grad_norm": 1.4562983326941146, + "learning_rate": 1.4911440975669444e-07, + "loss": 0.6148, + "step": 7700 + } + ], + "logging_steps": 1.0, + "max_steps": 21600, + "num_input_tokens_seen": 0, + "num_train_epochs": 40, + "save_steps": 50, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 6700262394036224.0, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +}