{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.7645305140795671, "eval_steps": 500, "global_step": 30517, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.000776630924942379, "grad_norm": 23.55321502685547, "learning_rate": 1.0157273918741808e-06, "loss": 8.3443, "step": 31 }, { "epoch": 0.001553261849884758, "grad_norm": 14.414477348327637, "learning_rate": 2.0314547837483616e-06, "loss": 7.198, "step": 62 }, { "epoch": 0.002329892774827137, "grad_norm": 10.739773750305176, "learning_rate": 3.0471821756225426e-06, "loss": 5.8639, "step": 93 }, { "epoch": 0.003106523699769516, "grad_norm": 5.378795623779297, "learning_rate": 4.062909567496723e-06, "loss": 4.8222, "step": 124 }, { "epoch": 0.003883154624711895, "grad_norm": 8.51831340789795, "learning_rate": 5.078636959370905e-06, "loss": 4.1983, "step": 155 }, { "epoch": 0.004659785549654274, "grad_norm": 6.32632303237915, "learning_rate": 6.094364351245085e-06, "loss": 3.8598, "step": 186 }, { "epoch": 0.005436416474596653, "grad_norm": 5.590794563293457, "learning_rate": 7.110091743119267e-06, "loss": 3.6646, "step": 217 }, { "epoch": 0.006213047399539032, "grad_norm": 5.9275383949279785, "learning_rate": 8.125819134993446e-06, "loss": 3.5212, "step": 248 }, { "epoch": 0.006989678324481411, "grad_norm": 5.041291236877441, "learning_rate": 9.141546526867629e-06, "loss": 3.3936, "step": 279 }, { "epoch": 0.00776630924942379, "grad_norm": 4.923459529876709, "learning_rate": 1.015727391874181e-05, "loss": 3.2713, "step": 310 }, { "epoch": 0.00854294017436617, "grad_norm": 8.011459350585938, "learning_rate": 1.117300131061599e-05, "loss": 3.2, "step": 341 }, { "epoch": 0.009319571099308548, "grad_norm": 5.261505603790283, "learning_rate": 1.218872870249017e-05, "loss": 3.1255, "step": 372 }, { "epoch": 0.010096202024250926, "grad_norm": 5.004406452178955, "learning_rate": 1.3204456094364351e-05, "loss": 3.0726, "step": 403 }, { "epoch": 0.010872832949193306, "grad_norm": 4.326263427734375, "learning_rate": 1.4220183486238533e-05, "loss": 2.9682, "step": 434 }, { "epoch": 0.011649463874135685, "grad_norm": 4.251937389373779, "learning_rate": 1.5235910878112714e-05, "loss": 2.9215, "step": 465 }, { "epoch": 0.012426094799078065, "grad_norm": 4.35053825378418, "learning_rate": 1.6251638269986893e-05, "loss": 2.8902, "step": 496 }, { "epoch": 0.013202725724020443, "grad_norm": 4.316098690032959, "learning_rate": 1.7267365661861077e-05, "loss": 2.8521, "step": 527 }, { "epoch": 0.013979356648962823, "grad_norm": 5.536881446838379, "learning_rate": 1.8283093053735257e-05, "loss": 2.8058, "step": 558 }, { "epoch": 0.014755987573905201, "grad_norm": 3.759843111038208, "learning_rate": 1.9298820445609438e-05, "loss": 2.7438, "step": 589 }, { "epoch": 0.01553261849884758, "grad_norm": 4.057703018188477, "learning_rate": 2.031454783748362e-05, "loss": 2.7508, "step": 620 }, { "epoch": 0.01630924942378996, "grad_norm": 3.427708625793457, "learning_rate": 2.13302752293578e-05, "loss": 2.7046, "step": 651 }, { "epoch": 0.01708588034873234, "grad_norm": 3.9012484550476074, "learning_rate": 2.234600262123198e-05, "loss": 2.6706, "step": 682 }, { "epoch": 0.017862511273674716, "grad_norm": 3.388504981994629, "learning_rate": 2.336173001310616e-05, "loss": 2.6377, "step": 713 }, { "epoch": 0.018639142198617096, "grad_norm": 3.2011327743530273, "learning_rate": 2.437745740498034e-05, "loss": 2.6214, "step": 744 }, { "epoch": 0.019415773123559476, "grad_norm": 3.5897209644317627, "learning_rate": 2.5393184796854525e-05, "loss": 2.6045, "step": 775 }, { "epoch": 0.020192404048501852, "grad_norm": 3.386697769165039, "learning_rate": 2.6408912188728702e-05, "loss": 2.6066, "step": 806 }, { "epoch": 0.020969034973444232, "grad_norm": 2.9135866165161133, "learning_rate": 2.7424639580602886e-05, "loss": 2.5697, "step": 837 }, { "epoch": 0.021745665898386612, "grad_norm": 9.212065696716309, "learning_rate": 2.8440366972477066e-05, "loss": 2.5773, "step": 868 }, { "epoch": 0.022522296823328992, "grad_norm": 3.0339765548706055, "learning_rate": 2.9456094364351244e-05, "loss": 2.5453, "step": 899 }, { "epoch": 0.02329892774827137, "grad_norm": 2.8007607460021973, "learning_rate": 3.0471821756225428e-05, "loss": 2.5297, "step": 930 }, { "epoch": 0.02407555867321375, "grad_norm": 3.251768112182617, "learning_rate": 3.148754914809961e-05, "loss": 2.5273, "step": 961 }, { "epoch": 0.02485218959815613, "grad_norm": 3.357011079788208, "learning_rate": 3.2503276539973785e-05, "loss": 2.4896, "step": 992 }, { "epoch": 0.025628820523098506, "grad_norm": 2.6069107055664062, "learning_rate": 3.351900393184797e-05, "loss": 2.4718, "step": 1023 }, { "epoch": 0.026405451448040886, "grad_norm": 4.663620948791504, "learning_rate": 3.453473132372215e-05, "loss": 2.4338, "step": 1054 }, { "epoch": 0.027182082372983266, "grad_norm": 3.819638252258301, "learning_rate": 3.555045871559633e-05, "loss": 2.4848, "step": 1085 }, { "epoch": 0.027958713297925646, "grad_norm": 2.6367456912994385, "learning_rate": 3.6566186107470514e-05, "loss": 2.4415, "step": 1116 }, { "epoch": 0.028735344222868022, "grad_norm": 2.6118173599243164, "learning_rate": 3.7581913499344695e-05, "loss": 2.4341, "step": 1147 }, { "epoch": 0.029511975147810402, "grad_norm": 9.820366859436035, "learning_rate": 3.8597640891218876e-05, "loss": 2.3773, "step": 1178 }, { "epoch": 0.030288606072752782, "grad_norm": 38.890506744384766, "learning_rate": 3.9613368283093056e-05, "loss": 2.3966, "step": 1209 }, { "epoch": 0.03106523699769516, "grad_norm": 2.5037872791290283, "learning_rate": 4.062909567496724e-05, "loss": 2.3421, "step": 1240 }, { "epoch": 0.03184186792263754, "grad_norm": 2.426544189453125, "learning_rate": 4.164482306684142e-05, "loss": 2.3568, "step": 1271 }, { "epoch": 0.03261849884757992, "grad_norm": 2.5375614166259766, "learning_rate": 4.26605504587156e-05, "loss": 2.3314, "step": 1302 }, { "epoch": 0.0333951297725223, "grad_norm": 3.555330276489258, "learning_rate": 4.367627785058978e-05, "loss": 2.3262, "step": 1333 }, { "epoch": 0.03417176069746468, "grad_norm": 2.2571263313293457, "learning_rate": 4.469200524246396e-05, "loss": 2.3167, "step": 1364 }, { "epoch": 0.03494839162240705, "grad_norm": 2.184305429458618, "learning_rate": 4.570773263433814e-05, "loss": 2.2936, "step": 1395 }, { "epoch": 0.03572502254734943, "grad_norm": 2.8295061588287354, "learning_rate": 4.672346002621232e-05, "loss": 2.3033, "step": 1426 }, { "epoch": 0.03650165347229181, "grad_norm": 2.0241498947143555, "learning_rate": 4.77391874180865e-05, "loss": 2.2833, "step": 1457 }, { "epoch": 0.03727828439723419, "grad_norm": 2.0029919147491455, "learning_rate": 4.875491480996068e-05, "loss": 2.2663, "step": 1488 }, { "epoch": 0.03805491532217657, "grad_norm": 2.033773899078369, "learning_rate": 4.977064220183487e-05, "loss": 2.2597, "step": 1519 }, { "epoch": 0.03883154624711895, "grad_norm": 2.047210216522217, "learning_rate": 4.9999915451558777e-05, "loss": 2.2231, "step": 1550 }, { "epoch": 0.03960817717206133, "grad_norm": 2.032099485397339, "learning_rate": 4.999955597496219e-05, "loss": 2.2411, "step": 1581 }, { "epoch": 0.040384808097003705, "grad_norm": 1.8384301662445068, "learning_rate": 4.9998914381774255e-05, "loss": 2.2409, "step": 1612 }, { "epoch": 0.041161439021946085, "grad_norm": 1.721187710762024, "learning_rate": 4.999799067923527e-05, "loss": 2.193, "step": 1643 }, { "epoch": 0.041938069946888465, "grad_norm": 1.8869743347167969, "learning_rate": 4.999678487776908e-05, "loss": 2.1917, "step": 1674 }, { "epoch": 0.042714700871830845, "grad_norm": 1.796583652496338, "learning_rate": 4.9995296990983006e-05, "loss": 2.1794, "step": 1705 }, { "epoch": 0.043491331796773225, "grad_norm": 1.6308804750442505, "learning_rate": 4.999352703566763e-05, "loss": 2.1854, "step": 1736 }, { "epoch": 0.044267962721715605, "grad_norm": 1.683474063873291, "learning_rate": 4.999147503179668e-05, "loss": 2.1797, "step": 1767 }, { "epoch": 0.045044593646657985, "grad_norm": 1.7442835569381714, "learning_rate": 4.998914100252672e-05, "loss": 2.1695, "step": 1798 }, { "epoch": 0.04582122457160036, "grad_norm": 1.704970359802246, "learning_rate": 4.998652497419696e-05, "loss": 2.1365, "step": 1829 }, { "epoch": 0.04659785549654274, "grad_norm": 1.7213023900985718, "learning_rate": 4.9983626976328927e-05, "loss": 2.1524, "step": 1860 }, { "epoch": 0.04737448642148512, "grad_norm": 4.01851224899292, "learning_rate": 4.998044704162613e-05, "loss": 2.1171, "step": 1891 }, { "epoch": 0.0481511173464275, "grad_norm": 1.5619858503341675, "learning_rate": 4.9976985205973705e-05, "loss": 2.1608, "step": 1922 }, { "epoch": 0.04892774827136988, "grad_norm": 1.684493064880371, "learning_rate": 4.997324150843799e-05, "loss": 2.1278, "step": 1953 }, { "epoch": 0.04970437919631226, "grad_norm": 1.660992980003357, "learning_rate": 4.99692159912661e-05, "loss": 2.1136, "step": 1984 }, { "epoch": 0.05048101012125464, "grad_norm": 1.5491065979003906, "learning_rate": 4.996490869988546e-05, "loss": 2.1086, "step": 2015 }, { "epoch": 0.05125764104619701, "grad_norm": 1.5400840044021606, "learning_rate": 4.996031968290326e-05, "loss": 2.1069, "step": 2046 }, { "epoch": 0.05203427197113939, "grad_norm": 1.5687416791915894, "learning_rate": 4.995544899210594e-05, "loss": 2.0888, "step": 2077 }, { "epoch": 0.05281090289608177, "grad_norm": 1.5218935012817383, "learning_rate": 4.9950296682458583e-05, "loss": 2.1032, "step": 2108 }, { "epoch": 0.05358753382102415, "grad_norm": 1.763666033744812, "learning_rate": 4.994486281210429e-05, "loss": 2.0799, "step": 2139 }, { "epoch": 0.05436416474596653, "grad_norm": 1.5606274604797363, "learning_rate": 4.9939147442363566e-05, "loss": 2.0614, "step": 2170 }, { "epoch": 0.05514079567090891, "grad_norm": 1.4900213479995728, "learning_rate": 4.9933150637733574e-05, "loss": 2.0565, "step": 2201 }, { "epoch": 0.05591742659585129, "grad_norm": 1.4454214572906494, "learning_rate": 4.992687246588743e-05, "loss": 2.0381, "step": 2232 }, { "epoch": 0.056694057520793664, "grad_norm": 1.4872742891311646, "learning_rate": 4.992031299767347e-05, "loss": 2.0227, "step": 2263 }, { "epoch": 0.057470688445736044, "grad_norm": 1.4538527727127075, "learning_rate": 4.9913472307114386e-05, "loss": 2.0395, "step": 2294 }, { "epoch": 0.058247319370678424, "grad_norm": 1.432769536972046, "learning_rate": 4.9906350471406446e-05, "loss": 2.0352, "step": 2325 }, { "epoch": 0.059023950295620804, "grad_norm": 1.4535977840423584, "learning_rate": 4.989894757091861e-05, "loss": 2.0341, "step": 2356 }, { "epoch": 0.059800581220563184, "grad_norm": 1.4004398584365845, "learning_rate": 4.989126368919158e-05, "loss": 2.0185, "step": 2387 }, { "epoch": 0.060577212145505564, "grad_norm": 1.433273434638977, "learning_rate": 4.988329891293693e-05, "loss": 2.036, "step": 2418 }, { "epoch": 0.06135384307044794, "grad_norm": 1.3791290521621704, "learning_rate": 4.987505333203608e-05, "loss": 2.0421, "step": 2449 }, { "epoch": 0.06213047399539032, "grad_norm": 2.008694648742676, "learning_rate": 4.9866527039539276e-05, "loss": 1.9948, "step": 2480 }, { "epoch": 0.0629071049203327, "grad_norm": 1.3773164749145508, "learning_rate": 4.9857720131664594e-05, "loss": 2.0158, "step": 2511 }, { "epoch": 0.06368373584527508, "grad_norm": 1.4489777088165283, "learning_rate": 4.9848632707796773e-05, "loss": 2.0216, "step": 2542 }, { "epoch": 0.06446036677021745, "grad_norm": 2.1266908645629883, "learning_rate": 4.9839264870486155e-05, "loss": 1.9962, "step": 2573 }, { "epoch": 0.06523699769515984, "grad_norm": 1.3704332113265991, "learning_rate": 4.9829616725447526e-05, "loss": 1.9801, "step": 2604 }, { "epoch": 0.06601362862010221, "grad_norm": 1.3643279075622559, "learning_rate": 4.981968838155888e-05, "loss": 1.9902, "step": 2635 }, { "epoch": 0.0667902595450446, "grad_norm": 1.3495409488677979, "learning_rate": 4.980947995086024e-05, "loss": 1.993, "step": 2666 }, { "epoch": 0.06756689046998697, "grad_norm": 1.3918874263763428, "learning_rate": 4.979899154855234e-05, "loss": 1.9782, "step": 2697 }, { "epoch": 0.06834352139492936, "grad_norm": 1.3205535411834717, "learning_rate": 4.9788223292995386e-05, "loss": 1.9638, "step": 2728 }, { "epoch": 0.06912015231987173, "grad_norm": 1.3123114109039307, "learning_rate": 4.977717530570768e-05, "loss": 1.9826, "step": 2759 }, { "epoch": 0.0698967832448141, "grad_norm": 1.3138995170593262, "learning_rate": 4.976584771136425e-05, "loss": 1.9938, "step": 2790 }, { "epoch": 0.07067341416975649, "grad_norm": 1.2964009046554565, "learning_rate": 4.975424063779547e-05, "loss": 1.9924, "step": 2821 }, { "epoch": 0.07145004509469886, "grad_norm": 1.3010739088058472, "learning_rate": 4.974235421598557e-05, "loss": 1.9592, "step": 2852 }, { "epoch": 0.07222667601964125, "grad_norm": 1.2549103498458862, "learning_rate": 4.973018858007122e-05, "loss": 1.9394, "step": 2883 }, { "epoch": 0.07300330694458362, "grad_norm": 1.3198585510253906, "learning_rate": 4.9717743867339963e-05, "loss": 1.989, "step": 2914 }, { "epoch": 0.07377993786952601, "grad_norm": 1.3653398752212524, "learning_rate": 4.9705020218228695e-05, "loss": 1.9499, "step": 2945 }, { "epoch": 0.07455656879446838, "grad_norm": 4.606905937194824, "learning_rate": 4.969201777632205e-05, "loss": 1.9504, "step": 2976 }, { "epoch": 0.07533319971941076, "grad_norm": 1.2823129892349243, "learning_rate": 4.9678736688350846e-05, "loss": 1.9483, "step": 3007 }, { "epoch": 0.07610983064435314, "grad_norm": 1.2803308963775635, "learning_rate": 4.966517710419033e-05, "loss": 1.9604, "step": 3038 }, { "epoch": 0.07688646156929552, "grad_norm": 1.2765967845916748, "learning_rate": 4.965133917685858e-05, "loss": 1.941, "step": 3069 }, { "epoch": 0.0776630924942379, "grad_norm": 2.622286796569824, "learning_rate": 4.9637223062514714e-05, "loss": 1.9426, "step": 3100 }, { "epoch": 0.07843972341918028, "grad_norm": 1.231387972831726, "learning_rate": 4.962282892045718e-05, "loss": 1.9416, "step": 3131 }, { "epoch": 0.07921635434412266, "grad_norm": 1.2499909400939941, "learning_rate": 4.9608156913121904e-05, "loss": 1.9226, "step": 3162 }, { "epoch": 0.07999298526906504, "grad_norm": 1.296276330947876, "learning_rate": 4.959320720608049e-05, "loss": 1.935, "step": 3193 }, { "epoch": 0.08076961619400741, "grad_norm": 1.2642258405685425, "learning_rate": 4.9577979968038354e-05, "loss": 1.9185, "step": 3224 }, { "epoch": 0.0815462471189498, "grad_norm": 1.2396386861801147, "learning_rate": 4.956247537083282e-05, "loss": 1.9361, "step": 3255 }, { "epoch": 0.08232287804389217, "grad_norm": 1.1811271905899048, "learning_rate": 4.9546693589431145e-05, "loss": 1.9276, "step": 3286 }, { "epoch": 0.08309950896883456, "grad_norm": 1.2897577285766602, "learning_rate": 4.9530634801928595e-05, "loss": 1.9277, "step": 3317 }, { "epoch": 0.08387613989377693, "grad_norm": 1.1453903913497925, "learning_rate": 4.9514299189546395e-05, "loss": 1.9109, "step": 3348 }, { "epoch": 0.08465277081871932, "grad_norm": 1.1600992679595947, "learning_rate": 4.949768693662973e-05, "loss": 1.9248, "step": 3379 }, { "epoch": 0.08542940174366169, "grad_norm": 1.234217882156372, "learning_rate": 4.948079823064559e-05, "loss": 1.9236, "step": 3410 }, { "epoch": 0.08620603266860406, "grad_norm": 1.1933766603469849, "learning_rate": 4.946363326218074e-05, "loss": 1.9284, "step": 3441 }, { "epoch": 0.08698266359354645, "grad_norm": 1.2087618112564087, "learning_rate": 4.9446192224939525e-05, "loss": 1.8971, "step": 3472 }, { "epoch": 0.08775929451848882, "grad_norm": 1.227864146232605, "learning_rate": 4.942847531574167e-05, "loss": 1.9353, "step": 3503 }, { "epoch": 0.08853592544343121, "grad_norm": 1.2828229665756226, "learning_rate": 4.941048273452008e-05, "loss": 1.9085, "step": 3534 }, { "epoch": 0.08931255636837358, "grad_norm": 1.2086588144302368, "learning_rate": 4.9392214684318605e-05, "loss": 1.9044, "step": 3565 }, { "epoch": 0.09008918729331597, "grad_norm": 1.1942152976989746, "learning_rate": 4.93736713712897e-05, "loss": 1.9182, "step": 3596 }, { "epoch": 0.09086581821825834, "grad_norm": 1.2399394512176514, "learning_rate": 4.9354853004692124e-05, "loss": 1.8943, "step": 3627 }, { "epoch": 0.09164244914320072, "grad_norm": 1.227376937866211, "learning_rate": 4.93357597968886e-05, "loss": 1.8881, "step": 3658 }, { "epoch": 0.0924190800681431, "grad_norm": 1.1314102411270142, "learning_rate": 4.931639196334338e-05, "loss": 1.8878, "step": 3689 }, { "epoch": 0.09319571099308548, "grad_norm": 1.192134141921997, "learning_rate": 4.9296749722619826e-05, "loss": 1.8857, "step": 3720 }, { "epoch": 0.09397234191802786, "grad_norm": 1.1905503273010254, "learning_rate": 4.9276833296377966e-05, "loss": 1.8701, "step": 3751 }, { "epoch": 0.09474897284297024, "grad_norm": 1.2335184812545776, "learning_rate": 4.925664290937196e-05, "loss": 1.8741, "step": 3782 }, { "epoch": 0.09552560376791262, "grad_norm": 1.1403234004974365, "learning_rate": 4.9236178789447576e-05, "loss": 1.8658, "step": 3813 }, { "epoch": 0.096302234692855, "grad_norm": 1.1734315156936646, "learning_rate": 4.921544116753962e-05, "loss": 1.8724, "step": 3844 }, { "epoch": 0.09707886561779737, "grad_norm": 1.2637231349945068, "learning_rate": 4.919443027766935e-05, "loss": 1.867, "step": 3875 }, { "epoch": 0.09785549654273976, "grad_norm": 1.1608805656433105, "learning_rate": 4.91731463569418e-05, "loss": 1.8764, "step": 3906 }, { "epoch": 0.09863212746768213, "grad_norm": 1.1562029123306274, "learning_rate": 4.915158964554312e-05, "loss": 1.8612, "step": 3937 }, { "epoch": 0.09940875839262452, "grad_norm": 1.1918668746948242, "learning_rate": 4.912976038673786e-05, "loss": 1.866, "step": 3968 }, { "epoch": 0.10018538931756689, "grad_norm": 1.180206537246704, "learning_rate": 4.9107658826866254e-05, "loss": 1.849, "step": 3999 }, { "epoch": 0.10096202024250928, "grad_norm": 1.0747199058532715, "learning_rate": 4.908528521534139e-05, "loss": 1.8547, "step": 4030 }, { "epoch": 0.10173865116745165, "grad_norm": 1.103786826133728, "learning_rate": 4.906263980464644e-05, "loss": 1.864, "step": 4061 }, { "epoch": 0.10251528209239402, "grad_norm": 1.1390912532806396, "learning_rate": 4.903972285033178e-05, "loss": 1.8669, "step": 4092 }, { "epoch": 0.10329191301733641, "grad_norm": 1.1488468647003174, "learning_rate": 4.901653461101213e-05, "loss": 1.8445, "step": 4123 }, { "epoch": 0.10406854394227878, "grad_norm": 1.1589927673339844, "learning_rate": 4.8993075348363626e-05, "loss": 1.8529, "step": 4154 }, { "epoch": 0.10484517486722117, "grad_norm": 1.1793478727340698, "learning_rate": 4.896934532712084e-05, "loss": 1.8818, "step": 4185 }, { "epoch": 0.10562180579216354, "grad_norm": 1.165292739868164, "learning_rate": 4.8945344815073846e-05, "loss": 1.8425, "step": 4216 }, { "epoch": 0.10639843671710593, "grad_norm": 1.105946660041809, "learning_rate": 4.892107408306516e-05, "loss": 1.8327, "step": 4247 }, { "epoch": 0.1071750676420483, "grad_norm": 1.1690527200698853, "learning_rate": 4.889653340498669e-05, "loss": 1.8428, "step": 4278 }, { "epoch": 0.10795169856699068, "grad_norm": 1.1106913089752197, "learning_rate": 4.8871723057776664e-05, "loss": 1.839, "step": 4309 }, { "epoch": 0.10872832949193306, "grad_norm": 1.1590090990066528, "learning_rate": 4.8846643321416476e-05, "loss": 1.8658, "step": 4340 }, { "epoch": 0.10950496041687544, "grad_norm": 1.1315875053405762, "learning_rate": 4.882129447892753e-05, "loss": 1.8282, "step": 4371 }, { "epoch": 0.11028159134181782, "grad_norm": 1.6055009365081787, "learning_rate": 4.8795676816368076e-05, "loss": 1.8294, "step": 4402 }, { "epoch": 0.1110582222667602, "grad_norm": 1.1236201524734497, "learning_rate": 4.876979062282995e-05, "loss": 1.8325, "step": 4433 }, { "epoch": 0.11183485319170258, "grad_norm": 1.0830161571502686, "learning_rate": 4.8743636190435325e-05, "loss": 1.8451, "step": 4464 }, { "epoch": 0.11261148411664496, "grad_norm": 1.1663005352020264, "learning_rate": 4.871721381433344e-05, "loss": 1.8349, "step": 4495 }, { "epoch": 0.11338811504158733, "grad_norm": 1.2522406578063965, "learning_rate": 4.869052379269719e-05, "loss": 1.8162, "step": 4526 }, { "epoch": 0.11416474596652972, "grad_norm": 1.1271142959594727, "learning_rate": 4.866356642671985e-05, "loss": 1.8348, "step": 4557 }, { "epoch": 0.11494137689147209, "grad_norm": 1.053753137588501, "learning_rate": 4.8636342020611634e-05, "loss": 1.8376, "step": 4588 }, { "epoch": 0.11571800781641448, "grad_norm": 1.1469056606292725, "learning_rate": 4.860885088159626e-05, "loss": 1.8248, "step": 4619 }, { "epoch": 0.11649463874135685, "grad_norm": 1.116128921508789, "learning_rate": 4.858109331990751e-05, "loss": 1.8454, "step": 4650 }, { "epoch": 0.11727126966629922, "grad_norm": 1.0742651224136353, "learning_rate": 4.855306964878567e-05, "loss": 1.8231, "step": 4681 }, { "epoch": 0.11804790059124161, "grad_norm": 1.101657748222351, "learning_rate": 4.8524780184474084e-05, "loss": 1.8421, "step": 4712 }, { "epoch": 0.11882453151618398, "grad_norm": 1.0891473293304443, "learning_rate": 4.8496225246215496e-05, "loss": 1.8053, "step": 4743 }, { "epoch": 0.11960116244112637, "grad_norm": 1.1044034957885742, "learning_rate": 4.8467405156248505e-05, "loss": 1.8184, "step": 4774 }, { "epoch": 0.12037779336606874, "grad_norm": 1.0644783973693848, "learning_rate": 4.843832023980392e-05, "loss": 1.8237, "step": 4805 }, { "epoch": 0.12115442429101113, "grad_norm": 1.0427252054214478, "learning_rate": 4.840897082510106e-05, "loss": 1.8189, "step": 4836 }, { "epoch": 0.1219310552159535, "grad_norm": 1.0806516408920288, "learning_rate": 4.8379357243344084e-05, "loss": 1.7958, "step": 4867 }, { "epoch": 0.12270768614089587, "grad_norm": 1.092581868171692, "learning_rate": 4.8349479828718236e-05, "loss": 1.8229, "step": 4898 }, { "epoch": 0.12348431706583826, "grad_norm": 1.0858657360076904, "learning_rate": 4.8319338918386075e-05, "loss": 1.8165, "step": 4929 }, { "epoch": 0.12426094799078063, "grad_norm": 1.073818325996399, "learning_rate": 4.828893485248369e-05, "loss": 1.8352, "step": 4960 }, { "epoch": 0.125037578915723, "grad_norm": 1.069009780883789, "learning_rate": 4.825826797411682e-05, "loss": 1.8, "step": 4991 }, { "epoch": 0.1258142098406654, "grad_norm": 1.0416266918182373, "learning_rate": 4.822733862935702e-05, "loss": 1.8224, "step": 5022 }, { "epoch": 0.12659084076560778, "grad_norm": 1.093551516532898, "learning_rate": 4.819614716723775e-05, "loss": 1.8005, "step": 5053 }, { "epoch": 0.12736747169055015, "grad_norm": 1.026353359222412, "learning_rate": 4.8164693939750425e-05, "loss": 1.844, "step": 5084 }, { "epoch": 0.12814410261549253, "grad_norm": 1.102858304977417, "learning_rate": 4.813297930184042e-05, "loss": 1.8045, "step": 5115 }, { "epoch": 0.1289207335404349, "grad_norm": 1.0537272691726685, "learning_rate": 4.810100361140314e-05, "loss": 1.8054, "step": 5146 }, { "epoch": 0.1296973644653773, "grad_norm": 1.0389387607574463, "learning_rate": 4.8068767229279885e-05, "loss": 1.7734, "step": 5177 }, { "epoch": 0.13047399539031967, "grad_norm": 1.0274319648742676, "learning_rate": 4.8036270519253854e-05, "loss": 1.788, "step": 5208 }, { "epoch": 0.13125062631526205, "grad_norm": 1.0932128429412842, "learning_rate": 4.8003513848046e-05, "loss": 1.7813, "step": 5239 }, { "epoch": 0.13202725724020442, "grad_norm": 1.0294640064239502, "learning_rate": 4.79704975853109e-05, "loss": 1.81, "step": 5270 }, { "epoch": 0.13280388816514682, "grad_norm": 1.0617210865020752, "learning_rate": 4.793722210363262e-05, "loss": 1.8017, "step": 5301 }, { "epoch": 0.1335805190900892, "grad_norm": 1.0657798051834106, "learning_rate": 4.7903687778520414e-05, "loss": 1.7884, "step": 5332 }, { "epoch": 0.13435715001503157, "grad_norm": 0.9954230189323425, "learning_rate": 4.7869894988404593e-05, "loss": 1.7848, "step": 5363 }, { "epoch": 0.13513378093997394, "grad_norm": 1.117066740989685, "learning_rate": 4.783584411463221e-05, "loss": 1.7984, "step": 5394 }, { "epoch": 0.13591041186491631, "grad_norm": 1.0706870555877686, "learning_rate": 4.780153554146274e-05, "loss": 1.8008, "step": 5425 }, { "epoch": 0.13668704278985871, "grad_norm": 1.063238501548767, "learning_rate": 4.7766969656063766e-05, "loss": 1.7754, "step": 5456 }, { "epoch": 0.1374636737148011, "grad_norm": 1.0702152252197266, "learning_rate": 4.773214684850662e-05, "loss": 1.8037, "step": 5487 }, { "epoch": 0.13824030463974346, "grad_norm": 1.0318788290023804, "learning_rate": 4.769706751176193e-05, "loss": 1.8174, "step": 5518 }, { "epoch": 0.13901693556468583, "grad_norm": 1.0097755193710327, "learning_rate": 4.7661732041695264e-05, "loss": 1.7648, "step": 5549 }, { "epoch": 0.1397935664896282, "grad_norm": 1.063873291015625, "learning_rate": 4.762614083706258e-05, "loss": 1.809, "step": 5580 }, { "epoch": 0.1405701974145706, "grad_norm": 1.0409623384475708, "learning_rate": 4.759029429950581e-05, "loss": 1.7853, "step": 5611 }, { "epoch": 0.14134682833951298, "grad_norm": 1.0131454467773438, "learning_rate": 4.7554192833548235e-05, "loss": 1.7759, "step": 5642 }, { "epoch": 0.14212345926445535, "grad_norm": 1.1056182384490967, "learning_rate": 4.751783684659e-05, "loss": 1.7959, "step": 5673 }, { "epoch": 0.14290009018939773, "grad_norm": 1.0572400093078613, "learning_rate": 4.748122674890348e-05, "loss": 1.7597, "step": 5704 }, { "epoch": 0.14367672111434013, "grad_norm": 1.0375362634658813, "learning_rate": 4.7444362953628654e-05, "loss": 1.7719, "step": 5735 }, { "epoch": 0.1444533520392825, "grad_norm": 1.0238806009292603, "learning_rate": 4.7407245876768424e-05, "loss": 1.7966, "step": 5766 }, { "epoch": 0.14522998296422487, "grad_norm": 1.0335805416107178, "learning_rate": 4.736987593718397e-05, "loss": 1.7805, "step": 5797 }, { "epoch": 0.14600661388916725, "grad_norm": 1.08907151222229, "learning_rate": 4.733225355658999e-05, "loss": 1.7905, "step": 5828 }, { "epoch": 0.14678324481410962, "grad_norm": 1.0151680707931519, "learning_rate": 4.7294379159549926e-05, "loss": 1.7766, "step": 5859 }, { "epoch": 0.14755987573905202, "grad_norm": 0.9963664412498474, "learning_rate": 4.725625317347119e-05, "loss": 1.7708, "step": 5890 }, { "epoch": 0.1483365066639944, "grad_norm": 1.0762684345245361, "learning_rate": 4.7217876028600374e-05, "loss": 1.7998, "step": 5921 }, { "epoch": 0.14911313758893677, "grad_norm": 1.1151126623153687, "learning_rate": 4.717924815801832e-05, "loss": 1.782, "step": 5952 }, { "epoch": 0.14988976851387914, "grad_norm": 0.9970519542694092, "learning_rate": 4.714036999763532e-05, "loss": 1.7779, "step": 5983 }, { "epoch": 0.1506663994388215, "grad_norm": 1.0146524906158447, "learning_rate": 4.7101241986186116e-05, "loss": 1.7631, "step": 6014 }, { "epoch": 0.15144303036376391, "grad_norm": 1.0554558038711548, "learning_rate": 4.7061864565225e-05, "loss": 1.7729, "step": 6045 }, { "epoch": 0.1522196612887063, "grad_norm": 1.0413624048233032, "learning_rate": 4.702223817912081e-05, "loss": 1.79, "step": 6076 }, { "epoch": 0.15299629221364866, "grad_norm": 1.07778000831604, "learning_rate": 4.698236327505195e-05, "loss": 1.787, "step": 6107 }, { "epoch": 0.15377292313859103, "grad_norm": 1.0084989070892334, "learning_rate": 4.694224030300127e-05, "loss": 1.7695, "step": 6138 }, { "epoch": 0.1545495540635334, "grad_norm": 1.0067577362060547, "learning_rate": 4.690186971575107e-05, "loss": 1.7611, "step": 6169 }, { "epoch": 0.1553261849884758, "grad_norm": 1.0149894952774048, "learning_rate": 4.6861251968877916e-05, "loss": 1.7734, "step": 6200 }, { "epoch": 0.15610281591341818, "grad_norm": 0.996447741985321, "learning_rate": 4.68203875207476e-05, "loss": 1.7609, "step": 6231 }, { "epoch": 0.15687944683836055, "grad_norm": 1.0334714651107788, "learning_rate": 4.677927683250983e-05, "loss": 1.7725, "step": 6262 }, { "epoch": 0.15765607776330293, "grad_norm": 0.949685275554657, "learning_rate": 4.6737920368093156e-05, "loss": 1.7597, "step": 6293 }, { "epoch": 0.15843270868824533, "grad_norm": 1.0946801900863647, "learning_rate": 4.669631859419965e-05, "loss": 1.755, "step": 6324 }, { "epoch": 0.1592093396131877, "grad_norm": 1.0390102863311768, "learning_rate": 4.6654471980299676e-05, "loss": 1.7755, "step": 6355 }, { "epoch": 0.15998597053813007, "grad_norm": 0.974733293056488, "learning_rate": 4.661238099862658e-05, "loss": 1.7619, "step": 6386 }, { "epoch": 0.16076260146307245, "grad_norm": 1.0192877054214478, "learning_rate": 4.657004612417138e-05, "loss": 1.7676, "step": 6417 }, { "epoch": 0.16153923238801482, "grad_norm": 1.0147426128387451, "learning_rate": 4.6527467834677374e-05, "loss": 1.7473, "step": 6448 }, { "epoch": 0.16231586331295722, "grad_norm": 1.0231157541275024, "learning_rate": 4.648464661063478e-05, "loss": 1.766, "step": 6479 }, { "epoch": 0.1630924942378996, "grad_norm": 0.9995896816253662, "learning_rate": 4.6441582935275264e-05, "loss": 1.7594, "step": 6510 }, { "epoch": 0.16386912516284197, "grad_norm": 0.9674972295761108, "learning_rate": 4.6398277294566586e-05, "loss": 1.7347, "step": 6541 }, { "epoch": 0.16464575608778434, "grad_norm": 0.9876355528831482, "learning_rate": 4.6354730177207e-05, "loss": 1.748, "step": 6572 }, { "epoch": 0.1654223870127267, "grad_norm": 1.0107293128967285, "learning_rate": 4.6310942074619787e-05, "loss": 1.7743, "step": 6603 }, { "epoch": 0.1661990179376691, "grad_norm": 0.9731600284576416, "learning_rate": 4.626691348094777e-05, "loss": 1.7515, "step": 6634 }, { "epoch": 0.1669756488626115, "grad_norm": 0.9843081831932068, "learning_rate": 4.622264489304762e-05, "loss": 1.7498, "step": 6665 }, { "epoch": 0.16775227978755386, "grad_norm": 0.9850636720657349, "learning_rate": 4.617813681048434e-05, "loss": 1.7563, "step": 6696 }, { "epoch": 0.16852891071249623, "grad_norm": 0.964599609375, "learning_rate": 4.61333897355256e-05, "loss": 1.7699, "step": 6727 }, { "epoch": 0.16930554163743863, "grad_norm": 0.9415149688720703, "learning_rate": 4.608840417313604e-05, "loss": 1.745, "step": 6758 }, { "epoch": 0.170082172562381, "grad_norm": 0.9608363509178162, "learning_rate": 4.6043180630971646e-05, "loss": 1.7519, "step": 6789 }, { "epoch": 0.17085880348732338, "grad_norm": 0.9133123159408569, "learning_rate": 4.599771961937391e-05, "loss": 1.7432, "step": 6820 }, { "epoch": 0.17163543441226575, "grad_norm": 0.943702220916748, "learning_rate": 4.5952021651364204e-05, "loss": 1.7231, "step": 6851 }, { "epoch": 0.17241206533720813, "grad_norm": 0.9676947593688965, "learning_rate": 4.590608724263786e-05, "loss": 1.7563, "step": 6882 }, { "epoch": 0.17318869626215053, "grad_norm": 0.964959442615509, "learning_rate": 4.585991691155845e-05, "loss": 1.7514, "step": 6913 }, { "epoch": 0.1739653271870929, "grad_norm": 1.0078297853469849, "learning_rate": 4.581351117915188e-05, "loss": 1.7538, "step": 6944 }, { "epoch": 0.17474195811203527, "grad_norm": 0.9631285071372986, "learning_rate": 4.5766870569100534e-05, "loss": 1.7363, "step": 6975 }, { "epoch": 0.17551858903697765, "grad_norm": 0.9542151093482971, "learning_rate": 4.571999560773736e-05, "loss": 1.753, "step": 7006 }, { "epoch": 0.17629521996192002, "grad_norm": 0.9770635366439819, "learning_rate": 4.5672886824039915e-05, "loss": 1.7435, "step": 7037 }, { "epoch": 0.17707185088686242, "grad_norm": 0.979009211063385, "learning_rate": 4.5625544749624435e-05, "loss": 1.7591, "step": 7068 }, { "epoch": 0.1778484818118048, "grad_norm": 0.9663967490196228, "learning_rate": 4.5577969918739794e-05, "loss": 1.7129, "step": 7099 }, { "epoch": 0.17862511273674717, "grad_norm": 1.0181857347488403, "learning_rate": 4.5530162868261486e-05, "loss": 1.7396, "step": 7130 }, { "epoch": 0.17940174366168954, "grad_norm": 1.0090553760528564, "learning_rate": 4.548212413768558e-05, "loss": 1.7488, "step": 7161 }, { "epoch": 0.18017837458663194, "grad_norm": 1.0095964670181274, "learning_rate": 4.543385426912261e-05, "loss": 1.7342, "step": 7192 }, { "epoch": 0.1809550055115743, "grad_norm": 0.994108259677887, "learning_rate": 4.53853538072915e-05, "loss": 1.7445, "step": 7223 }, { "epoch": 0.18173163643651669, "grad_norm": 0.9566254019737244, "learning_rate": 4.533662329951336e-05, "loss": 1.7272, "step": 7254 }, { "epoch": 0.18250826736145906, "grad_norm": 0.9655300974845886, "learning_rate": 4.528766329570536e-05, "loss": 1.7112, "step": 7285 }, { "epoch": 0.18328489828640143, "grad_norm": 0.9693655371665955, "learning_rate": 4.523847434837447e-05, "loss": 1.74, "step": 7316 }, { "epoch": 0.18406152921134383, "grad_norm": 0.9668208956718445, "learning_rate": 4.518905701261128e-05, "loss": 1.7239, "step": 7347 }, { "epoch": 0.1848381601362862, "grad_norm": 1.0193735361099243, "learning_rate": 4.5139411846083715e-05, "loss": 1.7273, "step": 7378 }, { "epoch": 0.18561479106122858, "grad_norm": 0.9525153636932373, "learning_rate": 4.508953940903073e-05, "loss": 1.7356, "step": 7409 }, { "epoch": 0.18639142198617095, "grad_norm": 0.9590099453926086, "learning_rate": 4.5039440264255994e-05, "loss": 1.7296, "step": 7440 }, { "epoch": 0.18716805291111333, "grad_norm": 0.9519559741020203, "learning_rate": 4.498911497712155e-05, "loss": 1.7183, "step": 7471 }, { "epoch": 0.18794468383605573, "grad_norm": 0.9692674279212952, "learning_rate": 4.493856411554142e-05, "loss": 1.7329, "step": 7502 }, { "epoch": 0.1887213147609981, "grad_norm": 0.9676815271377563, "learning_rate": 4.4887788249975206e-05, "loss": 1.7194, "step": 7533 }, { "epoch": 0.18949794568594047, "grad_norm": 0.9612167477607727, "learning_rate": 4.4836787953421656e-05, "loss": 1.7297, "step": 7564 }, { "epoch": 0.19027457661088285, "grad_norm": 0.9661709666252136, "learning_rate": 4.478556380141218e-05, "loss": 1.728, "step": 7595 }, { "epoch": 0.19105120753582525, "grad_norm": 0.95663982629776, "learning_rate": 4.4734116372004375e-05, "loss": 1.7387, "step": 7626 }, { "epoch": 0.19182783846076762, "grad_norm": 0.9469471573829651, "learning_rate": 4.4682446245775477e-05, "loss": 1.7034, "step": 7657 }, { "epoch": 0.19260446938571, "grad_norm": 1.03744637966156, "learning_rate": 4.463055400581586e-05, "loss": 1.7394, "step": 7688 }, { "epoch": 0.19338110031065237, "grad_norm": 0.932919979095459, "learning_rate": 4.4578440237722374e-05, "loss": 1.7113, "step": 7719 }, { "epoch": 0.19415773123559474, "grad_norm": 0.9527629017829895, "learning_rate": 4.452610552959183e-05, "loss": 1.7228, "step": 7750 }, { "epoch": 0.19493436216053714, "grad_norm": 0.9649306535720825, "learning_rate": 4.447355047201428e-05, "loss": 1.7434, "step": 7781 }, { "epoch": 0.1957109930854795, "grad_norm": 0.9027903079986572, "learning_rate": 4.4420775658066414e-05, "loss": 1.7022, "step": 7812 }, { "epoch": 0.19648762401042189, "grad_norm": 0.9486429691314697, "learning_rate": 4.436778168330484e-05, "loss": 1.7277, "step": 7843 }, { "epoch": 0.19726425493536426, "grad_norm": 0.9951580762863159, "learning_rate": 4.4314569145759353e-05, "loss": 1.7354, "step": 7874 }, { "epoch": 0.19804088586030663, "grad_norm": 0.984990656375885, "learning_rate": 4.42611386459262e-05, "loss": 1.7247, "step": 7905 }, { "epoch": 0.19881751678524903, "grad_norm": 1.0146019458770752, "learning_rate": 4.420749078676133e-05, "loss": 1.7048, "step": 7936 }, { "epoch": 0.1995941477101914, "grad_norm": 0.9386274814605713, "learning_rate": 4.4153626173673516e-05, "loss": 1.7035, "step": 7967 }, { "epoch": 0.20037077863513378, "grad_norm": 0.9656661152839661, "learning_rate": 4.409954541451762e-05, "loss": 1.7138, "step": 7998 }, { "epoch": 0.20114740956007615, "grad_norm": 0.9721409678459167, "learning_rate": 4.404524911958764e-05, "loss": 1.7174, "step": 8029 }, { "epoch": 0.20192404048501855, "grad_norm": 0.9685981869697571, "learning_rate": 4.399073790160989e-05, "loss": 1.7162, "step": 8060 }, { "epoch": 0.20270067140996093, "grad_norm": 0.9472204446792603, "learning_rate": 4.393601237573607e-05, "loss": 1.7068, "step": 8091 }, { "epoch": 0.2034773023349033, "grad_norm": 0.9472411870956421, "learning_rate": 4.388107315953628e-05, "loss": 1.7244, "step": 8122 }, { "epoch": 0.20425393325984567, "grad_norm": 0.9380378723144531, "learning_rate": 4.382592087299212e-05, "loss": 1.7169, "step": 8153 }, { "epoch": 0.20503056418478804, "grad_norm": 0.9613087177276611, "learning_rate": 4.377055613848964e-05, "loss": 1.7108, "step": 8184 }, { "epoch": 0.20580719510973045, "grad_norm": 0.9507735967636108, "learning_rate": 4.3714979580812355e-05, "loss": 1.7147, "step": 8215 }, { "epoch": 0.20658382603467282, "grad_norm": 0.9669816493988037, "learning_rate": 4.365919182713416e-05, "loss": 1.7058, "step": 8246 }, { "epoch": 0.2073604569596152, "grad_norm": 0.8880124688148499, "learning_rate": 4.360319350701226e-05, "loss": 1.7049, "step": 8277 }, { "epoch": 0.20813708788455756, "grad_norm": 4.802671909332275, "learning_rate": 4.3546985252380115e-05, "loss": 1.7349, "step": 8308 }, { "epoch": 0.20891371880949994, "grad_norm": 0.9394221901893616, "learning_rate": 4.349056769754021e-05, "loss": 1.6932, "step": 8339 }, { "epoch": 0.20969034973444234, "grad_norm": 0.9521833062171936, "learning_rate": 4.3433941479156994e-05, "loss": 1.695, "step": 8370 }, { "epoch": 0.2104669806593847, "grad_norm": 0.9627339839935303, "learning_rate": 4.3377107236249647e-05, "loss": 1.7329, "step": 8401 }, { "epoch": 0.21124361158432708, "grad_norm": 0.9096025228500366, "learning_rate": 4.332006561018488e-05, "loss": 1.6885, "step": 8432 }, { "epoch": 0.21202024250926946, "grad_norm": 0.9301928877830505, "learning_rate": 4.3262817244669683e-05, "loss": 1.7112, "step": 8463 }, { "epoch": 0.21279687343421186, "grad_norm": 1.0362213850021362, "learning_rate": 4.3205362785744083e-05, "loss": 1.7061, "step": 8494 }, { "epoch": 0.21357350435915423, "grad_norm": 1.986802339553833, "learning_rate": 4.314770288177384e-05, "loss": 1.7095, "step": 8525 }, { "epoch": 0.2143501352840966, "grad_norm": 0.9070325493812561, "learning_rate": 4.308983818344313e-05, "loss": 1.7087, "step": 8556 }, { "epoch": 0.21512676620903898, "grad_norm": 0.985433042049408, "learning_rate": 4.3031769343747206e-05, "loss": 1.7176, "step": 8587 }, { "epoch": 0.21590339713398135, "grad_norm": 0.9339021444320679, "learning_rate": 4.297349701798505e-05, "loss": 1.6823, "step": 8618 }, { "epoch": 0.21668002805892375, "grad_norm": 0.9413474798202515, "learning_rate": 4.2915021863751916e-05, "loss": 1.7205, "step": 8649 }, { "epoch": 0.21745665898386612, "grad_norm": 1.023187279701233, "learning_rate": 4.285634454093198e-05, "loss": 1.6943, "step": 8680 }, { "epoch": 0.2182332899088085, "grad_norm": 0.9431496262550354, "learning_rate": 4.279746571169086e-05, "loss": 1.7012, "step": 8711 }, { "epoch": 0.21900992083375087, "grad_norm": 0.9386807680130005, "learning_rate": 4.2738386040468136e-05, "loss": 1.7069, "step": 8742 }, { "epoch": 0.21978655175869324, "grad_norm": 0.9307882785797119, "learning_rate": 4.2679106193969866e-05, "loss": 1.6914, "step": 8773 }, { "epoch": 0.22056318268363564, "grad_norm": 1.0440980195999146, "learning_rate": 4.261962684116106e-05, "loss": 1.6892, "step": 8804 }, { "epoch": 0.22133981360857802, "grad_norm": 0.9271411895751953, "learning_rate": 4.2559948653258145e-05, "loss": 1.7117, "step": 8835 }, { "epoch": 0.2221164445335204, "grad_norm": 0.9182149767875671, "learning_rate": 4.250007230372134e-05, "loss": 1.7092, "step": 8866 }, { "epoch": 0.22289307545846276, "grad_norm": 0.9707378149032593, "learning_rate": 4.2439998468247126e-05, "loss": 1.7055, "step": 8897 }, { "epoch": 0.22366970638340516, "grad_norm": 0.9413692951202393, "learning_rate": 4.2379727824760566e-05, "loss": 1.6968, "step": 8928 }, { "epoch": 0.22444633730834754, "grad_norm": 1.1943906545639038, "learning_rate": 4.231926105340768e-05, "loss": 1.7174, "step": 8959 }, { "epoch": 0.2252229682332899, "grad_norm": 0.9508203268051147, "learning_rate": 4.225859883654776e-05, "loss": 1.6941, "step": 8990 }, { "epoch": 0.22599959915823228, "grad_norm": 0.9774237275123596, "learning_rate": 4.219774185874569e-05, "loss": 1.6996, "step": 9021 }, { "epoch": 0.22677623008317466, "grad_norm": 0.9906953573226929, "learning_rate": 4.213669080676418e-05, "loss": 1.7186, "step": 9052 }, { "epoch": 0.22755286100811706, "grad_norm": 0.966076135635376, "learning_rate": 4.2075446369556056e-05, "loss": 1.6828, "step": 9083 }, { "epoch": 0.22832949193305943, "grad_norm": 0.9266401529312134, "learning_rate": 4.201400923825648e-05, "loss": 1.6969, "step": 9114 }, { "epoch": 0.2291061228580018, "grad_norm": 0.8797683119773865, "learning_rate": 4.195238010617511e-05, "loss": 1.6835, "step": 9145 }, { "epoch": 0.22988275378294418, "grad_norm": 0.92194664478302, "learning_rate": 4.1890559668788344e-05, "loss": 1.7028, "step": 9176 }, { "epoch": 0.23065938470788655, "grad_norm": 0.9278755187988281, "learning_rate": 4.1828548623731405e-05, "loss": 1.6853, "step": 9207 }, { "epoch": 0.23143601563282895, "grad_norm": 0.9345482587814331, "learning_rate": 4.1766347670790506e-05, "loss": 1.6879, "step": 9238 }, { "epoch": 0.23221264655777132, "grad_norm": 0.9217596054077148, "learning_rate": 4.170395751189495e-05, "loss": 1.679, "step": 9269 }, { "epoch": 0.2329892774827137, "grad_norm": 0.9330849647521973, "learning_rate": 4.164137885110921e-05, "loss": 1.6891, "step": 9300 }, { "epoch": 0.23376590840765607, "grad_norm": 0.923973023891449, "learning_rate": 4.157861239462495e-05, "loss": 1.6929, "step": 9331 }, { "epoch": 0.23454253933259844, "grad_norm": 0.9245807528495789, "learning_rate": 4.1515658850753114e-05, "loss": 1.7162, "step": 9362 }, { "epoch": 0.23531917025754084, "grad_norm": 0.9212767481803894, "learning_rate": 4.145251892991588e-05, "loss": 1.7169, "step": 9393 }, { "epoch": 0.23609580118248322, "grad_norm": 0.9439975023269653, "learning_rate": 4.138919334463868e-05, "loss": 1.6869, "step": 9424 }, { "epoch": 0.2368724321074256, "grad_norm": 0.9876207709312439, "learning_rate": 4.1325682809542124e-05, "loss": 1.6989, "step": 9455 }, { "epoch": 0.23764906303236796, "grad_norm": 0.8833792805671692, "learning_rate": 4.126198804133398e-05, "loss": 1.6917, "step": 9486 }, { "epoch": 0.23842569395731036, "grad_norm": 0.9133450388908386, "learning_rate": 4.1198109758801055e-05, "loss": 1.7038, "step": 9517 }, { "epoch": 0.23920232488225274, "grad_norm": 0.8729329705238342, "learning_rate": 4.113404868280107e-05, "loss": 1.672, "step": 9548 }, { "epoch": 0.2399789558071951, "grad_norm": 0.892375111579895, "learning_rate": 4.106980553625457e-05, "loss": 1.6829, "step": 9579 }, { "epoch": 0.24075558673213748, "grad_norm": 0.8939517140388489, "learning_rate": 4.100538104413674e-05, "loss": 1.6965, "step": 9610 }, { "epoch": 0.24153221765707986, "grad_norm": 0.8656035661697388, "learning_rate": 4.09407759334692e-05, "loss": 1.7063, "step": 9641 }, { "epoch": 0.24230884858202226, "grad_norm": 0.8904181122779846, "learning_rate": 4.087599093331186e-05, "loss": 1.6719, "step": 9672 }, { "epoch": 0.24308547950696463, "grad_norm": 0.9176281690597534, "learning_rate": 4.081102677475462e-05, "loss": 1.6648, "step": 9703 }, { "epoch": 0.243862110431907, "grad_norm": 0.9038643836975098, "learning_rate": 4.0745884190909194e-05, "loss": 1.6756, "step": 9734 }, { "epoch": 0.24463874135684938, "grad_norm": 0.9255871772766113, "learning_rate": 4.0680563916900796e-05, "loss": 1.6752, "step": 9765 }, { "epoch": 0.24541537228179175, "grad_norm": 0.9124768972396851, "learning_rate": 4.0615066689859815e-05, "loss": 1.6982, "step": 9796 }, { "epoch": 0.24619200320673415, "grad_norm": 0.9149096608161926, "learning_rate": 4.0549393248913584e-05, "loss": 1.6692, "step": 9827 }, { "epoch": 0.24696863413167652, "grad_norm": 0.9468510150909424, "learning_rate": 4.048354433517794e-05, "loss": 1.6775, "step": 9858 }, { "epoch": 0.2477452650566189, "grad_norm": 1.2210197448730469, "learning_rate": 4.0417520691748916e-05, "loss": 1.692, "step": 9889 }, { "epoch": 0.24852189598156127, "grad_norm": 0.9717695116996765, "learning_rate": 4.035132306369438e-05, "loss": 1.6583, "step": 9920 }, { "epoch": 0.24929852690650367, "grad_norm": 0.9524528980255127, "learning_rate": 4.028495219804555e-05, "loss": 1.6606, "step": 9951 }, { "epoch": 0.250075157831446, "grad_norm": 0.937414824962616, "learning_rate": 4.021840884378864e-05, "loss": 1.7063, "step": 9982 }, { "epoch": 0.2508517887563884, "grad_norm": 0.9392083883285522, "learning_rate": 4.015169375185633e-05, "loss": 1.6813, "step": 10013 }, { "epoch": 0.2516284196813308, "grad_norm": 0.9055363535881042, "learning_rate": 4.0084807675119396e-05, "loss": 1.6622, "step": 10044 }, { "epoch": 0.25240505060627316, "grad_norm": 0.8967134356498718, "learning_rate": 4.0017751368378106e-05, "loss": 1.7052, "step": 10075 }, { "epoch": 0.25318168153121556, "grad_norm": 0.9304991364479065, "learning_rate": 3.995052558835377e-05, "loss": 1.6779, "step": 10106 }, { "epoch": 0.2539583124561579, "grad_norm": 0.9946010708808899, "learning_rate": 3.988313109368017e-05, "loss": 1.6727, "step": 10137 }, { "epoch": 0.2547349433811003, "grad_norm": 0.936072051525116, "learning_rate": 3.981556864489504e-05, "loss": 1.6861, "step": 10168 }, { "epoch": 0.2555115743060427, "grad_norm": 0.9157513380050659, "learning_rate": 3.974783900443142e-05, "loss": 1.6871, "step": 10199 }, { "epoch": 0.25628820523098506, "grad_norm": 0.8980918526649475, "learning_rate": 3.9679942936609095e-05, "loss": 1.674, "step": 10230 }, { "epoch": 0.25706483615592746, "grad_norm": 0.9256001114845276, "learning_rate": 3.961188120762596e-05, "loss": 1.6858, "step": 10261 }, { "epoch": 0.2578414670808698, "grad_norm": 0.8827298879623413, "learning_rate": 3.954365458554938e-05, "loss": 1.6734, "step": 10292 }, { "epoch": 0.2586180980058122, "grad_norm": 0.9297879338264465, "learning_rate": 3.947526384030751e-05, "loss": 1.6648, "step": 10323 }, { "epoch": 0.2593947289307546, "grad_norm": 0.91744065284729, "learning_rate": 3.9406709743680624e-05, "loss": 1.6713, "step": 10354 }, { "epoch": 0.26017135985569695, "grad_norm": 0.949213445186615, "learning_rate": 3.9337993069292366e-05, "loss": 1.6633, "step": 10385 }, { "epoch": 0.26094799078063935, "grad_norm": 0.9231769442558289, "learning_rate": 3.926911459260109e-05, "loss": 1.6815, "step": 10416 }, { "epoch": 0.2617246217055817, "grad_norm": 0.9055944681167603, "learning_rate": 3.920007509089102e-05, "loss": 1.6621, "step": 10447 }, { "epoch": 0.2625012526305241, "grad_norm": 1.7065776586532593, "learning_rate": 3.913087534326357e-05, "loss": 1.6832, "step": 10478 }, { "epoch": 0.2632778835554665, "grad_norm": 0.9098711013793945, "learning_rate": 3.9061516130628475e-05, "loss": 1.7106, "step": 10509 }, { "epoch": 0.26405451448040884, "grad_norm": 0.9403277635574341, "learning_rate": 3.8991998235695025e-05, "loss": 1.708, "step": 10540 }, { "epoch": 0.26483114540535124, "grad_norm": 0.9076827168464661, "learning_rate": 3.8922322442963224e-05, "loss": 1.6571, "step": 10571 }, { "epoch": 0.26560777633029364, "grad_norm": 0.9140092730522156, "learning_rate": 3.885248953871491e-05, "loss": 1.6626, "step": 10602 }, { "epoch": 0.266384407255236, "grad_norm": 0.9277907013893127, "learning_rate": 3.8782500311004915e-05, "loss": 1.6625, "step": 10633 }, { "epoch": 0.2671610381801784, "grad_norm": 0.9199085831642151, "learning_rate": 3.871235554965218e-05, "loss": 1.663, "step": 10664 }, { "epoch": 0.26793766910512073, "grad_norm": 0.9358407855033875, "learning_rate": 3.864205604623078e-05, "loss": 1.6743, "step": 10695 }, { "epoch": 0.26871430003006314, "grad_norm": 0.8855702877044678, "learning_rate": 3.857160259406107e-05, "loss": 1.6619, "step": 10726 }, { "epoch": 0.26949093095500554, "grad_norm": 0.9157874584197998, "learning_rate": 3.8500995988200674e-05, "loss": 1.6772, "step": 10757 }, { "epoch": 0.2702675618799479, "grad_norm": 0.9023963809013367, "learning_rate": 3.843023702543556e-05, "loss": 1.6559, "step": 10788 }, { "epoch": 0.2710441928048903, "grad_norm": 0.9174203872680664, "learning_rate": 3.8359326504270984e-05, "loss": 1.6393, "step": 10819 }, { "epoch": 0.27182082372983263, "grad_norm": 0.9019688963890076, "learning_rate": 3.828826522492255e-05, "loss": 1.6763, "step": 10850 }, { "epoch": 0.27259745465477503, "grad_norm": 0.8955567479133606, "learning_rate": 3.821705398930713e-05, "loss": 1.6679, "step": 10881 }, { "epoch": 0.27337408557971743, "grad_norm": 0.9326659440994263, "learning_rate": 3.814569360103385e-05, "loss": 1.6765, "step": 10912 }, { "epoch": 0.2741507165046598, "grad_norm": 0.8885530233383179, "learning_rate": 3.807418486539499e-05, "loss": 1.674, "step": 10943 }, { "epoch": 0.2749273474296022, "grad_norm": 0.909267783164978, "learning_rate": 3.80025285893569e-05, "loss": 1.6602, "step": 10974 }, { "epoch": 0.2757039783545445, "grad_norm": 0.9005350470542908, "learning_rate": 3.793072558155093e-05, "loss": 1.6628, "step": 11005 }, { "epoch": 0.2764806092794869, "grad_norm": 0.919683575630188, "learning_rate": 3.785877665226426e-05, "loss": 1.6647, "step": 11036 }, { "epoch": 0.2772572402044293, "grad_norm": 0.8954905867576599, "learning_rate": 3.778668261343079e-05, "loss": 1.6763, "step": 11067 }, { "epoch": 0.27803387112937167, "grad_norm": 0.9120632410049438, "learning_rate": 3.771444427862192e-05, "loss": 1.6552, "step": 11098 }, { "epoch": 0.27881050205431407, "grad_norm": 0.9041403532028198, "learning_rate": 3.7642062463037465e-05, "loss": 1.6604, "step": 11129 }, { "epoch": 0.2795871329792564, "grad_norm": 0.9149596095085144, "learning_rate": 3.7569537983496373e-05, "loss": 1.6666, "step": 11160 }, { "epoch": 0.2803637639041988, "grad_norm": 0.9379541277885437, "learning_rate": 3.749687165842753e-05, "loss": 1.654, "step": 11191 }, { "epoch": 0.2811403948291412, "grad_norm": 0.8942285180091858, "learning_rate": 3.7424064307860536e-05, "loss": 1.6558, "step": 11222 }, { "epoch": 0.28191702575408356, "grad_norm": 0.9407718181610107, "learning_rate": 3.735111675341645e-05, "loss": 1.6574, "step": 11253 }, { "epoch": 0.28269365667902596, "grad_norm": 0.9167685508728027, "learning_rate": 3.7278029818298524e-05, "loss": 1.6767, "step": 11284 }, { "epoch": 0.2834702876039683, "grad_norm": 0.8957439661026001, "learning_rate": 3.720480432728287e-05, "loss": 1.6846, "step": 11315 }, { "epoch": 0.2842469185289107, "grad_norm": 0.9044497609138489, "learning_rate": 3.71314411067092e-05, "loss": 1.6323, "step": 11346 }, { "epoch": 0.2850235494538531, "grad_norm": 0.9024799466133118, "learning_rate": 3.70579409844715e-05, "loss": 1.682, "step": 11377 }, { "epoch": 0.28580018037879545, "grad_norm": 0.9126555323600769, "learning_rate": 3.698430479000865e-05, "loss": 1.6802, "step": 11408 }, { "epoch": 0.28657681130373786, "grad_norm": 0.924310564994812, "learning_rate": 3.691053335429509e-05, "loss": 1.6568, "step": 11439 }, { "epoch": 0.28735344222868026, "grad_norm": 0.8566572666168213, "learning_rate": 3.683662750983147e-05, "loss": 1.6682, "step": 11470 }, { "epoch": 0.2881300731536226, "grad_norm": 0.8287222981452942, "learning_rate": 3.676258809063518e-05, "loss": 1.6484, "step": 11501 }, { "epoch": 0.288906704078565, "grad_norm": 1.0163201093673706, "learning_rate": 3.6688415932231004e-05, "loss": 1.635, "step": 11532 }, { "epoch": 0.28968333500350735, "grad_norm": 0.8813551068305969, "learning_rate": 3.661411187164166e-05, "loss": 1.6654, "step": 11563 }, { "epoch": 0.29045996592844975, "grad_norm": 0.8688541650772095, "learning_rate": 3.65396767473784e-05, "loss": 1.6787, "step": 11594 }, { "epoch": 0.29123659685339215, "grad_norm": 0.8820333480834961, "learning_rate": 3.6465111399431465e-05, "loss": 1.6418, "step": 11625 }, { "epoch": 0.2920132277783345, "grad_norm": 0.9284527897834778, "learning_rate": 3.6390416669260674e-05, "loss": 1.6727, "step": 11656 }, { "epoch": 0.2927898587032769, "grad_norm": 0.8665077090263367, "learning_rate": 3.63155933997859e-05, "loss": 1.6377, "step": 11687 }, { "epoch": 0.29356648962821924, "grad_norm": 0.9927760362625122, "learning_rate": 3.624064243537758e-05, "loss": 1.6628, "step": 11718 }, { "epoch": 0.29434312055316164, "grad_norm": 0.9006340503692627, "learning_rate": 3.616556462184716e-05, "loss": 1.6589, "step": 11749 }, { "epoch": 0.29511975147810404, "grad_norm": 0.9278168082237244, "learning_rate": 3.609036080643755e-05, "loss": 1.6519, "step": 11780 }, { "epoch": 0.2958963824030464, "grad_norm": 0.9146036505699158, "learning_rate": 3.60150318378136e-05, "loss": 1.6541, "step": 11811 }, { "epoch": 0.2966730133279888, "grad_norm": 0.8698804378509521, "learning_rate": 3.5939578566052465e-05, "loss": 1.6722, "step": 11842 }, { "epoch": 0.29744964425293113, "grad_norm": 0.9304215312004089, "learning_rate": 3.586400184263408e-05, "loss": 1.6392, "step": 11873 }, { "epoch": 0.29822627517787353, "grad_norm": 0.8917405605316162, "learning_rate": 3.578830252043148e-05, "loss": 1.6373, "step": 11904 }, { "epoch": 0.29900290610281594, "grad_norm": 0.9075546860694885, "learning_rate": 3.571248145370125e-05, "loss": 1.6401, "step": 11935 }, { "epoch": 0.2997795370277583, "grad_norm": 0.9235963225364685, "learning_rate": 3.5636539498073794e-05, "loss": 1.667, "step": 11966 }, { "epoch": 0.3005561679527007, "grad_norm": 0.9514668583869934, "learning_rate": 3.556047751054378e-05, "loss": 1.6654, "step": 11997 }, { "epoch": 0.301332798877643, "grad_norm": 0.8843172192573547, "learning_rate": 3.548429634946039e-05, "loss": 1.6609, "step": 12028 }, { "epoch": 0.3021094298025854, "grad_norm": 0.9325070977210999, "learning_rate": 3.540799687451768e-05, "loss": 1.6392, "step": 12059 }, { "epoch": 0.30288606072752783, "grad_norm": 0.8855581879615784, "learning_rate": 3.533157994674485e-05, "loss": 1.6607, "step": 12090 }, { "epoch": 0.3036626916524702, "grad_norm": 0.9044367671012878, "learning_rate": 3.5255046428496546e-05, "loss": 1.6417, "step": 12121 }, { "epoch": 0.3044393225774126, "grad_norm": 0.9446975588798523, "learning_rate": 3.517839718344311e-05, "loss": 1.6279, "step": 12152 }, { "epoch": 0.3052159535023549, "grad_norm": 0.8764945864677429, "learning_rate": 3.510163307656086e-05, "loss": 1.6297, "step": 12183 }, { "epoch": 0.3059925844272973, "grad_norm": 0.8633154034614563, "learning_rate": 3.5024754974122324e-05, "loss": 1.648, "step": 12214 }, { "epoch": 0.3067692153522397, "grad_norm": 0.8891461491584778, "learning_rate": 3.494776374368643e-05, "loss": 1.6492, "step": 12245 }, { "epoch": 0.30754584627718207, "grad_norm": 0.8888112306594849, "learning_rate": 3.4870660254088724e-05, "loss": 1.6507, "step": 12276 }, { "epoch": 0.30832247720212447, "grad_norm": 0.8980075120925903, "learning_rate": 3.479344537543164e-05, "loss": 1.6571, "step": 12307 }, { "epoch": 0.3090991081270668, "grad_norm": 0.86537104845047, "learning_rate": 3.4716119979074565e-05, "loss": 1.6331, "step": 12338 }, { "epoch": 0.3098757390520092, "grad_norm": 0.9001527428627014, "learning_rate": 3.463868493762412e-05, "loss": 1.6453, "step": 12369 }, { "epoch": 0.3106523699769516, "grad_norm": 0.8868582248687744, "learning_rate": 3.456114112492418e-05, "loss": 1.6623, "step": 12400 }, { "epoch": 0.31142900090189396, "grad_norm": 0.899174690246582, "learning_rate": 3.4483489416046164e-05, "loss": 1.6451, "step": 12431 }, { "epoch": 0.31220563182683636, "grad_norm": 0.9003099203109741, "learning_rate": 3.440573068727905e-05, "loss": 1.6544, "step": 12462 }, { "epoch": 0.31298226275177876, "grad_norm": 0.8912012577056885, "learning_rate": 3.4327865816119495e-05, "loss": 1.6579, "step": 12493 }, { "epoch": 0.3137588936767211, "grad_norm": 0.8915206789970398, "learning_rate": 3.4249895681262025e-05, "loss": 1.65, "step": 12524 }, { "epoch": 0.3145355246016635, "grad_norm": 0.9975365400314331, "learning_rate": 3.417182116258899e-05, "loss": 1.6506, "step": 12555 }, { "epoch": 0.31531215552660585, "grad_norm": 0.8904858827590942, "learning_rate": 3.409364314116074e-05, "loss": 1.6402, "step": 12586 }, { "epoch": 0.31608878645154825, "grad_norm": 0.8596284985542297, "learning_rate": 3.401536249920559e-05, "loss": 1.6473, "step": 12617 }, { "epoch": 0.31686541737649065, "grad_norm": 0.8970181345939636, "learning_rate": 3.393698012010998e-05, "loss": 1.6347, "step": 12648 }, { "epoch": 0.317642048301433, "grad_norm": 0.890839695930481, "learning_rate": 3.385849688840839e-05, "loss": 1.6323, "step": 12679 }, { "epoch": 0.3184186792263754, "grad_norm": 0.914010226726532, "learning_rate": 3.3779913689773414e-05, "loss": 1.6396, "step": 12710 }, { "epoch": 0.31919531015131775, "grad_norm": 0.8762277960777283, "learning_rate": 3.370123141100578e-05, "loss": 1.6057, "step": 12741 }, { "epoch": 0.31997194107626015, "grad_norm": 0.9001442193984985, "learning_rate": 3.3622450940024305e-05, "loss": 1.647, "step": 12772 }, { "epoch": 0.32074857200120255, "grad_norm": 0.8792625069618225, "learning_rate": 3.35435731658559e-05, "loss": 1.6441, "step": 12803 }, { "epoch": 0.3215252029261449, "grad_norm": 0.889738142490387, "learning_rate": 3.346459897862552e-05, "loss": 1.6433, "step": 12834 }, { "epoch": 0.3223018338510873, "grad_norm": 0.9831507205963135, "learning_rate": 3.338552926954613e-05, "loss": 1.6635, "step": 12865 }, { "epoch": 0.32307846477602964, "grad_norm": 0.8929276466369629, "learning_rate": 3.330636493090868e-05, "loss": 1.6439, "step": 12896 }, { "epoch": 0.32385509570097204, "grad_norm": 0.8983376026153564, "learning_rate": 3.322710685607193e-05, "loss": 1.641, "step": 12927 }, { "epoch": 0.32463172662591444, "grad_norm": 0.8629191517829895, "learning_rate": 3.314775593945251e-05, "loss": 1.6336, "step": 12958 }, { "epoch": 0.3254083575508568, "grad_norm": 0.9323459267616272, "learning_rate": 3.3068313076514714e-05, "loss": 1.6304, "step": 12989 }, { "epoch": 0.3261849884757992, "grad_norm": 0.8520583510398865, "learning_rate": 3.298877916376047e-05, "loss": 1.6277, "step": 13020 }, { "epoch": 0.32696161940074153, "grad_norm": 0.8893365263938904, "learning_rate": 3.290915509871915e-05, "loss": 1.6508, "step": 13051 }, { "epoch": 0.32773825032568393, "grad_norm": 0.8612290024757385, "learning_rate": 3.282944177993753e-05, "loss": 1.6319, "step": 13082 }, { "epoch": 0.32851488125062633, "grad_norm": 0.9378547072410583, "learning_rate": 3.274964010696957e-05, "loss": 1.637, "step": 13113 }, { "epoch": 0.3292915121755687, "grad_norm": 0.8851338028907776, "learning_rate": 3.266975098036629e-05, "loss": 1.6326, "step": 13144 }, { "epoch": 0.3300681431005111, "grad_norm": 0.9052163362503052, "learning_rate": 3.258977530166562e-05, "loss": 1.6507, "step": 13175 }, { "epoch": 0.3308447740254534, "grad_norm": 0.9593476057052612, "learning_rate": 3.250971397338227e-05, "loss": 1.6497, "step": 13206 }, { "epoch": 0.3316214049503958, "grad_norm": 0.8861746788024902, "learning_rate": 3.2429567898997404e-05, "loss": 1.6487, "step": 13237 }, { "epoch": 0.3323980358753382, "grad_norm": 0.8835756778717041, "learning_rate": 3.234933798294859e-05, "loss": 1.6328, "step": 13268 }, { "epoch": 0.33317466680028057, "grad_norm": 0.9133350849151611, "learning_rate": 3.2269025130619535e-05, "loss": 1.6326, "step": 13299 }, { "epoch": 0.333951297725223, "grad_norm": 0.8866952657699585, "learning_rate": 3.218863024832985e-05, "loss": 1.6418, "step": 13330 }, { "epoch": 0.3347279286501654, "grad_norm": 0.8961243033409119, "learning_rate": 3.2108154243324864e-05, "loss": 1.6436, "step": 13361 }, { "epoch": 0.3355045595751077, "grad_norm": 0.871108889579773, "learning_rate": 3.2027598023765345e-05, "loss": 1.6205, "step": 13392 }, { "epoch": 0.3362811905000501, "grad_norm": 0.8666812777519226, "learning_rate": 3.194696249871729e-05, "loss": 1.6478, "step": 13423 }, { "epoch": 0.33705782142499247, "grad_norm": 0.8760045766830444, "learning_rate": 3.186624857814164e-05, "loss": 1.63, "step": 13454 }, { "epoch": 0.33783445234993487, "grad_norm": 0.9416362047195435, "learning_rate": 3.178545717288401e-05, "loss": 1.6268, "step": 13485 }, { "epoch": 0.33861108327487727, "grad_norm": 0.8591843843460083, "learning_rate": 3.170458919466444e-05, "loss": 1.6206, "step": 13516 }, { "epoch": 0.3393877141998196, "grad_norm": 0.9452426433563232, "learning_rate": 3.1623645556067063e-05, "loss": 1.6351, "step": 13547 }, { "epoch": 0.340164345124762, "grad_norm": 0.8680084943771362, "learning_rate": 3.154262717052985e-05, "loss": 1.6389, "step": 13578 }, { "epoch": 0.34094097604970436, "grad_norm": 0.8631142973899841, "learning_rate": 3.146153495233426e-05, "loss": 1.6198, "step": 13609 }, { "epoch": 0.34171760697464676, "grad_norm": 0.905168354511261, "learning_rate": 3.1380369816594944e-05, "loss": 1.6395, "step": 13640 }, { "epoch": 0.34249423789958916, "grad_norm": 0.8644616007804871, "learning_rate": 3.129913267924946e-05, "loss": 1.6282, "step": 13671 }, { "epoch": 0.3432708688245315, "grad_norm": 0.8989911675453186, "learning_rate": 3.121782445704782e-05, "loss": 1.6378, "step": 13702 }, { "epoch": 0.3440474997494739, "grad_norm": 0.908371090888977, "learning_rate": 3.11364460675423e-05, "loss": 1.6152, "step": 13733 }, { "epoch": 0.34482413067441625, "grad_norm": 0.876471221446991, "learning_rate": 3.1054998429076934e-05, "loss": 1.6533, "step": 13764 }, { "epoch": 0.34560076159935865, "grad_norm": 0.879152238368988, "learning_rate": 3.097348246077728e-05, "loss": 1.6196, "step": 13795 }, { "epoch": 0.34637739252430105, "grad_norm": 0.8737320303916931, "learning_rate": 3.0891899082539924e-05, "loss": 1.6122, "step": 13826 }, { "epoch": 0.3471540234492434, "grad_norm": 0.894914448261261, "learning_rate": 3.0810249215022233e-05, "loss": 1.6231, "step": 13857 }, { "epoch": 0.3479306543741858, "grad_norm": 0.8699216842651367, "learning_rate": 3.0728533779631865e-05, "loss": 1.6374, "step": 13888 }, { "epoch": 0.34870728529912814, "grad_norm": 0.8811262249946594, "learning_rate": 3.064675369851637e-05, "loss": 1.635, "step": 13919 }, { "epoch": 0.34948391622407055, "grad_norm": 0.9017398953437805, "learning_rate": 3.056490989455289e-05, "loss": 1.6377, "step": 13950 }, { "epoch": 0.35026054714901295, "grad_norm": 0.8944412469863892, "learning_rate": 3.0483003291337596e-05, "loss": 1.6401, "step": 13981 }, { "epoch": 0.3510371780739553, "grad_norm": 0.9429489970207214, "learning_rate": 3.040103481317539e-05, "loss": 1.6313, "step": 14012 }, { "epoch": 0.3518138089988977, "grad_norm": 0.8750442862510681, "learning_rate": 3.03190053850694e-05, "loss": 1.6305, "step": 14043 }, { "epoch": 0.35259043992384004, "grad_norm": 0.8219168782234192, "learning_rate": 3.0236915932710573e-05, "loss": 1.6289, "step": 14074 }, { "epoch": 0.35336707084878244, "grad_norm": 0.9778533577919006, "learning_rate": 3.0154767382467232e-05, "loss": 1.6221, "step": 14105 }, { "epoch": 0.35414370177372484, "grad_norm": 0.8655532598495483, "learning_rate": 3.0072560661374582e-05, "loss": 1.6352, "step": 14136 }, { "epoch": 0.3549203326986672, "grad_norm": 0.8914876580238342, "learning_rate": 2.999029669712431e-05, "loss": 1.6283, "step": 14167 }, { "epoch": 0.3556969636236096, "grad_norm": 0.9823268055915833, "learning_rate": 2.990797641805408e-05, "loss": 1.6037, "step": 14198 }, { "epoch": 0.356473594548552, "grad_norm": 0.8770225048065186, "learning_rate": 2.982560075313704e-05, "loss": 1.6199, "step": 14229 }, { "epoch": 0.35725022547349433, "grad_norm": 0.8294642567634583, "learning_rate": 2.9743170631971368e-05, "loss": 1.587, "step": 14260 }, { "epoch": 0.35802685639843673, "grad_norm": 0.8678600192070007, "learning_rate": 2.9660686984769792e-05, "loss": 1.6006, "step": 14291 }, { "epoch": 0.3588034873233791, "grad_norm": 0.8829056620597839, "learning_rate": 2.9578150742349047e-05, "loss": 1.6354, "step": 14322 }, { "epoch": 0.3595801182483215, "grad_norm": 0.8785590529441833, "learning_rate": 2.949556283611942e-05, "loss": 1.6293, "step": 14353 }, { "epoch": 0.3603567491732639, "grad_norm": 0.8895497918128967, "learning_rate": 2.9412924198074206e-05, "loss": 1.6121, "step": 14384 }, { "epoch": 0.3611333800982062, "grad_norm": 0.8808982372283936, "learning_rate": 2.9330235760779208e-05, "loss": 1.6251, "step": 14415 }, { "epoch": 0.3619100110231486, "grad_norm": 0.8854517936706543, "learning_rate": 2.9247498457362188e-05, "loss": 1.617, "step": 14446 }, { "epoch": 0.36268664194809097, "grad_norm": 0.9169306755065918, "learning_rate": 2.9164713221502373e-05, "loss": 1.632, "step": 14477 }, { "epoch": 0.36346327287303337, "grad_norm": 0.9230756163597107, "learning_rate": 2.9081880987419912e-05, "loss": 1.6227, "step": 14508 }, { "epoch": 0.3642399037979758, "grad_norm": 0.8728978037834167, "learning_rate": 2.8999002689865296e-05, "loss": 1.6175, "step": 14539 }, { "epoch": 0.3650165347229181, "grad_norm": 0.9083303213119507, "learning_rate": 2.8916079264108852e-05, "loss": 1.6147, "step": 14570 }, { "epoch": 0.3657931656478605, "grad_norm": 0.8577157855033875, "learning_rate": 2.883311164593017e-05, "loss": 1.6357, "step": 14601 }, { "epoch": 0.36656979657280286, "grad_norm": 0.9495333433151245, "learning_rate": 2.875010077160754e-05, "loss": 1.6192, "step": 14632 }, { "epoch": 0.36734642749774526, "grad_norm": 0.9099650979042053, "learning_rate": 2.866704757790741e-05, "loss": 1.6417, "step": 14663 }, { "epoch": 0.36812305842268767, "grad_norm": 0.8684884905815125, "learning_rate": 2.858395300207376e-05, "loss": 1.6171, "step": 14694 }, { "epoch": 0.36889968934763, "grad_norm": 0.8524884581565857, "learning_rate": 2.8500817981817607e-05, "loss": 1.6153, "step": 14725 }, { "epoch": 0.3696763202725724, "grad_norm": 0.9122047424316406, "learning_rate": 2.8417643455306336e-05, "loss": 1.6223, "step": 14756 }, { "epoch": 0.37045295119751476, "grad_norm": 0.875884473323822, "learning_rate": 2.8334430361153185e-05, "loss": 1.6364, "step": 14787 }, { "epoch": 0.37122958212245716, "grad_norm": 0.8689435720443726, "learning_rate": 2.8251179638406612e-05, "loss": 1.639, "step": 14818 }, { "epoch": 0.37200621304739956, "grad_norm": 0.8948261141777039, "learning_rate": 2.8167892226539704e-05, "loss": 1.6091, "step": 14849 }, { "epoch": 0.3727828439723419, "grad_norm": 0.8971095085144043, "learning_rate": 2.8084569065439588e-05, "loss": 1.6527, "step": 14880 }, { "epoch": 0.3735594748972843, "grad_norm": 0.9517078995704651, "learning_rate": 2.8001211095396807e-05, "loss": 1.6201, "step": 14911 }, { "epoch": 0.37433610582222665, "grad_norm": 0.8965676426887512, "learning_rate": 2.791781925709473e-05, "loss": 1.6104, "step": 14942 }, { "epoch": 0.37511273674716905, "grad_norm": 0.9771518707275391, "learning_rate": 2.7834394491598908e-05, "loss": 1.6329, "step": 14973 }, { "epoch": 0.37588936767211145, "grad_norm": 0.8746607899665833, "learning_rate": 2.7750937740346485e-05, "loss": 1.6135, "step": 15004 }, { "epoch": 0.3766659985970538, "grad_norm": 0.860730767250061, "learning_rate": 2.7667449945135564e-05, "loss": 1.6163, "step": 15035 }, { "epoch": 0.3774426295219962, "grad_norm": 0.8879541158676147, "learning_rate": 2.7583932048114557e-05, "loss": 1.6036, "step": 15066 }, { "epoch": 0.37821926044693854, "grad_norm": 0.936347246170044, "learning_rate": 2.7500384991771587e-05, "loss": 1.5914, "step": 15097 }, { "epoch": 0.37899589137188094, "grad_norm": 0.8772497773170471, "learning_rate": 2.7416809718923825e-05, "loss": 1.5994, "step": 15128 }, { "epoch": 0.37977252229682334, "grad_norm": 0.8785439729690552, "learning_rate": 2.7333207172706864e-05, "loss": 1.6057, "step": 15159 }, { "epoch": 0.3805491532217657, "grad_norm": 1.1660430431365967, "learning_rate": 2.7249578296564088e-05, "loss": 1.6257, "step": 15190 }, { "epoch": 0.3813257841467081, "grad_norm": 0.8710660934448242, "learning_rate": 2.7165924034235973e-05, "loss": 1.6223, "step": 15221 }, { "epoch": 0.3821024150716505, "grad_norm": 0.9040597081184387, "learning_rate": 2.708224532974953e-05, "loss": 1.6054, "step": 15252 }, { "epoch": 0.38287904599659284, "grad_norm": 0.8942875862121582, "learning_rate": 2.6998543127407538e-05, "loss": 1.6307, "step": 15283 }, { "epoch": 0.38365567692153524, "grad_norm": 0.8733716011047363, "learning_rate": 2.6914818371777988e-05, "loss": 1.6298, "step": 15314 }, { "epoch": 0.3844323078464776, "grad_norm": 0.9183559417724609, "learning_rate": 2.6831072007683373e-05, "loss": 1.6056, "step": 15345 }, { "epoch": 0.38520893877142, "grad_norm": 0.8594227433204651, "learning_rate": 2.6747304980190018e-05, "loss": 1.6092, "step": 15376 }, { "epoch": 0.3859855696963624, "grad_norm": 0.8866741061210632, "learning_rate": 2.6663518234597453e-05, "loss": 1.6013, "step": 15407 }, { "epoch": 0.38676220062130473, "grad_norm": 0.8940442204475403, "learning_rate": 2.6579712716427696e-05, "loss": 1.6203, "step": 15438 }, { "epoch": 0.38753883154624713, "grad_norm": 0.8408336639404297, "learning_rate": 2.6495889371414652e-05, "loss": 1.5997, "step": 15469 }, { "epoch": 0.3883154624711895, "grad_norm": 0.8949300646781921, "learning_rate": 2.6412049145493367e-05, "loss": 1.6062, "step": 15500 }, { "epoch": 0.3890920933961319, "grad_norm": 0.8919069766998291, "learning_rate": 2.632819298478939e-05, "loss": 1.6197, "step": 15531 }, { "epoch": 0.3898687243210743, "grad_norm": 0.9069515466690063, "learning_rate": 2.6244321835608105e-05, "loss": 1.6147, "step": 15562 }, { "epoch": 0.3906453552460166, "grad_norm": 0.909631073474884, "learning_rate": 2.6160436644424024e-05, "loss": 1.6182, "step": 15593 }, { "epoch": 0.391421986170959, "grad_norm": 0.8957046866416931, "learning_rate": 2.6076538357870133e-05, "loss": 1.5959, "step": 15624 }, { "epoch": 0.39219861709590137, "grad_norm": 0.8709155321121216, "learning_rate": 2.5992627922727196e-05, "loss": 1.6211, "step": 15655 }, { "epoch": 0.39297524802084377, "grad_norm": 0.8602601885795593, "learning_rate": 2.5908706285913066e-05, "loss": 1.6032, "step": 15686 }, { "epoch": 0.39375187894578617, "grad_norm": 0.853210985660553, "learning_rate": 2.5824774394472008e-05, "loss": 1.5994, "step": 15717 }, { "epoch": 0.3945285098707285, "grad_norm": 0.8873618245124817, "learning_rate": 2.5740833195563996e-05, "loss": 1.5949, "step": 15748 }, { "epoch": 0.3953051407956709, "grad_norm": 0.8576481342315674, "learning_rate": 2.5656883636454067e-05, "loss": 1.6029, "step": 15779 }, { "epoch": 0.39608177172061326, "grad_norm": 0.8468286395072937, "learning_rate": 2.557292666450159e-05, "loss": 1.6108, "step": 15810 }, { "epoch": 0.39685840264555566, "grad_norm": 0.9147023558616638, "learning_rate": 2.5488963227149566e-05, "loss": 1.6175, "step": 15841 }, { "epoch": 0.39763503357049806, "grad_norm": 0.8911499977111816, "learning_rate": 2.5404994271913983e-05, "loss": 1.6232, "step": 15872 }, { "epoch": 0.3984116644954404, "grad_norm": 0.8820773363113403, "learning_rate": 2.5321020746373085e-05, "loss": 1.6064, "step": 15903 }, { "epoch": 0.3991882954203828, "grad_norm": 0.8574590086936951, "learning_rate": 2.52370435981567e-05, "loss": 1.6107, "step": 15934 }, { "epoch": 0.39996492634532516, "grad_norm": 0.8728383183479309, "learning_rate": 2.5153063774935533e-05, "loss": 1.5737, "step": 15965 }, { "epoch": 0.40074155727026756, "grad_norm": 0.9165883660316467, "learning_rate": 2.506908222441045e-05, "loss": 1.601, "step": 15996 }, { "epoch": 0.40151818819520996, "grad_norm": 0.8354183435440063, "learning_rate": 2.498509989430187e-05, "loss": 1.5931, "step": 16027 }, { "epoch": 0.4022948191201523, "grad_norm": 0.881255567073822, "learning_rate": 2.4901117732338958e-05, "loss": 1.6065, "step": 16058 }, { "epoch": 0.4030714500450947, "grad_norm": 0.8858495950698853, "learning_rate": 2.481713668624899e-05, "loss": 1.6148, "step": 16089 }, { "epoch": 0.4038480809700371, "grad_norm": 0.8628421425819397, "learning_rate": 2.4733157703746663e-05, "loss": 1.6098, "step": 16120 }, { "epoch": 0.40462471189497945, "grad_norm": 0.8520597815513611, "learning_rate": 2.4649181732523392e-05, "loss": 1.6127, "step": 16151 }, { "epoch": 0.40540134281992185, "grad_norm": 0.8664495348930359, "learning_rate": 2.4565209720236582e-05, "loss": 1.6039, "step": 16182 }, { "epoch": 0.4061779737448642, "grad_norm": 0.8572731614112854, "learning_rate": 2.4481242614498975e-05, "loss": 1.6023, "step": 16213 }, { "epoch": 0.4069546046698066, "grad_norm": 0.8819786906242371, "learning_rate": 2.439728136286796e-05, "loss": 1.6105, "step": 16244 }, { "epoch": 0.407731235594749, "grad_norm": 0.8858864903450012, "learning_rate": 2.4313326912834852e-05, "loss": 1.5995, "step": 16275 }, { "epoch": 0.40850786651969134, "grad_norm": 0.8981665372848511, "learning_rate": 2.4229380211814206e-05, "loss": 1.6041, "step": 16306 }, { "epoch": 0.40928449744463374, "grad_norm": 0.8813167214393616, "learning_rate": 2.4145442207133124e-05, "loss": 1.5976, "step": 16337 }, { "epoch": 0.4100611283695761, "grad_norm": 0.8797750473022461, "learning_rate": 2.406151384602059e-05, "loss": 1.602, "step": 16368 }, { "epoch": 0.4108377592945185, "grad_norm": 0.8421767950057983, "learning_rate": 2.3977596075596747e-05, "loss": 1.5781, "step": 16399 }, { "epoch": 0.4116143902194609, "grad_norm": 0.856469452381134, "learning_rate": 2.3893689842862223e-05, "loss": 1.588, "step": 16430 }, { "epoch": 0.41239102114440324, "grad_norm": 0.8485891819000244, "learning_rate": 2.3809796094687475e-05, "loss": 1.589, "step": 16461 }, { "epoch": 0.41316765206934564, "grad_norm": 0.8968520760536194, "learning_rate": 2.372591577780202e-05, "loss": 1.6012, "step": 16492 }, { "epoch": 0.413944282994288, "grad_norm": 0.875661313533783, "learning_rate": 2.3642049838783838e-05, "loss": 1.6067, "step": 16523 }, { "epoch": 0.4147209139192304, "grad_norm": 0.9094383716583252, "learning_rate": 2.3558199224048666e-05, "loss": 1.5836, "step": 16554 }, { "epoch": 0.4154975448441728, "grad_norm": 0.8707971572875977, "learning_rate": 2.347436487983929e-05, "loss": 1.6115, "step": 16585 }, { "epoch": 0.41627417576911513, "grad_norm": 0.8473120331764221, "learning_rate": 2.3390547752214888e-05, "loss": 1.6227, "step": 16616 }, { "epoch": 0.41705080669405753, "grad_norm": 0.9022457003593445, "learning_rate": 2.330674878704035e-05, "loss": 1.5926, "step": 16647 }, { "epoch": 0.4178274376189999, "grad_norm": 0.8699711561203003, "learning_rate": 2.322296892997561e-05, "loss": 1.6033, "step": 16678 }, { "epoch": 0.4186040685439423, "grad_norm": 0.8574140667915344, "learning_rate": 2.313920912646497e-05, "loss": 1.5861, "step": 16709 }, { "epoch": 0.4193806994688847, "grad_norm": 0.9694622159004211, "learning_rate": 2.305547032172643e-05, "loss": 1.5801, "step": 16740 }, { "epoch": 0.420157330393827, "grad_norm": 0.8770648837089539, "learning_rate": 2.2971753460741014e-05, "loss": 1.5877, "step": 16771 }, { "epoch": 0.4209339613187694, "grad_norm": 0.8674799799919128, "learning_rate": 2.288805948824212e-05, "loss": 1.6172, "step": 16802 }, { "epoch": 0.42171059224371177, "grad_norm": 0.9614238142967224, "learning_rate": 2.2804389348704858e-05, "loss": 1.6031, "step": 16833 }, { "epoch": 0.42248722316865417, "grad_norm": 0.8952891826629639, "learning_rate": 2.2720743986335374e-05, "loss": 1.5921, "step": 16864 }, { "epoch": 0.42326385409359657, "grad_norm": 0.8816654682159424, "learning_rate": 2.2637124345060233e-05, "loss": 1.5838, "step": 16895 }, { "epoch": 0.4240404850185389, "grad_norm": 0.8586190938949585, "learning_rate": 2.2553531368515695e-05, "loss": 1.5965, "step": 16926 }, { "epoch": 0.4248171159434813, "grad_norm": 0.9344453811645508, "learning_rate": 2.2469966000037144e-05, "loss": 1.6041, "step": 16957 }, { "epoch": 0.4255937468684237, "grad_norm": 0.924696683883667, "learning_rate": 2.2386429182648417e-05, "loss": 1.5944, "step": 16988 }, { "epoch": 0.42637037779336606, "grad_norm": 0.8722828030586243, "learning_rate": 2.230292185905114e-05, "loss": 1.6213, "step": 17019 }, { "epoch": 0.42714700871830846, "grad_norm": 0.8846574425697327, "learning_rate": 2.2219444971614116e-05, "loss": 1.6087, "step": 17050 }, { "epoch": 0.4279236396432508, "grad_norm": 0.8602490425109863, "learning_rate": 2.2135999462362655e-05, "loss": 1.5915, "step": 17081 }, { "epoch": 0.4287002705681932, "grad_norm": 0.8821462988853455, "learning_rate": 2.2052586272968003e-05, "loss": 1.6104, "step": 17112 }, { "epoch": 0.4294769014931356, "grad_norm": 0.8978081345558167, "learning_rate": 2.196920634473666e-05, "loss": 1.5932, "step": 17143 }, { "epoch": 0.43025353241807796, "grad_norm": 0.9008092284202576, "learning_rate": 2.1885860618599787e-05, "loss": 1.5915, "step": 17174 }, { "epoch": 0.43103016334302036, "grad_norm": 0.9174037575721741, "learning_rate": 2.1802550035102577e-05, "loss": 1.6062, "step": 17205 }, { "epoch": 0.4318067942679627, "grad_norm": 0.8990745544433594, "learning_rate": 2.171927553439363e-05, "loss": 1.6019, "step": 17236 }, { "epoch": 0.4325834251929051, "grad_norm": 0.867262601852417, "learning_rate": 2.1636038056214376e-05, "loss": 1.5862, "step": 17267 }, { "epoch": 0.4333600561178475, "grad_norm": 0.8440841436386108, "learning_rate": 2.155283853988844e-05, "loss": 1.5698, "step": 17298 }, { "epoch": 0.43413668704278985, "grad_norm": 0.9019381999969482, "learning_rate": 2.146967792431106e-05, "loss": 1.6102, "step": 17329 }, { "epoch": 0.43491331796773225, "grad_norm": 0.8731846213340759, "learning_rate": 2.138655714793849e-05, "loss": 1.5867, "step": 17360 }, { "epoch": 0.4356899488926746, "grad_norm": 0.8628113269805908, "learning_rate": 2.1303477148777367e-05, "loss": 1.5992, "step": 17391 }, { "epoch": 0.436466579817617, "grad_norm": 0.8760020732879639, "learning_rate": 2.122043886437421e-05, "loss": 1.5897, "step": 17422 }, { "epoch": 0.4372432107425594, "grad_norm": 0.8777160048484802, "learning_rate": 2.1137443231804765e-05, "loss": 1.5809, "step": 17453 }, { "epoch": 0.43801984166750174, "grad_norm": 0.8788965344429016, "learning_rate": 2.105449118766347e-05, "loss": 1.6006, "step": 17484 }, { "epoch": 0.43879647259244414, "grad_norm": 0.9438194036483765, "learning_rate": 2.097158366805287e-05, "loss": 1.5916, "step": 17515 }, { "epoch": 0.4395731035173865, "grad_norm": 0.9142041206359863, "learning_rate": 2.0888721608573047e-05, "loss": 1.5874, "step": 17546 }, { "epoch": 0.4403497344423289, "grad_norm": 0.8893091082572937, "learning_rate": 2.0805905944311087e-05, "loss": 1.6032, "step": 17577 }, { "epoch": 0.4411263653672713, "grad_norm": 0.8822838068008423, "learning_rate": 2.0723137609830497e-05, "loss": 1.6018, "step": 17608 }, { "epoch": 0.44190299629221363, "grad_norm": 0.8874905109405518, "learning_rate": 2.0640417539160686e-05, "loss": 1.5894, "step": 17639 }, { "epoch": 0.44267962721715604, "grad_norm": 0.8892821669578552, "learning_rate": 2.0557746665786427e-05, "loss": 1.5873, "step": 17670 }, { "epoch": 0.4434562581420984, "grad_norm": 0.8921502232551575, "learning_rate": 2.0475125922637256e-05, "loss": 1.6111, "step": 17701 }, { "epoch": 0.4442328890670408, "grad_norm": 0.908928394317627, "learning_rate": 2.0392556242077047e-05, "loss": 1.5754, "step": 17732 }, { "epoch": 0.4450095199919832, "grad_norm": 0.9088190793991089, "learning_rate": 2.031003855589343e-05, "loss": 1.589, "step": 17763 }, { "epoch": 0.4457861509169255, "grad_norm": 0.9038158655166626, "learning_rate": 2.022757379528727e-05, "loss": 1.6003, "step": 17794 }, { "epoch": 0.44656278184186793, "grad_norm": 0.9049093127250671, "learning_rate": 2.0145162890862184e-05, "loss": 1.5752, "step": 17825 }, { "epoch": 0.44733941276681033, "grad_norm": 0.827835202217102, "learning_rate": 2.0062806772614022e-05, "loss": 1.584, "step": 17856 }, { "epoch": 0.4481160436917527, "grad_norm": 0.951187789440155, "learning_rate": 1.9980506369920392e-05, "loss": 1.5783, "step": 17887 }, { "epoch": 0.4488926746166951, "grad_norm": 0.8784689903259277, "learning_rate": 1.989826261153015e-05, "loss": 1.5744, "step": 17918 }, { "epoch": 0.4496693055416374, "grad_norm": 0.8900132179260254, "learning_rate": 1.9816076425552923e-05, "loss": 1.5695, "step": 17949 }, { "epoch": 0.4504459364665798, "grad_norm": 0.8698625564575195, "learning_rate": 1.9733948739448676e-05, "loss": 1.5884, "step": 17980 }, { "epoch": 0.4512225673915222, "grad_norm": 0.8933484554290771, "learning_rate": 1.9651880480017155e-05, "loss": 1.5761, "step": 18011 }, { "epoch": 0.45199919831646457, "grad_norm": 0.8772753477096558, "learning_rate": 1.9569872573387516e-05, "loss": 1.5765, "step": 18042 }, { "epoch": 0.45277582924140697, "grad_norm": 0.8958901762962341, "learning_rate": 1.9487925945007854e-05, "loss": 1.5815, "step": 18073 }, { "epoch": 0.4535524601663493, "grad_norm": 0.8793728947639465, "learning_rate": 1.9406041519634726e-05, "loss": 1.5662, "step": 18104 }, { "epoch": 0.4543290910912917, "grad_norm": 0.8501383066177368, "learning_rate": 1.932422022132275e-05, "loss": 1.5755, "step": 18135 }, { "epoch": 0.4551057220162341, "grad_norm": 0.8817576766014099, "learning_rate": 1.924246297341414e-05, "loss": 1.6053, "step": 18166 }, { "epoch": 0.45588235294117646, "grad_norm": 0.8871498703956604, "learning_rate": 1.9160770698528338e-05, "loss": 1.5758, "step": 18197 }, { "epoch": 0.45665898386611886, "grad_norm": 0.8694120049476624, "learning_rate": 1.907914431855156e-05, "loss": 1.5969, "step": 18228 }, { "epoch": 0.4574356147910612, "grad_norm": 0.8998878598213196, "learning_rate": 1.8997584754626412e-05, "loss": 1.5888, "step": 18259 }, { "epoch": 0.4582122457160036, "grad_norm": 0.90102618932724, "learning_rate": 1.8916092927141486e-05, "loss": 1.589, "step": 18290 }, { "epoch": 0.458988876640946, "grad_norm": 0.9209203720092773, "learning_rate": 1.883466975572098e-05, "loss": 1.569, "step": 18321 }, { "epoch": 0.45976550756588835, "grad_norm": 0.906152069568634, "learning_rate": 1.8753316159214312e-05, "loss": 1.5681, "step": 18352 }, { "epoch": 0.46054213849083075, "grad_norm": 0.8538137078285217, "learning_rate": 1.8672033055685766e-05, "loss": 1.5675, "step": 18383 }, { "epoch": 0.4613187694157731, "grad_norm": 0.8828029036521912, "learning_rate": 1.8590821362404116e-05, "loss": 1.5978, "step": 18414 }, { "epoch": 0.4620954003407155, "grad_norm": 0.9327139854431152, "learning_rate": 1.8509681995832294e-05, "loss": 1.5828, "step": 18445 }, { "epoch": 0.4628720312656579, "grad_norm": 0.8733485341072083, "learning_rate": 1.8428615871617004e-05, "loss": 1.6059, "step": 18476 }, { "epoch": 0.46364866219060025, "grad_norm": 0.8775501847267151, "learning_rate": 1.8347623904578448e-05, "loss": 1.5851, "step": 18507 }, { "epoch": 0.46442529311554265, "grad_norm": 0.9166297316551208, "learning_rate": 1.8266707008699975e-05, "loss": 1.5957, "step": 18538 }, { "epoch": 0.465201924040485, "grad_norm": 0.8798413276672363, "learning_rate": 1.818586609711774e-05, "loss": 1.5733, "step": 18569 }, { "epoch": 0.4659785549654274, "grad_norm": 0.8816075921058655, "learning_rate": 1.8105102082110462e-05, "loss": 1.5783, "step": 18600 }, { "epoch": 0.4667551858903698, "grad_norm": 0.897884726524353, "learning_rate": 1.8024415875089058e-05, "loss": 1.5814, "step": 18631 }, { "epoch": 0.46753181681531214, "grad_norm": 0.9144574999809265, "learning_rate": 1.7943808386586407e-05, "loss": 1.57, "step": 18662 }, { "epoch": 0.46830844774025454, "grad_norm": 0.9229109883308411, "learning_rate": 1.7863280526247073e-05, "loss": 1.5678, "step": 18693 }, { "epoch": 0.4690850786651969, "grad_norm": 0.893386721611023, "learning_rate": 1.7782833202817003e-05, "loss": 1.5851, "step": 18724 }, { "epoch": 0.4698617095901393, "grad_norm": 0.8726303577423096, "learning_rate": 1.7702467324133327e-05, "loss": 1.5823, "step": 18755 }, { "epoch": 0.4706383405150817, "grad_norm": 0.8656278252601624, "learning_rate": 1.7622183797114042e-05, "loss": 1.5632, "step": 18786 }, { "epoch": 0.47141497144002403, "grad_norm": 0.8968300819396973, "learning_rate": 1.7541983527747838e-05, "loss": 1.5617, "step": 18817 }, { "epoch": 0.47219160236496643, "grad_norm": 0.8675340414047241, "learning_rate": 1.746186742108387e-05, "loss": 1.5846, "step": 18848 }, { "epoch": 0.47296823328990883, "grad_norm": 0.9016551971435547, "learning_rate": 1.73818363812215e-05, "loss": 1.5821, "step": 18879 }, { "epoch": 0.4737448642148512, "grad_norm": 0.9066888093948364, "learning_rate": 1.7301891311300153e-05, "loss": 1.5856, "step": 18910 }, { "epoch": 0.4745214951397936, "grad_norm": 0.8297645449638367, "learning_rate": 1.7222033113489055e-05, "loss": 1.5766, "step": 18941 }, { "epoch": 0.4752981260647359, "grad_norm": 0.8680350184440613, "learning_rate": 1.7142262688977127e-05, "loss": 1.6083, "step": 18972 }, { "epoch": 0.4760747569896783, "grad_norm": 0.8944342136383057, "learning_rate": 1.7062580937962764e-05, "loss": 1.5802, "step": 19003 }, { "epoch": 0.47685138791462073, "grad_norm": 0.8671095967292786, "learning_rate": 1.698298875964369e-05, "loss": 1.5858, "step": 19034 }, { "epoch": 0.4776280188395631, "grad_norm": 0.8976803421974182, "learning_rate": 1.690348705220684e-05, "loss": 1.5546, "step": 19065 }, { "epoch": 0.4784046497645055, "grad_norm": 0.8853389024734497, "learning_rate": 1.6824076712818156e-05, "loss": 1.5767, "step": 19096 }, { "epoch": 0.4791812806894478, "grad_norm": 0.8673645853996277, "learning_rate": 1.6744758637612533e-05, "loss": 1.5868, "step": 19127 }, { "epoch": 0.4799579116143902, "grad_norm": 0.8737816214561462, "learning_rate": 1.6665533721683664e-05, "loss": 1.5826, "step": 19158 }, { "epoch": 0.4807345425393326, "grad_norm": 0.8424599170684814, "learning_rate": 1.6586402859073974e-05, "loss": 1.5718, "step": 19189 }, { "epoch": 0.48151117346427497, "grad_norm": 0.8880824446678162, "learning_rate": 1.6507366942764463e-05, "loss": 1.5749, "step": 19220 }, { "epoch": 0.48228780438921737, "grad_norm": 0.8780614137649536, "learning_rate": 1.6428426864664732e-05, "loss": 1.5848, "step": 19251 }, { "epoch": 0.4830644353141597, "grad_norm": 0.8731892108917236, "learning_rate": 1.6349583515602816e-05, "loss": 1.5735, "step": 19282 }, { "epoch": 0.4838410662391021, "grad_norm": 0.9195248484611511, "learning_rate": 1.6270837785315208e-05, "loss": 1.6181, "step": 19313 }, { "epoch": 0.4846176971640445, "grad_norm": 0.8955127000808716, "learning_rate": 1.619219056243676e-05, "loss": 1.5962, "step": 19344 }, { "epoch": 0.48539432808898686, "grad_norm": 0.864900529384613, "learning_rate": 1.6113642734490698e-05, "loss": 1.5664, "step": 19375 }, { "epoch": 0.48617095901392926, "grad_norm": 0.8904951214790344, "learning_rate": 1.6035195187878577e-05, "loss": 1.5695, "step": 19406 }, { "epoch": 0.4869475899388716, "grad_norm": 0.9016303420066833, "learning_rate": 1.5956848807870305e-05, "loss": 1.5536, "step": 19437 }, { "epoch": 0.487724220863814, "grad_norm": 0.8600819706916809, "learning_rate": 1.587860447859413e-05, "loss": 1.5722, "step": 19468 }, { "epoch": 0.4885008517887564, "grad_norm": 0.8715513348579407, "learning_rate": 1.5800463083026686e-05, "loss": 1.5776, "step": 19499 }, { "epoch": 0.48927748271369875, "grad_norm": 0.8642942905426025, "learning_rate": 1.572242550298298e-05, "loss": 1.5711, "step": 19530 }, { "epoch": 0.49005411363864115, "grad_norm": 0.877073347568512, "learning_rate": 1.56444926191065e-05, "loss": 1.549, "step": 19561 }, { "epoch": 0.4908307445635835, "grad_norm": 0.8812013268470764, "learning_rate": 1.5566665310859257e-05, "loss": 1.5701, "step": 19592 }, { "epoch": 0.4916073754885259, "grad_norm": 0.9232684373855591, "learning_rate": 1.5488944456511846e-05, "loss": 1.5736, "step": 19623 }, { "epoch": 0.4923840064134683, "grad_norm": 0.9008351564407349, "learning_rate": 1.5411330933133546e-05, "loss": 1.5806, "step": 19654 }, { "epoch": 0.49316063733841065, "grad_norm": 0.909685492515564, "learning_rate": 1.533382561658241e-05, "loss": 1.5643, "step": 19685 }, { "epoch": 0.49393726826335305, "grad_norm": 0.8817586898803711, "learning_rate": 1.525642938149541e-05, "loss": 1.5751, "step": 19716 }, { "epoch": 0.49471389918829545, "grad_norm": 0.853233814239502, "learning_rate": 1.5179143101278536e-05, "loss": 1.5655, "step": 19747 }, { "epoch": 0.4954905301132378, "grad_norm": 0.9284148812294006, "learning_rate": 1.5101967648096955e-05, "loss": 1.5781, "step": 19778 }, { "epoch": 0.4962671610381802, "grad_norm": 0.9339998364448547, "learning_rate": 1.5024903892865172e-05, "loss": 1.5732, "step": 19809 }, { "epoch": 0.49704379196312254, "grad_norm": 0.9051497578620911, "learning_rate": 1.4947952705237184e-05, "loss": 1.5475, "step": 19840 }, { "epoch": 0.49782042288806494, "grad_norm": 0.8712489008903503, "learning_rate": 1.4871114953596682e-05, "loss": 1.559, "step": 19871 }, { "epoch": 0.49859705381300734, "grad_norm": 0.9316422939300537, "learning_rate": 1.4794391505047256e-05, "loss": 1.5615, "step": 19902 }, { "epoch": 0.4993736847379497, "grad_norm": 0.8911099433898926, "learning_rate": 1.4717783225402596e-05, "loss": 1.5776, "step": 19933 }, { "epoch": 0.500150315662892, "grad_norm": 0.8842763900756836, "learning_rate": 1.4641290979176735e-05, "loss": 1.5799, "step": 19964 }, { "epoch": 0.5009269465878344, "grad_norm": 0.9785636067390442, "learning_rate": 1.4564915629574246e-05, "loss": 1.5841, "step": 19995 }, { "epoch": 0.5017035775127768, "grad_norm": 0.8933955430984497, "learning_rate": 1.4488658038480601e-05, "loss": 1.5825, "step": 20026 }, { "epoch": 0.5024802084377192, "grad_norm": 0.9056789875030518, "learning_rate": 1.4412519066452323e-05, "loss": 1.549, "step": 20057 }, { "epoch": 0.5032568393626616, "grad_norm": 0.9087361693382263, "learning_rate": 1.4336499572707373e-05, "loss": 1.5448, "step": 20088 }, { "epoch": 0.5040334702876039, "grad_norm": 0.8783806562423706, "learning_rate": 1.4260600415115433e-05, "loss": 1.5512, "step": 20119 }, { "epoch": 0.5048101012125463, "grad_norm": 0.872285008430481, "learning_rate": 1.4184822450188137e-05, "loss": 1.5703, "step": 20150 }, { "epoch": 0.5055867321374887, "grad_norm": 0.8820748329162598, "learning_rate": 1.410916653306954e-05, "loss": 1.5595, "step": 20181 }, { "epoch": 0.5063633630624311, "grad_norm": 0.912257194519043, "learning_rate": 1.403363351752639e-05, "loss": 1.5795, "step": 20212 }, { "epoch": 0.5071399939873735, "grad_norm": 0.9199615716934204, "learning_rate": 1.3958224255938485e-05, "loss": 1.5686, "step": 20243 }, { "epoch": 0.5079166249123158, "grad_norm": 0.895979106426239, "learning_rate": 1.388293959928911e-05, "loss": 1.5777, "step": 20274 }, { "epoch": 0.5086932558372582, "grad_norm": 0.9053602814674377, "learning_rate": 1.3807780397155379e-05, "loss": 1.5762, "step": 20305 }, { "epoch": 0.5094698867622006, "grad_norm": 0.9254106283187866, "learning_rate": 1.3732747497698655e-05, "loss": 1.5789, "step": 20336 }, { "epoch": 0.510246517687143, "grad_norm": 0.867800235748291, "learning_rate": 1.3657841747655038e-05, "loss": 1.5648, "step": 20367 }, { "epoch": 0.5110231486120854, "grad_norm": 0.9261404871940613, "learning_rate": 1.3583063992325706e-05, "loss": 1.5754, "step": 20398 }, { "epoch": 0.5117997795370277, "grad_norm": 0.8558968901634216, "learning_rate": 1.3508415075567496e-05, "loss": 1.5706, "step": 20429 }, { "epoch": 0.5125764104619701, "grad_norm": 0.8890355229377747, "learning_rate": 1.343389583978327e-05, "loss": 1.5528, "step": 20460 }, { "epoch": 0.5133530413869125, "grad_norm": 0.8697656989097595, "learning_rate": 1.3359507125912468e-05, "loss": 1.5647, "step": 20491 }, { "epoch": 0.5141296723118549, "grad_norm": 0.8828274011611938, "learning_rate": 1.3285249773421627e-05, "loss": 1.5748, "step": 20522 }, { "epoch": 0.5149063032367973, "grad_norm": 0.9029169678688049, "learning_rate": 1.3211124620294884e-05, "loss": 1.5452, "step": 20553 }, { "epoch": 0.5156829341617396, "grad_norm": 2.4660046100616455, "learning_rate": 1.313713250302451e-05, "loss": 1.5479, "step": 20584 }, { "epoch": 0.516459565086682, "grad_norm": 0.9502973556518555, "learning_rate": 1.3063274256601479e-05, "loss": 1.5751, "step": 20615 }, { "epoch": 0.5172361960116244, "grad_norm": 0.8808090090751648, "learning_rate": 1.2989550714506086e-05, "loss": 1.5822, "step": 20646 }, { "epoch": 0.5180128269365668, "grad_norm": 0.8710846304893494, "learning_rate": 1.291596270869846e-05, "loss": 1.559, "step": 20677 }, { "epoch": 0.5187894578615092, "grad_norm": 0.9257534742355347, "learning_rate": 1.284251106960927e-05, "loss": 1.579, "step": 20708 }, { "epoch": 0.5195660887864515, "grad_norm": 0.9311069250106812, "learning_rate": 1.2769196626130263e-05, "loss": 1.5672, "step": 20739 }, { "epoch": 0.5203427197113939, "grad_norm": 0.8663937449455261, "learning_rate": 1.2696020205604969e-05, "loss": 1.5601, "step": 20770 }, { "epoch": 0.5211193506363363, "grad_norm": 0.8988776206970215, "learning_rate": 1.2622982633819359e-05, "loss": 1.5538, "step": 20801 }, { "epoch": 0.5218959815612787, "grad_norm": 0.8996877670288086, "learning_rate": 1.2550084734992484e-05, "loss": 1.5715, "step": 20832 }, { "epoch": 0.5226726124862211, "grad_norm": 0.8598377704620361, "learning_rate": 1.247732733176724e-05, "loss": 1.5494, "step": 20863 }, { "epoch": 0.5234492434111634, "grad_norm": 0.8787708282470703, "learning_rate": 1.2404711245201044e-05, "loss": 1.5401, "step": 20894 }, { "epoch": 0.5242258743361058, "grad_norm": 0.9126632809638977, "learning_rate": 1.2332237294756535e-05, "loss": 1.5567, "step": 20925 }, { "epoch": 0.5250025052610482, "grad_norm": 0.8951886296272278, "learning_rate": 1.225990629829241e-05, "loss": 1.5513, "step": 20956 }, { "epoch": 0.5257791361859906, "grad_norm": 0.8855202794075012, "learning_rate": 1.2187719072054136e-05, "loss": 1.5408, "step": 20987 }, { "epoch": 0.526555767110933, "grad_norm": 0.8703266978263855, "learning_rate": 1.2115676430664735e-05, "loss": 1.5481, "step": 21018 }, { "epoch": 0.5273323980358753, "grad_norm": 0.8899471163749695, "learning_rate": 1.2043779187115647e-05, "loss": 1.5589, "step": 21049 }, { "epoch": 0.5281090289608177, "grad_norm": 0.946517825126648, "learning_rate": 1.1972028152757476e-05, "loss": 1.56, "step": 21080 }, { "epoch": 0.5288856598857601, "grad_norm": 0.9399780631065369, "learning_rate": 1.1900424137290889e-05, "loss": 1.5744, "step": 21111 }, { "epoch": 0.5296622908107025, "grad_norm": 0.9172708988189697, "learning_rate": 1.1828967948757482e-05, "loss": 1.5737, "step": 21142 }, { "epoch": 0.5304389217356449, "grad_norm": 0.9447731375694275, "learning_rate": 1.175766039353062e-05, "loss": 1.5644, "step": 21173 }, { "epoch": 0.5312155526605873, "grad_norm": 0.8953673243522644, "learning_rate": 1.1686502276306382e-05, "loss": 1.5763, "step": 21204 }, { "epoch": 0.5319921835855296, "grad_norm": 0.9340639114379883, "learning_rate": 1.1615494400094445e-05, "loss": 1.5508, "step": 21235 }, { "epoch": 0.532768814510472, "grad_norm": 0.8861269950866699, "learning_rate": 1.1544637566209029e-05, "loss": 1.5558, "step": 21266 }, { "epoch": 0.5335454454354144, "grad_norm": 0.8895665407180786, "learning_rate": 1.1473932574259886e-05, "loss": 1.5614, "step": 21297 }, { "epoch": 0.5343220763603568, "grad_norm": 0.8842455148696899, "learning_rate": 1.1403380222143247e-05, "loss": 1.5585, "step": 21328 }, { "epoch": 0.5350987072852992, "grad_norm": 0.8868069052696228, "learning_rate": 1.1332981306032808e-05, "loss": 1.535, "step": 21359 }, { "epoch": 0.5358753382102415, "grad_norm": 0.9249818921089172, "learning_rate": 1.1262736620370762e-05, "loss": 1.5528, "step": 21390 }, { "epoch": 0.5366519691351839, "grad_norm": 0.8665826916694641, "learning_rate": 1.1192646957858854e-05, "loss": 1.5656, "step": 21421 }, { "epoch": 0.5374286000601263, "grad_norm": 0.9058834910392761, "learning_rate": 1.1122713109449381e-05, "loss": 1.5498, "step": 21452 }, { "epoch": 0.5382052309850687, "grad_norm": 0.8894379138946533, "learning_rate": 1.105293586433634e-05, "loss": 1.5667, "step": 21483 }, { "epoch": 0.5389818619100111, "grad_norm": 0.9372698068618774, "learning_rate": 1.0983316009946446e-05, "loss": 1.5463, "step": 21514 }, { "epoch": 0.5397584928349534, "grad_norm": 0.9113041758537292, "learning_rate": 1.0913854331930282e-05, "loss": 1.5809, "step": 21545 }, { "epoch": 0.5405351237598958, "grad_norm": 0.9373642802238464, "learning_rate": 1.0844551614153456e-05, "loss": 1.5664, "step": 21576 }, { "epoch": 0.5413117546848382, "grad_norm": 0.8933825492858887, "learning_rate": 1.0775408638687725e-05, "loss": 1.555, "step": 21607 }, { "epoch": 0.5420883856097806, "grad_norm": 0.9607747793197632, "learning_rate": 1.0706426185802165e-05, "loss": 1.5545, "step": 21638 }, { "epoch": 0.542865016534723, "grad_norm": 0.8767135143280029, "learning_rate": 1.0637605033954371e-05, "loss": 1.5552, "step": 21669 }, { "epoch": 0.5436416474596653, "grad_norm": 0.9311366081237793, "learning_rate": 1.05689459597817e-05, "loss": 1.5664, "step": 21700 }, { "epoch": 0.5444182783846077, "grad_norm": 0.9141988158226013, "learning_rate": 1.050044973809246e-05, "loss": 1.5567, "step": 21731 }, { "epoch": 0.5451949093095501, "grad_norm": 0.899506151676178, "learning_rate": 1.043211714185722e-05, "loss": 1.5702, "step": 21762 }, { "epoch": 0.5459715402344925, "grad_norm": 0.8849896788597107, "learning_rate": 1.036394894220003e-05, "loss": 1.5609, "step": 21793 }, { "epoch": 0.5467481711594349, "grad_norm": 0.904323399066925, "learning_rate": 1.0295945908389751e-05, "loss": 1.5607, "step": 21824 }, { "epoch": 0.5475248020843771, "grad_norm": 0.9160985350608826, "learning_rate": 1.0228108807831393e-05, "loss": 1.5752, "step": 21855 }, { "epoch": 0.5483014330093195, "grad_norm": 0.8923236131668091, "learning_rate": 1.01604384060574e-05, "loss": 1.5468, "step": 21886 }, { "epoch": 0.549078063934262, "grad_norm": 0.8595706820487976, "learning_rate": 1.009293546671907e-05, "loss": 1.5394, "step": 21917 }, { "epoch": 0.5498546948592044, "grad_norm": 0.9517648220062256, "learning_rate": 1.002560075157791e-05, "loss": 1.548, "step": 21948 }, { "epoch": 0.5506313257841468, "grad_norm": 0.86150723695755, "learning_rate": 9.958435020496995e-06, "loss": 1.5662, "step": 21979 }, { "epoch": 0.551407956709089, "grad_norm": 0.9183287620544434, "learning_rate": 9.89143903143249e-06, "loss": 1.5598, "step": 22010 }, { "epoch": 0.5521845876340314, "grad_norm": 0.8485851287841797, "learning_rate": 9.824613540425038e-06, "loss": 1.5521, "step": 22041 }, { "epoch": 0.5529612185589738, "grad_norm": 0.8906897306442261, "learning_rate": 9.757959301591197e-06, "loss": 1.5744, "step": 22072 }, { "epoch": 0.5537378494839162, "grad_norm": 0.9588461518287659, "learning_rate": 9.691477067115017e-06, "loss": 1.557, "step": 22103 }, { "epoch": 0.5545144804088586, "grad_norm": 0.9231610298156738, "learning_rate": 9.625167587239467e-06, "loss": 1.5609, "step": 22134 }, { "epoch": 0.5552911113338009, "grad_norm": 0.9148405194282532, "learning_rate": 9.559031610258007e-06, "loss": 1.5533, "step": 22165 }, { "epoch": 0.5560677422587433, "grad_norm": 0.9222816824913025, "learning_rate": 9.493069882506164e-06, "loss": 1.5558, "step": 22196 }, { "epoch": 0.5568443731836857, "grad_norm": 0.8628998398780823, "learning_rate": 9.427283148353056e-06, "loss": 1.5721, "step": 22227 }, { "epoch": 0.5576210041086281, "grad_norm": 0.8785997033119202, "learning_rate": 9.361672150193052e-06, "loss": 1.5538, "step": 22258 }, { "epoch": 0.5583976350335705, "grad_norm": 0.9264510273933411, "learning_rate": 9.29623762843734e-06, "loss": 1.5606, "step": 22289 }, { "epoch": 0.5591742659585128, "grad_norm": 0.9573788642883301, "learning_rate": 9.230980321505594e-06, "loss": 1.5596, "step": 22320 }, { "epoch": 0.5599508968834552, "grad_norm": 0.8998466730117798, "learning_rate": 9.165900965817668e-06, "loss": 1.5421, "step": 22351 }, { "epoch": 0.5607275278083976, "grad_norm": 0.9093721508979797, "learning_rate": 9.101000295785245e-06, "loss": 1.5575, "step": 22382 }, { "epoch": 0.56150415873334, "grad_norm": 0.8681372404098511, "learning_rate": 9.036279043803565e-06, "loss": 1.5501, "step": 22413 }, { "epoch": 0.5622807896582824, "grad_norm": 0.8595815896987915, "learning_rate": 8.971737940243147e-06, "loss": 1.5607, "step": 22444 }, { "epoch": 0.5630574205832247, "grad_norm": 0.894629955291748, "learning_rate": 8.907377713441592e-06, "loss": 1.5441, "step": 22475 }, { "epoch": 0.5638340515081671, "grad_norm": 0.8969973921775818, "learning_rate": 8.843199089695293e-06, "loss": 1.5566, "step": 22506 }, { "epoch": 0.5646106824331095, "grad_norm": 0.8613621592521667, "learning_rate": 8.779202793251311e-06, "loss": 1.5548, "step": 22537 }, { "epoch": 0.5653873133580519, "grad_norm": 0.942423403263092, "learning_rate": 8.715389546299149e-06, "loss": 1.5715, "step": 22568 }, { "epoch": 0.5661639442829943, "grad_norm": 0.8348737955093384, "learning_rate": 8.651760068962617e-06, "loss": 1.5554, "step": 22599 }, { "epoch": 0.5669405752079366, "grad_norm": 0.9314144849777222, "learning_rate": 8.588315079291733e-06, "loss": 1.5738, "step": 22630 }, { "epoch": 0.567717206132879, "grad_norm": 0.9175193309783936, "learning_rate": 8.52505529325457e-06, "loss": 1.5369, "step": 22661 }, { "epoch": 0.5684938370578214, "grad_norm": 0.9038921594619751, "learning_rate": 8.461981424729216e-06, "loss": 1.5518, "step": 22692 }, { "epoch": 0.5692704679827638, "grad_norm": 0.9067641496658325, "learning_rate": 8.399094185495725e-06, "loss": 1.5554, "step": 22723 }, { "epoch": 0.5700470989077062, "grad_norm": 0.8826406002044678, "learning_rate": 8.336394285228017e-06, "loss": 1.5648, "step": 22754 }, { "epoch": 0.5708237298326485, "grad_norm": 0.8925307393074036, "learning_rate": 8.273882431485952e-06, "loss": 1.5504, "step": 22785 }, { "epoch": 0.5716003607575909, "grad_norm": 0.9345889091491699, "learning_rate": 8.211559329707316e-06, "loss": 1.5451, "step": 22816 }, { "epoch": 0.5723769916825333, "grad_norm": 0.913657546043396, "learning_rate": 8.149425683199823e-06, "loss": 1.569, "step": 22847 }, { "epoch": 0.5731536226074757, "grad_norm": 0.8716319799423218, "learning_rate": 8.08748219313325e-06, "loss": 1.5714, "step": 22878 }, { "epoch": 0.5739302535324181, "grad_norm": 0.869755744934082, "learning_rate": 8.025729558531453e-06, "loss": 1.5588, "step": 22909 }, { "epoch": 0.5747068844573605, "grad_norm": 0.8963233232498169, "learning_rate": 7.964168476264508e-06, "loss": 1.5429, "step": 22940 }, { "epoch": 0.5754835153823028, "grad_norm": 0.8966054916381836, "learning_rate": 7.902799641040884e-06, "loss": 1.5607, "step": 22971 }, { "epoch": 0.5762601463072452, "grad_norm": 0.9066189527511597, "learning_rate": 7.841623745399523e-06, "loss": 1.5386, "step": 23002 }, { "epoch": 0.5770367772321876, "grad_norm": 0.9433830976486206, "learning_rate": 7.780641479702114e-06, "loss": 1.5594, "step": 23033 }, { "epoch": 0.57781340815713, "grad_norm": 0.9131210446357727, "learning_rate": 7.719853532125227e-06, "loss": 1.5359, "step": 23064 }, { "epoch": 0.5785900390820724, "grad_norm": 0.880100429058075, "learning_rate": 7.65926058865258e-06, "loss": 1.547, "step": 23095 }, { "epoch": 0.5793666700070147, "grad_norm": 0.960096538066864, "learning_rate": 7.598863333067313e-06, "loss": 1.5449, "step": 23126 }, { "epoch": 0.5801433009319571, "grad_norm": 0.9209834337234497, "learning_rate": 7.538662446944253e-06, "loss": 1.5648, "step": 23157 }, { "epoch": 0.5809199318568995, "grad_norm": 0.9270380139350891, "learning_rate": 7.478658609642211e-06, "loss": 1.5551, "step": 23188 }, { "epoch": 0.5816965627818419, "grad_norm": 0.9106172919273376, "learning_rate": 7.418852498296327e-06, "loss": 1.5482, "step": 23219 }, { "epoch": 0.5824731937067843, "grad_norm": 0.9246364831924438, "learning_rate": 7.359244787810457e-06, "loss": 1.5383, "step": 23250 }, { "epoch": 0.5832498246317266, "grad_norm": 0.945947527885437, "learning_rate": 7.299836150849493e-06, "loss": 1.5497, "step": 23281 }, { "epoch": 0.584026455556669, "grad_norm": 0.9162856340408325, "learning_rate": 7.240627257831847e-06, "loss": 1.5606, "step": 23312 }, { "epoch": 0.5848030864816114, "grad_norm": 0.898065447807312, "learning_rate": 7.1816187769218195e-06, "loss": 1.5512, "step": 23343 }, { "epoch": 0.5855797174065538, "grad_norm": 0.928261935710907, "learning_rate": 7.1228113740220895e-06, "loss": 1.5689, "step": 23374 }, { "epoch": 0.5863563483314962, "grad_norm": 0.8689351677894592, "learning_rate": 7.064205712766226e-06, "loss": 1.5365, "step": 23405 }, { "epoch": 0.5871329792564385, "grad_norm": 0.9178231954574585, "learning_rate": 7.005802454511129e-06, "loss": 1.5358, "step": 23436 }, { "epoch": 0.5879096101813809, "grad_norm": 0.892164409160614, "learning_rate": 6.947602258329639e-06, "loss": 1.5617, "step": 23467 }, { "epoch": 0.5886862411063233, "grad_norm": 0.8901685476303101, "learning_rate": 6.889605781003078e-06, "loss": 1.5314, "step": 23498 }, { "epoch": 0.5894628720312657, "grad_norm": 0.904805064201355, "learning_rate": 6.831813677013776e-06, "loss": 1.5589, "step": 23529 }, { "epoch": 0.5902395029562081, "grad_norm": 0.8700739145278931, "learning_rate": 6.774226598537792e-06, "loss": 1.5414, "step": 23560 }, { "epoch": 0.5910161338811504, "grad_norm": 0.9062610864639282, "learning_rate": 6.716845195437482e-06, "loss": 1.5579, "step": 23591 }, { "epoch": 0.5917927648060928, "grad_norm": 0.9072427749633789, "learning_rate": 6.659670115254168e-06, "loss": 1.5498, "step": 23622 }, { "epoch": 0.5925693957310352, "grad_norm": 0.8828601837158203, "learning_rate": 6.602702003200872e-06, "loss": 1.5546, "step": 23653 }, { "epoch": 0.5933460266559776, "grad_norm": 0.8832286596298218, "learning_rate": 6.545941502154992e-06, "loss": 1.5277, "step": 23684 }, { "epoch": 0.59412265758092, "grad_norm": 0.9346180558204651, "learning_rate": 6.489389252651057e-06, "loss": 1.5555, "step": 23715 }, { "epoch": 0.5948992885058623, "grad_norm": 0.9047152400016785, "learning_rate": 6.4330458928735325e-06, "loss": 1.5593, "step": 23746 }, { "epoch": 0.5956759194308047, "grad_norm": 0.9097047448158264, "learning_rate": 6.376912058649559e-06, "loss": 1.5481, "step": 23777 }, { "epoch": 0.5964525503557471, "grad_norm": 0.9316630959510803, "learning_rate": 6.320988383441845e-06, "loss": 1.5605, "step": 23808 }, { "epoch": 0.5972291812806895, "grad_norm": 0.9213578701019287, "learning_rate": 6.265275498341452e-06, "loss": 1.5532, "step": 23839 }, { "epoch": 0.5980058122056319, "grad_norm": 0.8936622738838196, "learning_rate": 6.209774032060714e-06, "loss": 1.576, "step": 23870 }, { "epoch": 0.5987824431305742, "grad_norm": 0.907272458076477, "learning_rate": 6.1544846109261365e-06, "loss": 1.5403, "step": 23901 }, { "epoch": 0.5995590740555166, "grad_norm": 0.943996012210846, "learning_rate": 6.099407858871342e-06, "loss": 1.5353, "step": 23932 }, { "epoch": 0.600335704980459, "grad_norm": 0.8588092923164368, "learning_rate": 6.044544397429958e-06, "loss": 1.5432, "step": 23963 }, { "epoch": 0.6011123359054014, "grad_norm": 0.8964900374412537, "learning_rate": 5.989894845728708e-06, "loss": 1.5383, "step": 23994 }, { "epoch": 0.6018889668303438, "grad_norm": 0.9199158549308777, "learning_rate": 5.9354598204803605e-06, "loss": 1.5479, "step": 24025 }, { "epoch": 0.602665597755286, "grad_norm": 0.9112527370452881, "learning_rate": 5.881239935976762e-06, "loss": 1.5592, "step": 24056 }, { "epoch": 0.6034422286802285, "grad_norm": 0.8948724865913391, "learning_rate": 5.827235804081954e-06, "loss": 1.5442, "step": 24087 }, { "epoch": 0.6042188596051709, "grad_norm": 0.8647116422653198, "learning_rate": 5.773448034225221e-06, "loss": 1.5399, "step": 24118 }, { "epoch": 0.6049954905301133, "grad_norm": 0.9131345748901367, "learning_rate": 5.719877233394228e-06, "loss": 1.5597, "step": 24149 }, { "epoch": 0.6057721214550557, "grad_norm": 0.9051525592803955, "learning_rate": 5.666524006128191e-06, "loss": 1.5403, "step": 24180 }, { "epoch": 0.606548752379998, "grad_norm": 0.9300383925437927, "learning_rate": 5.613388954511015e-06, "loss": 1.5424, "step": 24211 }, { "epoch": 0.6073253833049403, "grad_norm": 2.001582384109497, "learning_rate": 5.560472678164552e-06, "loss": 1.5169, "step": 24242 }, { "epoch": 0.6081020142298827, "grad_norm": 0.9331688284873962, "learning_rate": 5.507775774241775e-06, "loss": 1.5582, "step": 24273 }, { "epoch": 0.6088786451548251, "grad_norm": 0.8840691447257996, "learning_rate": 5.4552988374200945e-06, "loss": 1.5441, "step": 24304 }, { "epoch": 0.6096552760797675, "grad_norm": 0.8922248482704163, "learning_rate": 5.403042459894597e-06, "loss": 1.5109, "step": 24335 }, { "epoch": 0.6104319070047098, "grad_norm": 0.9129870533943176, "learning_rate": 5.3510072313714135e-06, "loss": 1.5311, "step": 24366 }, { "epoch": 0.6112085379296522, "grad_norm": 0.9153522849082947, "learning_rate": 5.2991937390610205e-06, "loss": 1.5589, "step": 24397 }, { "epoch": 0.6119851688545946, "grad_norm": 0.926564633846283, "learning_rate": 5.247602567671625e-06, "loss": 1.5615, "step": 24428 }, { "epoch": 0.612761799779537, "grad_norm": 0.8949169516563416, "learning_rate": 5.196234299402603e-06, "loss": 1.5302, "step": 24459 }, { "epoch": 0.6135384307044794, "grad_norm": 0.9150370955467224, "learning_rate": 5.145089513937865e-06, "loss": 1.5299, "step": 24490 }, { "epoch": 0.6143150616294217, "grad_norm": 0.9064559936523438, "learning_rate": 5.094168788439369e-06, "loss": 1.5589, "step": 24521 }, { "epoch": 0.6150916925543641, "grad_norm": 0.9159832000732422, "learning_rate": 5.043472697540594e-06, "loss": 1.5586, "step": 24552 }, { "epoch": 0.6158683234793065, "grad_norm": 0.920007586479187, "learning_rate": 4.993001813340012e-06, "loss": 1.5293, "step": 24583 }, { "epoch": 0.6166449544042489, "grad_norm": 0.8829004764556885, "learning_rate": 4.942756705394702e-06, "loss": 1.5531, "step": 24614 }, { "epoch": 0.6174215853291913, "grad_norm": 0.8751837611198425, "learning_rate": 4.892737940713884e-06, "loss": 1.538, "step": 24645 }, { "epoch": 0.6181982162541336, "grad_norm": 0.9076789617538452, "learning_rate": 4.842946083752511e-06, "loss": 1.5282, "step": 24676 }, { "epoch": 0.618974847179076, "grad_norm": 0.9124575853347778, "learning_rate": 4.79338169640493e-06, "loss": 1.5309, "step": 24707 }, { "epoch": 0.6197514781040184, "grad_norm": 0.8740583062171936, "learning_rate": 4.74404533799851e-06, "loss": 1.5308, "step": 24738 }, { "epoch": 0.6205281090289608, "grad_norm": 0.8814982771873474, "learning_rate": 4.694937565287344e-06, "loss": 1.5122, "step": 24769 }, { "epoch": 0.6213047399539032, "grad_norm": 0.9013394117355347, "learning_rate": 4.646058932445985e-06, "loss": 1.5372, "step": 24800 }, { "epoch": 0.6220813708788456, "grad_norm": 0.9024440050125122, "learning_rate": 4.597409991063148e-06, "loss": 1.5449, "step": 24831 }, { "epoch": 0.6228580018037879, "grad_norm": 0.9109915494918823, "learning_rate": 4.5489912901355375e-06, "loss": 1.5525, "step": 24862 }, { "epoch": 0.6236346327287303, "grad_norm": 0.8715227246284485, "learning_rate": 4.500803376061608e-06, "loss": 1.562, "step": 24893 }, { "epoch": 0.6244112636536727, "grad_norm": 0.924454391002655, "learning_rate": 4.45284679263541e-06, "loss": 1.5415, "step": 24924 }, { "epoch": 0.6251878945786151, "grad_norm": 0.9070348739624023, "learning_rate": 4.4051220810404775e-06, "loss": 1.5204, "step": 24955 }, { "epoch": 0.6259645255035575, "grad_norm": 0.8900454640388489, "learning_rate": 4.3576297798437025e-06, "loss": 1.5172, "step": 24986 }, { "epoch": 0.6267411564284998, "grad_norm": 0.9387392997741699, "learning_rate": 4.3103704249892436e-06, "loss": 1.5577, "step": 25017 }, { "epoch": 0.6275177873534422, "grad_norm": 0.9040491580963135, "learning_rate": 4.263344549792487e-06, "loss": 1.5575, "step": 25048 }, { "epoch": 0.6282944182783846, "grad_norm": 0.9115748405456543, "learning_rate": 4.216552684934056e-06, "loss": 1.5373, "step": 25079 }, { "epoch": 0.629071049203327, "grad_norm": 0.8901536464691162, "learning_rate": 4.169995358453777e-06, "loss": 1.5433, "step": 25110 }, { "epoch": 0.6298476801282694, "grad_norm": 0.9255604147911072, "learning_rate": 4.123673095744757e-06, "loss": 1.5328, "step": 25141 }, { "epoch": 0.6306243110532117, "grad_norm": 0.915124237537384, "learning_rate": 4.077586419547435e-06, "loss": 1.5506, "step": 25172 }, { "epoch": 0.6314009419781541, "grad_norm": 0.8989410996437073, "learning_rate": 4.03173584994368e-06, "loss": 1.5426, "step": 25203 }, { "epoch": 0.6321775729030965, "grad_norm": 0.8552297353744507, "learning_rate": 3.986121904350948e-06, "loss": 1.5271, "step": 25234 }, { "epoch": 0.6329542038280389, "grad_norm": 0.956108033657074, "learning_rate": 3.940745097516407e-06, "loss": 1.5405, "step": 25265 }, { "epoch": 0.6337308347529813, "grad_norm": 0.8881359696388245, "learning_rate": 3.89560594151116e-06, "loss": 1.535, "step": 25296 }, { "epoch": 0.6345074656779236, "grad_norm": 0.8930270075798035, "learning_rate": 3.850704945724456e-06, "loss": 1.5405, "step": 25327 }, { "epoch": 0.635284096602866, "grad_norm": 0.862516462802887, "learning_rate": 3.8060426168579077e-06, "loss": 1.5396, "step": 25358 }, { "epoch": 0.6360607275278084, "grad_norm": 0.9361773729324341, "learning_rate": 3.7616194589198407e-06, "loss": 1.5546, "step": 25389 }, { "epoch": 0.6368373584527508, "grad_norm": 0.897813081741333, "learning_rate": 3.7174359732195574e-06, "loss": 1.5479, "step": 25420 }, { "epoch": 0.6376139893776932, "grad_norm": 0.9219157099723816, "learning_rate": 3.673492658361677e-06, "loss": 1.5404, "step": 25451 }, { "epoch": 0.6383906203026355, "grad_norm": 0.9218947291374207, "learning_rate": 3.6297900102405467e-06, "loss": 1.5637, "step": 25482 }, { "epoch": 0.6391672512275779, "grad_norm": 0.9385016560554504, "learning_rate": 3.586328522034607e-06, "loss": 1.5455, "step": 25513 }, { "epoch": 0.6399438821525203, "grad_norm": 0.9211886525154114, "learning_rate": 3.543108684200838e-06, "loss": 1.5471, "step": 25544 }, { "epoch": 0.6407205130774627, "grad_norm": 0.9449164867401123, "learning_rate": 3.5001309844692464e-06, "loss": 1.5341, "step": 25575 }, { "epoch": 0.6414971440024051, "grad_norm": 0.9099578261375427, "learning_rate": 3.4573959078373215e-06, "loss": 1.5688, "step": 25606 }, { "epoch": 0.6422737749273474, "grad_norm": 0.8888375759124756, "learning_rate": 3.4149039365646063e-06, "loss": 1.5408, "step": 25637 }, { "epoch": 0.6430504058522898, "grad_norm": 0.9400342106819153, "learning_rate": 3.3726555501672143e-06, "loss": 1.5442, "step": 25668 }, { "epoch": 0.6438270367772322, "grad_norm": 0.9119865298271179, "learning_rate": 3.33065122541244e-06, "loss": 1.542, "step": 25699 }, { "epoch": 0.6446036677021746, "grad_norm": 0.9019779562950134, "learning_rate": 3.288891436313385e-06, "loss": 1.5281, "step": 25730 }, { "epoch": 0.645380298627117, "grad_norm": 0.8899328708648682, "learning_rate": 3.2473766541235963e-06, "loss": 1.5382, "step": 25761 }, { "epoch": 0.6461569295520593, "grad_norm": 0.9054051637649536, "learning_rate": 3.2061073473317466e-06, "loss": 1.5604, "step": 25792 }, { "epoch": 0.6469335604770017, "grad_norm": 0.863438606262207, "learning_rate": 3.1650839816563444e-06, "loss": 1.5296, "step": 25823 }, { "epoch": 0.6477101914019441, "grad_norm": 0.8931278586387634, "learning_rate": 3.1243070200405093e-06, "loss": 1.5461, "step": 25854 }, { "epoch": 0.6484868223268865, "grad_norm": 0.9182883501052856, "learning_rate": 3.0837769226467e-06, "loss": 1.5296, "step": 25885 }, { "epoch": 0.6492634532518289, "grad_norm": 0.9026657938957214, "learning_rate": 3.0434941468515666e-06, "loss": 1.5465, "step": 25916 }, { "epoch": 0.6500400841767712, "grad_norm": 0.8940238952636719, "learning_rate": 3.003459147240753e-06, "loss": 1.5197, "step": 25947 }, { "epoch": 0.6508167151017136, "grad_norm": 0.8756072521209717, "learning_rate": 2.9636723756037875e-06, "loss": 1.5416, "step": 25978 }, { "epoch": 0.651593346026656, "grad_norm": 0.9100384712219238, "learning_rate": 2.9241342809289833e-06, "loss": 1.5403, "step": 26009 }, { "epoch": 0.6523699769515984, "grad_norm": 0.9010562300682068, "learning_rate": 2.8848453093983594e-06, "loss": 1.5407, "step": 26040 }, { "epoch": 0.6531466078765408, "grad_norm": 0.8857883214950562, "learning_rate": 2.8458059043826257e-06, "loss": 1.5366, "step": 26071 }, { "epoch": 0.6539232388014831, "grad_norm": 0.8999698758125305, "learning_rate": 2.807016506436172e-06, "loss": 1.529, "step": 26102 }, { "epoch": 0.6546998697264255, "grad_norm": 0.9420500993728638, "learning_rate": 2.7684775532920566e-06, "loss": 1.5195, "step": 26133 }, { "epoch": 0.6554765006513679, "grad_norm": 0.9227805733680725, "learning_rate": 2.7301894798571425e-06, "loss": 1.5563, "step": 26164 }, { "epoch": 0.6562531315763103, "grad_norm": 0.9569822549819946, "learning_rate": 2.6921527182071386e-06, "loss": 1.5239, "step": 26195 }, { "epoch": 0.6570297625012527, "grad_norm": 0.9220489859580994, "learning_rate": 2.654367697581725e-06, "loss": 1.5485, "step": 26226 }, { "epoch": 0.657806393426195, "grad_norm": 0.9232506155967712, "learning_rate": 2.6168348443797175e-06, "loss": 1.5451, "step": 26257 }, { "epoch": 0.6585830243511374, "grad_norm": 0.9334014058113098, "learning_rate": 2.5795545821542757e-06, "loss": 1.5228, "step": 26288 }, { "epoch": 0.6593596552760798, "grad_norm": 0.9631486535072327, "learning_rate": 2.54252733160808e-06, "loss": 1.5595, "step": 26319 }, { "epoch": 0.6601362862010222, "grad_norm": 0.9698452353477478, "learning_rate": 2.5057535105886294e-06, "loss": 1.5273, "step": 26350 }, { "epoch": 0.6609129171259646, "grad_norm": 0.8972797989845276, "learning_rate": 2.4692335340834953e-06, "loss": 1.5373, "step": 26381 }, { "epoch": 0.6616895480509069, "grad_norm": 0.8990224599838257, "learning_rate": 2.432967814215639e-06, "loss": 1.541, "step": 26412 }, { "epoch": 0.6624661789758493, "grad_norm": 0.9032983183860779, "learning_rate": 2.396956760238794e-06, "loss": 1.5189, "step": 26443 }, { "epoch": 0.6632428099007917, "grad_norm": 0.9151228666305542, "learning_rate": 2.361200778532796e-06, "loss": 1.5277, "step": 26474 }, { "epoch": 0.664019440825734, "grad_norm": 0.8908092975616455, "learning_rate": 2.325700272599049e-06, "loss": 1.5431, "step": 26505 }, { "epoch": 0.6647960717506765, "grad_norm": 0.9372129440307617, "learning_rate": 2.2904556430559415e-06, "loss": 1.5373, "step": 26536 }, { "epoch": 0.6655727026756189, "grad_norm": 0.9178755879402161, "learning_rate": 2.2554672876343106e-06, "loss": 1.551, "step": 26567 }, { "epoch": 0.6663493336005611, "grad_norm": 0.9076238870620728, "learning_rate": 2.220735601173002e-06, "loss": 1.5471, "step": 26598 }, { "epoch": 0.6671259645255035, "grad_norm": 0.8949811458587646, "learning_rate": 2.186260975614382e-06, "loss": 1.5498, "step": 26629 }, { "epoch": 0.667902595450446, "grad_norm": 0.9235818386077881, "learning_rate": 2.1520437999999034e-06, "loss": 1.5445, "step": 26660 }, { "epoch": 0.6686792263753883, "grad_norm": 0.884668231010437, "learning_rate": 2.1180844604657526e-06, "loss": 1.5344, "step": 26691 }, { "epoch": 0.6694558573003307, "grad_norm": 0.8701885938644409, "learning_rate": 2.084383340238455e-06, "loss": 1.5268, "step": 26722 }, { "epoch": 0.670232488225273, "grad_norm": 0.9150304198265076, "learning_rate": 2.0509408196305704e-06, "loss": 1.5445, "step": 26753 }, { "epoch": 0.6710091191502154, "grad_norm": 0.9240284562110901, "learning_rate": 2.017757276036403e-06, "loss": 1.5694, "step": 26784 }, { "epoch": 0.6717857500751578, "grad_norm": 0.9412962794303894, "learning_rate": 1.984833083927726e-06, "loss": 1.5657, "step": 26815 }, { "epoch": 0.6725623810001002, "grad_norm": 0.9341692924499512, "learning_rate": 1.952168614849581e-06, "loss": 1.5396, "step": 26846 }, { "epoch": 0.6733390119250426, "grad_norm": 0.9173117876052856, "learning_rate": 1.919764237416058e-06, "loss": 1.5496, "step": 26877 }, { "epoch": 0.6741156428499849, "grad_norm": 0.8984019160270691, "learning_rate": 1.8876203173061463e-06, "loss": 1.5385, "step": 26908 }, { "epoch": 0.6748922737749273, "grad_norm": 0.9131044745445251, "learning_rate": 1.8557372172596206e-06, "loss": 1.5551, "step": 26939 }, { "epoch": 0.6756689046998697, "grad_norm": 0.934843897819519, "learning_rate": 1.8241152970729341e-06, "loss": 1.5497, "step": 26970 }, { "epoch": 0.6764455356248121, "grad_norm": 0.878828227519989, "learning_rate": 1.7927549135951572e-06, "loss": 1.5587, "step": 27001 }, { "epoch": 0.6772221665497545, "grad_norm": 0.9447210431098938, "learning_rate": 1.7616564207239477e-06, "loss": 1.5403, "step": 27032 }, { "epoch": 0.6779987974746968, "grad_norm": 0.9129860997200012, "learning_rate": 1.730820169401584e-06, "loss": 1.5343, "step": 27063 }, { "epoch": 0.6787754283996392, "grad_norm": 0.9327501058578491, "learning_rate": 1.7002465076109558e-06, "loss": 1.5279, "step": 27094 }, { "epoch": 0.6795520593245816, "grad_norm": 0.9292470216751099, "learning_rate": 1.6699357803716898e-06, "loss": 1.5316, "step": 27125 }, { "epoch": 0.680328690249524, "grad_norm": 0.9054301977157593, "learning_rate": 1.6398883297362305e-06, "loss": 1.5257, "step": 27156 }, { "epoch": 0.6811053211744664, "grad_norm": 0.8920509815216064, "learning_rate": 1.6101044947859606e-06, "loss": 1.5451, "step": 27187 }, { "epoch": 0.6818819520994087, "grad_norm": 0.8931530714035034, "learning_rate": 1.5805846116274114e-06, "loss": 1.5316, "step": 27218 }, { "epoch": 0.6826585830243511, "grad_norm": 0.9331194758415222, "learning_rate": 1.5513290133884611e-06, "loss": 1.5352, "step": 27249 }, { "epoch": 0.6834352139492935, "grad_norm": 0.9129217267036438, "learning_rate": 1.5223380302145512e-06, "loss": 1.536, "step": 27280 }, { "epoch": 0.6842118448742359, "grad_norm": 0.9494413137435913, "learning_rate": 1.4936119892649925e-06, "loss": 1.519, "step": 27311 }, { "epoch": 0.6849884757991783, "grad_norm": 0.9152242541313171, "learning_rate": 1.4651512147092482e-06, "loss": 1.5227, "step": 27342 }, { "epoch": 0.6857651067241206, "grad_norm": 0.9150066375732422, "learning_rate": 1.4369560277232908e-06, "loss": 1.5323, "step": 27373 }, { "epoch": 0.686541737649063, "grad_norm": 0.9378058910369873, "learning_rate": 1.409026746485978e-06, "loss": 1.5414, "step": 27404 }, { "epoch": 0.6873183685740054, "grad_norm": 0.9014245867729187, "learning_rate": 1.3813636861754464e-06, "loss": 1.5541, "step": 27435 }, { "epoch": 0.6880949994989478, "grad_norm": 0.9115684628486633, "learning_rate": 1.3539671589655773e-06, "loss": 1.5481, "step": 27466 }, { "epoch": 0.6888716304238902, "grad_norm": 0.9748349785804749, "learning_rate": 1.3268374740224548e-06, "loss": 1.5266, "step": 27497 }, { "epoch": 0.6896482613488325, "grad_norm": 0.9160287976264954, "learning_rate": 1.2999749375008807e-06, "loss": 1.5253, "step": 27528 }, { "epoch": 0.6904248922737749, "grad_norm": 0.9088099002838135, "learning_rate": 1.2733798525409346e-06, "loss": 1.5405, "step": 27559 }, { "epoch": 0.6912015231987173, "grad_norm": 0.9275810718536377, "learning_rate": 1.2470525192645383e-06, "loss": 1.538, "step": 27590 }, { "epoch": 0.6919781541236597, "grad_norm": 0.8710309863090515, "learning_rate": 1.2209932347720666e-06, "loss": 1.5489, "step": 27621 }, { "epoch": 0.6927547850486021, "grad_norm": 0.9207929968833923, "learning_rate": 1.1952022931389972e-06, "loss": 1.5322, "step": 27652 }, { "epoch": 0.6935314159735444, "grad_norm": 0.9486070871353149, "learning_rate": 1.1696799854126083e-06, "loss": 1.5305, "step": 27683 }, { "epoch": 0.6943080468984868, "grad_norm": 0.8901826739311218, "learning_rate": 1.1444265996086694e-06, "loss": 1.5498, "step": 27714 }, { "epoch": 0.6950846778234292, "grad_norm": 0.9303523302078247, "learning_rate": 1.119442420708211e-06, "loss": 1.5485, "step": 27745 }, { "epoch": 0.6958613087483716, "grad_norm": 1.046007513999939, "learning_rate": 1.0947277306542964e-06, "loss": 1.5512, "step": 27776 }, { "epoch": 0.696637939673314, "grad_norm": 0.9494882225990295, "learning_rate": 1.0702828083488353e-06, "loss": 1.5154, "step": 27807 }, { "epoch": 0.6974145705982563, "grad_norm": 0.9443824291229248, "learning_rate": 1.0461079296494647e-06, "loss": 1.5367, "step": 27838 }, { "epoch": 0.6981912015231987, "grad_norm": 0.9392052292823792, "learning_rate": 1.0222033673663978e-06, "loss": 1.5383, "step": 27869 }, { "epoch": 0.6989678324481411, "grad_norm": 0.8942481279373169, "learning_rate": 9.985693912593713e-07, "loss": 1.5513, "step": 27900 }, { "epoch": 0.6997444633730835, "grad_norm": 0.9549274444580078, "learning_rate": 9.752062680346035e-07, "loss": 1.5589, "step": 27931 }, { "epoch": 0.7005210942980259, "grad_norm": 0.8858839869499207, "learning_rate": 9.521142613417494e-07, "loss": 1.527, "step": 27962 }, { "epoch": 0.7012977252229682, "grad_norm": 0.9377241730690002, "learning_rate": 9.292936317709722e-07, "loss": 1.5354, "step": 27993 }, { "epoch": 0.7020743561479106, "grad_norm": 0.9183421730995178, "learning_rate": 9.067446368499793e-07, "loss": 1.5365, "step": 28024 }, { "epoch": 0.702850987072853, "grad_norm": 0.9407469630241394, "learning_rate": 8.844675310411055e-07, "loss": 1.5392, "step": 28055 }, { "epoch": 0.7036276179977954, "grad_norm": 0.9024103879928589, "learning_rate": 8.6246256573847e-07, "loss": 1.5275, "step": 28086 }, { "epoch": 0.7044042489227378, "grad_norm": 0.9077733755111694, "learning_rate": 8.407299892651127e-07, "loss": 1.5285, "step": 28117 }, { "epoch": 0.7051808798476801, "grad_norm": 0.8907241821289062, "learning_rate": 8.19270046870202e-07, "loss": 1.5283, "step": 28148 }, { "epoch": 0.7059575107726225, "grad_norm": 0.9342151284217834, "learning_rate": 7.980829807262752e-07, "loss": 1.5493, "step": 28179 }, { "epoch": 0.7067341416975649, "grad_norm": 0.9555601477622986, "learning_rate": 7.771690299264889e-07, "loss": 1.5452, "step": 28210 }, { "epoch": 0.7075107726225073, "grad_norm": 0.9291635155677795, "learning_rate": 7.565284304819426e-07, "loss": 1.5286, "step": 28241 }, { "epoch": 0.7082874035474497, "grad_norm": 0.9101645946502686, "learning_rate": 7.361614153189922e-07, "loss": 1.5085, "step": 28272 }, { "epoch": 0.709064034472392, "grad_norm": 0.9269605278968811, "learning_rate": 7.160682142766328e-07, "loss": 1.5366, "step": 28303 }, { "epoch": 0.7098406653973344, "grad_norm": 0.9426531791687012, "learning_rate": 6.962490541039091e-07, "loss": 1.5571, "step": 28334 }, { "epoch": 0.7106172963222768, "grad_norm": 0.9087902903556824, "learning_rate": 6.767041584573531e-07, "loss": 1.5253, "step": 28365 }, { "epoch": 0.7113939272472192, "grad_norm": 0.9450654983520508, "learning_rate": 6.574337478984532e-07, "loss": 1.5361, "step": 28396 }, { "epoch": 0.7121705581721616, "grad_norm": 0.9181643128395081, "learning_rate": 6.384380398911732e-07, "loss": 1.5346, "step": 28427 }, { "epoch": 0.712947189097104, "grad_norm": 0.9480751752853394, "learning_rate": 6.197172487994951e-07, "loss": 1.5218, "step": 28458 }, { "epoch": 0.7137238200220463, "grad_norm": 0.947593629360199, "learning_rate": 6.012715858850021e-07, "loss": 1.5611, "step": 28489 }, { "epoch": 0.7145004509469887, "grad_norm": 0.9011168479919434, "learning_rate": 5.831012593044971e-07, "loss": 1.5437, "step": 28520 }, { "epoch": 0.7152770818719311, "grad_norm": 0.8987276554107666, "learning_rate": 5.652064741076435e-07, "loss": 1.5348, "step": 28551 }, { "epoch": 0.7160537127968735, "grad_norm": 0.9200072288513184, "learning_rate": 5.475874322346558e-07, "loss": 1.5614, "step": 28582 }, { "epoch": 0.7168303437218159, "grad_norm": 0.9210062026977539, "learning_rate": 5.30244332514035e-07, "loss": 1.5238, "step": 28613 }, { "epoch": 0.7176069746467582, "grad_norm": 0.9360731244087219, "learning_rate": 5.131773706602977e-07, "loss": 1.544, "step": 28644 }, { "epoch": 0.7183836055717006, "grad_norm": 0.9032440185546875, "learning_rate": 4.963867392717897e-07, "loss": 1.5359, "step": 28675 }, { "epoch": 0.719160236496643, "grad_norm": 0.910786509513855, "learning_rate": 4.798726278285093e-07, "loss": 1.5312, "step": 28706 }, { "epoch": 0.7199368674215854, "grad_norm": 0.9738684296607971, "learning_rate": 4.6363522268995097e-07, "loss": 1.5173, "step": 28737 }, { "epoch": 0.7207134983465278, "grad_norm": 0.9202192425727844, "learning_rate": 4.4767470709302927e-07, "loss": 1.5445, "step": 28768 }, { "epoch": 0.72149012927147, "grad_norm": 0.9537370800971985, "learning_rate": 4.319912611499971e-07, "loss": 1.5318, "step": 28799 }, { "epoch": 0.7222667601964124, "grad_norm": 0.9033458232879639, "learning_rate": 4.1658506184640564e-07, "loss": 1.5215, "step": 28830 }, { "epoch": 0.7230433911213549, "grad_norm": 0.897890567779541, "learning_rate": 4.0145628303911996e-07, "loss": 1.532, "step": 28861 }, { "epoch": 0.7238200220462973, "grad_norm": 0.9329019784927368, "learning_rate": 3.866050954543565e-07, "loss": 1.5374, "step": 28892 }, { "epoch": 0.7245966529712397, "grad_norm": 0.9132122993469238, "learning_rate": 3.720316666857432e-07, "loss": 1.5292, "step": 28923 }, { "epoch": 0.7253732838961819, "grad_norm": 0.9273849129676819, "learning_rate": 3.5773616119244845e-07, "loss": 1.5229, "step": 28954 }, { "epoch": 0.7261499148211243, "grad_norm": 0.8980498909950256, "learning_rate": 3.437187402973052e-07, "loss": 1.5167, "step": 28985 }, { "epoch": 0.7269265457460667, "grad_norm": 0.9546352028846741, "learning_rate": 3.2997956218500104e-07, "loss": 1.5321, "step": 29016 }, { "epoch": 0.7277031766710091, "grad_norm": 0.87969970703125, "learning_rate": 3.165187819003018e-07, "loss": 1.5308, "step": 29047 }, { "epoch": 0.7284798075959515, "grad_norm": 0.9020631313323975, "learning_rate": 3.033365513462755e-07, "loss": 1.5387, "step": 29078 }, { "epoch": 0.7292564385208938, "grad_norm": 0.9020014405250549, "learning_rate": 2.9043301928260437e-07, "loss": 1.5278, "step": 29109 }, { "epoch": 0.7300330694458362, "grad_norm": 0.8693344593048096, "learning_rate": 2.7780833132389773e-07, "loss": 1.5211, "step": 29140 }, { "epoch": 0.7308097003707786, "grad_norm": 0.9200028777122498, "learning_rate": 2.6546262993803473e-07, "loss": 1.5627, "step": 29171 }, { "epoch": 0.731586331295721, "grad_norm": 0.8627704977989197, "learning_rate": 2.533960544445879e-07, "loss": 1.553, "step": 29202 }, { "epoch": 0.7323629622206634, "grad_norm": 0.9087816476821899, "learning_rate": 2.416087410132134e-07, "loss": 1.5218, "step": 29233 }, { "epoch": 0.7331395931456057, "grad_norm": 0.9687427282333374, "learning_rate": 2.301008226621465e-07, "loss": 1.5471, "step": 29264 }, { "epoch": 0.7339162240705481, "grad_norm": 0.9520723223686218, "learning_rate": 2.1887242925668073e-07, "loss": 1.5318, "step": 29295 }, { "epoch": 0.7346928549954905, "grad_norm": 0.916723906993866, "learning_rate": 2.0792368750770785e-07, "loss": 1.5413, "step": 29326 }, { "epoch": 0.7354694859204329, "grad_norm": 0.9351438283920288, "learning_rate": 1.9725472097028851e-07, "loss": 1.5393, "step": 29357 }, { "epoch": 0.7362461168453753, "grad_norm": 0.9236180186271667, "learning_rate": 1.8686565004226718e-07, "loss": 1.5314, "step": 29388 }, { "epoch": 0.7370227477703176, "grad_norm": 0.921168327331543, "learning_rate": 1.7675659196288995e-07, "loss": 1.5469, "step": 29419 }, { "epoch": 0.73779937869526, "grad_norm": 0.8722018599510193, "learning_rate": 1.6692766081150556e-07, "loss": 1.5183, "step": 29450 }, { "epoch": 0.7385760096202024, "grad_norm": 0.9068271517753601, "learning_rate": 1.5737896750626647e-07, "loss": 1.5467, "step": 29481 }, { "epoch": 0.7393526405451448, "grad_norm": 0.8897390961647034, "learning_rate": 1.4811061980287976e-07, "loss": 1.5517, "step": 29512 }, { "epoch": 0.7401292714700872, "grad_norm": 0.8943518996238708, "learning_rate": 1.3912272229338886e-07, "loss": 1.5143, "step": 29543 }, { "epoch": 0.7409059023950295, "grad_norm": 0.9387959241867065, "learning_rate": 1.3041537640499645e-07, "loss": 1.5579, "step": 29574 }, { "epoch": 0.7416825333199719, "grad_norm": 0.9111166000366211, "learning_rate": 1.2198868039891564e-07, "loss": 1.5281, "step": 29605 }, { "epoch": 0.7424591642449143, "grad_norm": 0.9361575245857239, "learning_rate": 1.138427293692651e-07, "loss": 1.5488, "step": 29636 }, { "epoch": 0.7432357951698567, "grad_norm": 0.9410037398338318, "learning_rate": 1.0597761524199778e-07, "loss": 1.5175, "step": 29667 }, { "epoch": 0.7440124260947991, "grad_norm": 0.9299407601356506, "learning_rate": 9.839342677385455e-08, "loss": 1.5295, "step": 29698 }, { "epoch": 0.7447890570197414, "grad_norm": 0.9386714100837708, "learning_rate": 9.109024955137325e-08, "loss": 1.5498, "step": 29729 }, { "epoch": 0.7455656879446838, "grad_norm": 0.9419563412666321, "learning_rate": 8.406816598991729e-08, "loss": 1.5222, "step": 29760 }, { "epoch": 0.7463423188696262, "grad_norm": 0.9137088656425476, "learning_rate": 7.73272553327431e-08, "loss": 1.5529, "step": 29791 }, { "epoch": 0.7471189497945686, "grad_norm": 0.9149454236030579, "learning_rate": 7.086759365011186e-08, "loss": 1.5208, "step": 29822 }, { "epoch": 0.747895580719511, "grad_norm": 0.9159924387931824, "learning_rate": 6.468925383842639e-08, "loss": 1.5515, "step": 29853 }, { "epoch": 0.7486722116444533, "grad_norm": 0.9082822203636169, "learning_rate": 5.8792305619415067e-08, "loss": 1.5258, "step": 29884 }, { "epoch": 0.7494488425693957, "grad_norm": 0.9176669120788574, "learning_rate": 5.317681553933529e-08, "loss": 1.5567, "step": 29915 }, { "epoch": 0.7502254734943381, "grad_norm": 0.9306588172912598, "learning_rate": 4.78428469682296e-08, "loss": 1.5395, "step": 29946 }, { "epoch": 0.7510021044192805, "grad_norm": 0.9428490996360779, "learning_rate": 4.2790460099206844e-08, "loss": 1.5498, "step": 29977 }, { "epoch": 0.7517787353442229, "grad_norm": 0.8921723961830139, "learning_rate": 3.801971194777043e-08, "loss": 1.5456, "step": 30008 }, { "epoch": 0.7525553662691652, "grad_norm": 0.9158675670623779, "learning_rate": 3.353065635115782e-08, "loss": 1.5148, "step": 30039 }, { "epoch": 0.7533319971941076, "grad_norm": 0.9300774335861206, "learning_rate": 2.93233439677576e-08, "loss": 1.5298, "step": 30070 }, { "epoch": 0.75410862811905, "grad_norm": 0.8786063194274902, "learning_rate": 2.539782227651555e-08, "loss": 1.5023, "step": 30101 }, { "epoch": 0.7548852590439924, "grad_norm": 0.9041824340820312, "learning_rate": 2.175413557641004e-08, "loss": 1.5372, "step": 30132 }, { "epoch": 0.7556618899689348, "grad_norm": 0.9231988787651062, "learning_rate": 1.839232498594967e-08, "loss": 1.5502, "step": 30163 }, { "epoch": 0.7564385208938771, "grad_norm": 0.9001336097717285, "learning_rate": 1.5312428442712522e-08, "loss": 1.5239, "step": 30194 }, { "epoch": 0.7572151518188195, "grad_norm": 0.9294939637184143, "learning_rate": 1.2514480702913168e-08, "loss": 1.5117, "step": 30225 }, { "epoch": 0.7579917827437619, "grad_norm": 0.874643087387085, "learning_rate": 9.998513341005766e-09, "loss": 1.5402, "step": 30256 }, { "epoch": 0.7587684136687043, "grad_norm": 0.8916767239570618, "learning_rate": 7.764554749345454e-09, "loss": 1.5303, "step": 30287 }, { "epoch": 0.7595450445936467, "grad_norm": 0.9058026075363159, "learning_rate": 5.812630137849717e-09, "loss": 1.5194, "step": 30318 }, { "epoch": 0.7603216755185891, "grad_norm": 0.898659348487854, "learning_rate": 4.142761533723616e-09, "loss": 1.509, "step": 30349 }, { "epoch": 0.7610983064435314, "grad_norm": 0.8996184468269348, "learning_rate": 2.7549677812044317e-09, "loss": 1.527, "step": 30380 }, { "epoch": 0.7618749373684738, "grad_norm": 0.9288005232810974, "learning_rate": 1.6492645413590525e-09, "loss": 1.5369, "step": 30411 }, { "epoch": 0.7626515682934162, "grad_norm": 0.9407172203063965, "learning_rate": 8.256642918980096e-10, "loss": 1.5501, "step": 30442 }, { "epoch": 0.7634281992183586, "grad_norm": 0.905489444732666, "learning_rate": 2.841763270367004e-10, "loss": 1.5464, "step": 30473 }, { "epoch": 0.764204830143301, "grad_norm": 0.8837622404098511, "learning_rate": 2.480675739269245e-11, "loss": 1.531, "step": 30504 } ], "logging_steps": 31, "max_steps": 30517, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 3052, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.131475882955925e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }