atsuki-yamaguchi's picture
Upload folder using huggingface_hub
0bccc5f verified
raw
history blame
173 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.7645305140795671,
"eval_steps": 500,
"global_step": 30517,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.000776630924942379,
"grad_norm": 23.55321502685547,
"learning_rate": 1.0157273918741808e-06,
"loss": 8.3443,
"step": 31
},
{
"epoch": 0.001553261849884758,
"grad_norm": 14.414477348327637,
"learning_rate": 2.0314547837483616e-06,
"loss": 7.198,
"step": 62
},
{
"epoch": 0.002329892774827137,
"grad_norm": 10.739773750305176,
"learning_rate": 3.0471821756225426e-06,
"loss": 5.8639,
"step": 93
},
{
"epoch": 0.003106523699769516,
"grad_norm": 5.378795623779297,
"learning_rate": 4.062909567496723e-06,
"loss": 4.8222,
"step": 124
},
{
"epoch": 0.003883154624711895,
"grad_norm": 8.51831340789795,
"learning_rate": 5.078636959370905e-06,
"loss": 4.1983,
"step": 155
},
{
"epoch": 0.004659785549654274,
"grad_norm": 6.32632303237915,
"learning_rate": 6.094364351245085e-06,
"loss": 3.8598,
"step": 186
},
{
"epoch": 0.005436416474596653,
"grad_norm": 5.590794563293457,
"learning_rate": 7.110091743119267e-06,
"loss": 3.6646,
"step": 217
},
{
"epoch": 0.006213047399539032,
"grad_norm": 5.9275383949279785,
"learning_rate": 8.125819134993446e-06,
"loss": 3.5212,
"step": 248
},
{
"epoch": 0.006989678324481411,
"grad_norm": 5.041291236877441,
"learning_rate": 9.141546526867629e-06,
"loss": 3.3936,
"step": 279
},
{
"epoch": 0.00776630924942379,
"grad_norm": 4.923459529876709,
"learning_rate": 1.015727391874181e-05,
"loss": 3.2713,
"step": 310
},
{
"epoch": 0.00854294017436617,
"grad_norm": 8.011459350585938,
"learning_rate": 1.117300131061599e-05,
"loss": 3.2,
"step": 341
},
{
"epoch": 0.009319571099308548,
"grad_norm": 5.261505603790283,
"learning_rate": 1.218872870249017e-05,
"loss": 3.1255,
"step": 372
},
{
"epoch": 0.010096202024250926,
"grad_norm": 5.004406452178955,
"learning_rate": 1.3204456094364351e-05,
"loss": 3.0726,
"step": 403
},
{
"epoch": 0.010872832949193306,
"grad_norm": 4.326263427734375,
"learning_rate": 1.4220183486238533e-05,
"loss": 2.9682,
"step": 434
},
{
"epoch": 0.011649463874135685,
"grad_norm": 4.251937389373779,
"learning_rate": 1.5235910878112714e-05,
"loss": 2.9215,
"step": 465
},
{
"epoch": 0.012426094799078065,
"grad_norm": 4.35053825378418,
"learning_rate": 1.6251638269986893e-05,
"loss": 2.8902,
"step": 496
},
{
"epoch": 0.013202725724020443,
"grad_norm": 4.316098690032959,
"learning_rate": 1.7267365661861077e-05,
"loss": 2.8521,
"step": 527
},
{
"epoch": 0.013979356648962823,
"grad_norm": 5.536881446838379,
"learning_rate": 1.8283093053735257e-05,
"loss": 2.8058,
"step": 558
},
{
"epoch": 0.014755987573905201,
"grad_norm": 3.759843111038208,
"learning_rate": 1.9298820445609438e-05,
"loss": 2.7438,
"step": 589
},
{
"epoch": 0.01553261849884758,
"grad_norm": 4.057703018188477,
"learning_rate": 2.031454783748362e-05,
"loss": 2.7508,
"step": 620
},
{
"epoch": 0.01630924942378996,
"grad_norm": 3.427708625793457,
"learning_rate": 2.13302752293578e-05,
"loss": 2.7046,
"step": 651
},
{
"epoch": 0.01708588034873234,
"grad_norm": 3.9012484550476074,
"learning_rate": 2.234600262123198e-05,
"loss": 2.6706,
"step": 682
},
{
"epoch": 0.017862511273674716,
"grad_norm": 3.388504981994629,
"learning_rate": 2.336173001310616e-05,
"loss": 2.6377,
"step": 713
},
{
"epoch": 0.018639142198617096,
"grad_norm": 3.2011327743530273,
"learning_rate": 2.437745740498034e-05,
"loss": 2.6214,
"step": 744
},
{
"epoch": 0.019415773123559476,
"grad_norm": 3.5897209644317627,
"learning_rate": 2.5393184796854525e-05,
"loss": 2.6045,
"step": 775
},
{
"epoch": 0.020192404048501852,
"grad_norm": 3.386697769165039,
"learning_rate": 2.6408912188728702e-05,
"loss": 2.6066,
"step": 806
},
{
"epoch": 0.020969034973444232,
"grad_norm": 2.9135866165161133,
"learning_rate": 2.7424639580602886e-05,
"loss": 2.5697,
"step": 837
},
{
"epoch": 0.021745665898386612,
"grad_norm": 9.212065696716309,
"learning_rate": 2.8440366972477066e-05,
"loss": 2.5773,
"step": 868
},
{
"epoch": 0.022522296823328992,
"grad_norm": 3.0339765548706055,
"learning_rate": 2.9456094364351244e-05,
"loss": 2.5453,
"step": 899
},
{
"epoch": 0.02329892774827137,
"grad_norm": 2.8007607460021973,
"learning_rate": 3.0471821756225428e-05,
"loss": 2.5297,
"step": 930
},
{
"epoch": 0.02407555867321375,
"grad_norm": 3.251768112182617,
"learning_rate": 3.148754914809961e-05,
"loss": 2.5273,
"step": 961
},
{
"epoch": 0.02485218959815613,
"grad_norm": 3.357011079788208,
"learning_rate": 3.2503276539973785e-05,
"loss": 2.4896,
"step": 992
},
{
"epoch": 0.025628820523098506,
"grad_norm": 2.6069107055664062,
"learning_rate": 3.351900393184797e-05,
"loss": 2.4718,
"step": 1023
},
{
"epoch": 0.026405451448040886,
"grad_norm": 4.663620948791504,
"learning_rate": 3.453473132372215e-05,
"loss": 2.4338,
"step": 1054
},
{
"epoch": 0.027182082372983266,
"grad_norm": 3.819638252258301,
"learning_rate": 3.555045871559633e-05,
"loss": 2.4848,
"step": 1085
},
{
"epoch": 0.027958713297925646,
"grad_norm": 2.6367456912994385,
"learning_rate": 3.6566186107470514e-05,
"loss": 2.4415,
"step": 1116
},
{
"epoch": 0.028735344222868022,
"grad_norm": 2.6118173599243164,
"learning_rate": 3.7581913499344695e-05,
"loss": 2.4341,
"step": 1147
},
{
"epoch": 0.029511975147810402,
"grad_norm": 9.820366859436035,
"learning_rate": 3.8597640891218876e-05,
"loss": 2.3773,
"step": 1178
},
{
"epoch": 0.030288606072752782,
"grad_norm": 38.890506744384766,
"learning_rate": 3.9613368283093056e-05,
"loss": 2.3966,
"step": 1209
},
{
"epoch": 0.03106523699769516,
"grad_norm": 2.5037872791290283,
"learning_rate": 4.062909567496724e-05,
"loss": 2.3421,
"step": 1240
},
{
"epoch": 0.03184186792263754,
"grad_norm": 2.426544189453125,
"learning_rate": 4.164482306684142e-05,
"loss": 2.3568,
"step": 1271
},
{
"epoch": 0.03261849884757992,
"grad_norm": 2.5375614166259766,
"learning_rate": 4.26605504587156e-05,
"loss": 2.3314,
"step": 1302
},
{
"epoch": 0.0333951297725223,
"grad_norm": 3.555330276489258,
"learning_rate": 4.367627785058978e-05,
"loss": 2.3262,
"step": 1333
},
{
"epoch": 0.03417176069746468,
"grad_norm": 2.2571263313293457,
"learning_rate": 4.469200524246396e-05,
"loss": 2.3167,
"step": 1364
},
{
"epoch": 0.03494839162240705,
"grad_norm": 2.184305429458618,
"learning_rate": 4.570773263433814e-05,
"loss": 2.2936,
"step": 1395
},
{
"epoch": 0.03572502254734943,
"grad_norm": 2.8295061588287354,
"learning_rate": 4.672346002621232e-05,
"loss": 2.3033,
"step": 1426
},
{
"epoch": 0.03650165347229181,
"grad_norm": 2.0241498947143555,
"learning_rate": 4.77391874180865e-05,
"loss": 2.2833,
"step": 1457
},
{
"epoch": 0.03727828439723419,
"grad_norm": 2.0029919147491455,
"learning_rate": 4.875491480996068e-05,
"loss": 2.2663,
"step": 1488
},
{
"epoch": 0.03805491532217657,
"grad_norm": 2.033773899078369,
"learning_rate": 4.977064220183487e-05,
"loss": 2.2597,
"step": 1519
},
{
"epoch": 0.03883154624711895,
"grad_norm": 2.047210216522217,
"learning_rate": 4.9999915451558777e-05,
"loss": 2.2231,
"step": 1550
},
{
"epoch": 0.03960817717206133,
"grad_norm": 2.032099485397339,
"learning_rate": 4.999955597496219e-05,
"loss": 2.2411,
"step": 1581
},
{
"epoch": 0.040384808097003705,
"grad_norm": 1.8384301662445068,
"learning_rate": 4.9998914381774255e-05,
"loss": 2.2409,
"step": 1612
},
{
"epoch": 0.041161439021946085,
"grad_norm": 1.721187710762024,
"learning_rate": 4.999799067923527e-05,
"loss": 2.193,
"step": 1643
},
{
"epoch": 0.041938069946888465,
"grad_norm": 1.8869743347167969,
"learning_rate": 4.999678487776908e-05,
"loss": 2.1917,
"step": 1674
},
{
"epoch": 0.042714700871830845,
"grad_norm": 1.796583652496338,
"learning_rate": 4.9995296990983006e-05,
"loss": 2.1794,
"step": 1705
},
{
"epoch": 0.043491331796773225,
"grad_norm": 1.6308804750442505,
"learning_rate": 4.999352703566763e-05,
"loss": 2.1854,
"step": 1736
},
{
"epoch": 0.044267962721715605,
"grad_norm": 1.683474063873291,
"learning_rate": 4.999147503179668e-05,
"loss": 2.1797,
"step": 1767
},
{
"epoch": 0.045044593646657985,
"grad_norm": 1.7442835569381714,
"learning_rate": 4.998914100252672e-05,
"loss": 2.1695,
"step": 1798
},
{
"epoch": 0.04582122457160036,
"grad_norm": 1.704970359802246,
"learning_rate": 4.998652497419696e-05,
"loss": 2.1365,
"step": 1829
},
{
"epoch": 0.04659785549654274,
"grad_norm": 1.7213023900985718,
"learning_rate": 4.9983626976328927e-05,
"loss": 2.1524,
"step": 1860
},
{
"epoch": 0.04737448642148512,
"grad_norm": 4.01851224899292,
"learning_rate": 4.998044704162613e-05,
"loss": 2.1171,
"step": 1891
},
{
"epoch": 0.0481511173464275,
"grad_norm": 1.5619858503341675,
"learning_rate": 4.9976985205973705e-05,
"loss": 2.1608,
"step": 1922
},
{
"epoch": 0.04892774827136988,
"grad_norm": 1.684493064880371,
"learning_rate": 4.997324150843799e-05,
"loss": 2.1278,
"step": 1953
},
{
"epoch": 0.04970437919631226,
"grad_norm": 1.660992980003357,
"learning_rate": 4.99692159912661e-05,
"loss": 2.1136,
"step": 1984
},
{
"epoch": 0.05048101012125464,
"grad_norm": 1.5491065979003906,
"learning_rate": 4.996490869988546e-05,
"loss": 2.1086,
"step": 2015
},
{
"epoch": 0.05125764104619701,
"grad_norm": 1.5400840044021606,
"learning_rate": 4.996031968290326e-05,
"loss": 2.1069,
"step": 2046
},
{
"epoch": 0.05203427197113939,
"grad_norm": 1.5687416791915894,
"learning_rate": 4.995544899210594e-05,
"loss": 2.0888,
"step": 2077
},
{
"epoch": 0.05281090289608177,
"grad_norm": 1.5218935012817383,
"learning_rate": 4.9950296682458583e-05,
"loss": 2.1032,
"step": 2108
},
{
"epoch": 0.05358753382102415,
"grad_norm": 1.763666033744812,
"learning_rate": 4.994486281210429e-05,
"loss": 2.0799,
"step": 2139
},
{
"epoch": 0.05436416474596653,
"grad_norm": 1.5606274604797363,
"learning_rate": 4.9939147442363566e-05,
"loss": 2.0614,
"step": 2170
},
{
"epoch": 0.05514079567090891,
"grad_norm": 1.4900213479995728,
"learning_rate": 4.9933150637733574e-05,
"loss": 2.0565,
"step": 2201
},
{
"epoch": 0.05591742659585129,
"grad_norm": 1.4454214572906494,
"learning_rate": 4.992687246588743e-05,
"loss": 2.0381,
"step": 2232
},
{
"epoch": 0.056694057520793664,
"grad_norm": 1.4872742891311646,
"learning_rate": 4.992031299767347e-05,
"loss": 2.0227,
"step": 2263
},
{
"epoch": 0.057470688445736044,
"grad_norm": 1.4538527727127075,
"learning_rate": 4.9913472307114386e-05,
"loss": 2.0395,
"step": 2294
},
{
"epoch": 0.058247319370678424,
"grad_norm": 1.432769536972046,
"learning_rate": 4.9906350471406446e-05,
"loss": 2.0352,
"step": 2325
},
{
"epoch": 0.059023950295620804,
"grad_norm": 1.4535977840423584,
"learning_rate": 4.989894757091861e-05,
"loss": 2.0341,
"step": 2356
},
{
"epoch": 0.059800581220563184,
"grad_norm": 1.4004398584365845,
"learning_rate": 4.989126368919158e-05,
"loss": 2.0185,
"step": 2387
},
{
"epoch": 0.060577212145505564,
"grad_norm": 1.433273434638977,
"learning_rate": 4.988329891293693e-05,
"loss": 2.036,
"step": 2418
},
{
"epoch": 0.06135384307044794,
"grad_norm": 1.3791290521621704,
"learning_rate": 4.987505333203608e-05,
"loss": 2.0421,
"step": 2449
},
{
"epoch": 0.06213047399539032,
"grad_norm": 2.008694648742676,
"learning_rate": 4.9866527039539276e-05,
"loss": 1.9948,
"step": 2480
},
{
"epoch": 0.0629071049203327,
"grad_norm": 1.3773164749145508,
"learning_rate": 4.9857720131664594e-05,
"loss": 2.0158,
"step": 2511
},
{
"epoch": 0.06368373584527508,
"grad_norm": 1.4489777088165283,
"learning_rate": 4.9848632707796773e-05,
"loss": 2.0216,
"step": 2542
},
{
"epoch": 0.06446036677021745,
"grad_norm": 2.1266908645629883,
"learning_rate": 4.9839264870486155e-05,
"loss": 1.9962,
"step": 2573
},
{
"epoch": 0.06523699769515984,
"grad_norm": 1.3704332113265991,
"learning_rate": 4.9829616725447526e-05,
"loss": 1.9801,
"step": 2604
},
{
"epoch": 0.06601362862010221,
"grad_norm": 1.3643279075622559,
"learning_rate": 4.981968838155888e-05,
"loss": 1.9902,
"step": 2635
},
{
"epoch": 0.0667902595450446,
"grad_norm": 1.3495409488677979,
"learning_rate": 4.980947995086024e-05,
"loss": 1.993,
"step": 2666
},
{
"epoch": 0.06756689046998697,
"grad_norm": 1.3918874263763428,
"learning_rate": 4.979899154855234e-05,
"loss": 1.9782,
"step": 2697
},
{
"epoch": 0.06834352139492936,
"grad_norm": 1.3205535411834717,
"learning_rate": 4.9788223292995386e-05,
"loss": 1.9638,
"step": 2728
},
{
"epoch": 0.06912015231987173,
"grad_norm": 1.3123114109039307,
"learning_rate": 4.977717530570768e-05,
"loss": 1.9826,
"step": 2759
},
{
"epoch": 0.0698967832448141,
"grad_norm": 1.3138995170593262,
"learning_rate": 4.976584771136425e-05,
"loss": 1.9938,
"step": 2790
},
{
"epoch": 0.07067341416975649,
"grad_norm": 1.2964009046554565,
"learning_rate": 4.975424063779547e-05,
"loss": 1.9924,
"step": 2821
},
{
"epoch": 0.07145004509469886,
"grad_norm": 1.3010739088058472,
"learning_rate": 4.974235421598557e-05,
"loss": 1.9592,
"step": 2852
},
{
"epoch": 0.07222667601964125,
"grad_norm": 1.2549103498458862,
"learning_rate": 4.973018858007122e-05,
"loss": 1.9394,
"step": 2883
},
{
"epoch": 0.07300330694458362,
"grad_norm": 1.3198585510253906,
"learning_rate": 4.9717743867339963e-05,
"loss": 1.989,
"step": 2914
},
{
"epoch": 0.07377993786952601,
"grad_norm": 1.3653398752212524,
"learning_rate": 4.9705020218228695e-05,
"loss": 1.9499,
"step": 2945
},
{
"epoch": 0.07455656879446838,
"grad_norm": 4.606905937194824,
"learning_rate": 4.969201777632205e-05,
"loss": 1.9504,
"step": 2976
},
{
"epoch": 0.07533319971941076,
"grad_norm": 1.2823129892349243,
"learning_rate": 4.9678736688350846e-05,
"loss": 1.9483,
"step": 3007
},
{
"epoch": 0.07610983064435314,
"grad_norm": 1.2803308963775635,
"learning_rate": 4.966517710419033e-05,
"loss": 1.9604,
"step": 3038
},
{
"epoch": 0.07688646156929552,
"grad_norm": 1.2765967845916748,
"learning_rate": 4.965133917685858e-05,
"loss": 1.941,
"step": 3069
},
{
"epoch": 0.0776630924942379,
"grad_norm": 2.622286796569824,
"learning_rate": 4.9637223062514714e-05,
"loss": 1.9426,
"step": 3100
},
{
"epoch": 0.07843972341918028,
"grad_norm": 1.231387972831726,
"learning_rate": 4.962282892045718e-05,
"loss": 1.9416,
"step": 3131
},
{
"epoch": 0.07921635434412266,
"grad_norm": 1.2499909400939941,
"learning_rate": 4.9608156913121904e-05,
"loss": 1.9226,
"step": 3162
},
{
"epoch": 0.07999298526906504,
"grad_norm": 1.296276330947876,
"learning_rate": 4.959320720608049e-05,
"loss": 1.935,
"step": 3193
},
{
"epoch": 0.08076961619400741,
"grad_norm": 1.2642258405685425,
"learning_rate": 4.9577979968038354e-05,
"loss": 1.9185,
"step": 3224
},
{
"epoch": 0.0815462471189498,
"grad_norm": 1.2396386861801147,
"learning_rate": 4.956247537083282e-05,
"loss": 1.9361,
"step": 3255
},
{
"epoch": 0.08232287804389217,
"grad_norm": 1.1811271905899048,
"learning_rate": 4.9546693589431145e-05,
"loss": 1.9276,
"step": 3286
},
{
"epoch": 0.08309950896883456,
"grad_norm": 1.2897577285766602,
"learning_rate": 4.9530634801928595e-05,
"loss": 1.9277,
"step": 3317
},
{
"epoch": 0.08387613989377693,
"grad_norm": 1.1453903913497925,
"learning_rate": 4.9514299189546395e-05,
"loss": 1.9109,
"step": 3348
},
{
"epoch": 0.08465277081871932,
"grad_norm": 1.1600992679595947,
"learning_rate": 4.949768693662973e-05,
"loss": 1.9248,
"step": 3379
},
{
"epoch": 0.08542940174366169,
"grad_norm": 1.234217882156372,
"learning_rate": 4.948079823064559e-05,
"loss": 1.9236,
"step": 3410
},
{
"epoch": 0.08620603266860406,
"grad_norm": 1.1933766603469849,
"learning_rate": 4.946363326218074e-05,
"loss": 1.9284,
"step": 3441
},
{
"epoch": 0.08698266359354645,
"grad_norm": 1.2087618112564087,
"learning_rate": 4.9446192224939525e-05,
"loss": 1.8971,
"step": 3472
},
{
"epoch": 0.08775929451848882,
"grad_norm": 1.227864146232605,
"learning_rate": 4.942847531574167e-05,
"loss": 1.9353,
"step": 3503
},
{
"epoch": 0.08853592544343121,
"grad_norm": 1.2828229665756226,
"learning_rate": 4.941048273452008e-05,
"loss": 1.9085,
"step": 3534
},
{
"epoch": 0.08931255636837358,
"grad_norm": 1.2086588144302368,
"learning_rate": 4.9392214684318605e-05,
"loss": 1.9044,
"step": 3565
},
{
"epoch": 0.09008918729331597,
"grad_norm": 1.1942152976989746,
"learning_rate": 4.93736713712897e-05,
"loss": 1.9182,
"step": 3596
},
{
"epoch": 0.09086581821825834,
"grad_norm": 1.2399394512176514,
"learning_rate": 4.9354853004692124e-05,
"loss": 1.8943,
"step": 3627
},
{
"epoch": 0.09164244914320072,
"grad_norm": 1.227376937866211,
"learning_rate": 4.93357597968886e-05,
"loss": 1.8881,
"step": 3658
},
{
"epoch": 0.0924190800681431,
"grad_norm": 1.1314102411270142,
"learning_rate": 4.931639196334338e-05,
"loss": 1.8878,
"step": 3689
},
{
"epoch": 0.09319571099308548,
"grad_norm": 1.192134141921997,
"learning_rate": 4.9296749722619826e-05,
"loss": 1.8857,
"step": 3720
},
{
"epoch": 0.09397234191802786,
"grad_norm": 1.1905503273010254,
"learning_rate": 4.9276833296377966e-05,
"loss": 1.8701,
"step": 3751
},
{
"epoch": 0.09474897284297024,
"grad_norm": 1.2335184812545776,
"learning_rate": 4.925664290937196e-05,
"loss": 1.8741,
"step": 3782
},
{
"epoch": 0.09552560376791262,
"grad_norm": 1.1403234004974365,
"learning_rate": 4.9236178789447576e-05,
"loss": 1.8658,
"step": 3813
},
{
"epoch": 0.096302234692855,
"grad_norm": 1.1734315156936646,
"learning_rate": 4.921544116753962e-05,
"loss": 1.8724,
"step": 3844
},
{
"epoch": 0.09707886561779737,
"grad_norm": 1.2637231349945068,
"learning_rate": 4.919443027766935e-05,
"loss": 1.867,
"step": 3875
},
{
"epoch": 0.09785549654273976,
"grad_norm": 1.1608805656433105,
"learning_rate": 4.91731463569418e-05,
"loss": 1.8764,
"step": 3906
},
{
"epoch": 0.09863212746768213,
"grad_norm": 1.1562029123306274,
"learning_rate": 4.915158964554312e-05,
"loss": 1.8612,
"step": 3937
},
{
"epoch": 0.09940875839262452,
"grad_norm": 1.1918668746948242,
"learning_rate": 4.912976038673786e-05,
"loss": 1.866,
"step": 3968
},
{
"epoch": 0.10018538931756689,
"grad_norm": 1.180206537246704,
"learning_rate": 4.9107658826866254e-05,
"loss": 1.849,
"step": 3999
},
{
"epoch": 0.10096202024250928,
"grad_norm": 1.0747199058532715,
"learning_rate": 4.908528521534139e-05,
"loss": 1.8547,
"step": 4030
},
{
"epoch": 0.10173865116745165,
"grad_norm": 1.103786826133728,
"learning_rate": 4.906263980464644e-05,
"loss": 1.864,
"step": 4061
},
{
"epoch": 0.10251528209239402,
"grad_norm": 1.1390912532806396,
"learning_rate": 4.903972285033178e-05,
"loss": 1.8669,
"step": 4092
},
{
"epoch": 0.10329191301733641,
"grad_norm": 1.1488468647003174,
"learning_rate": 4.901653461101213e-05,
"loss": 1.8445,
"step": 4123
},
{
"epoch": 0.10406854394227878,
"grad_norm": 1.1589927673339844,
"learning_rate": 4.8993075348363626e-05,
"loss": 1.8529,
"step": 4154
},
{
"epoch": 0.10484517486722117,
"grad_norm": 1.1793478727340698,
"learning_rate": 4.896934532712084e-05,
"loss": 1.8818,
"step": 4185
},
{
"epoch": 0.10562180579216354,
"grad_norm": 1.165292739868164,
"learning_rate": 4.8945344815073846e-05,
"loss": 1.8425,
"step": 4216
},
{
"epoch": 0.10639843671710593,
"grad_norm": 1.105946660041809,
"learning_rate": 4.892107408306516e-05,
"loss": 1.8327,
"step": 4247
},
{
"epoch": 0.1071750676420483,
"grad_norm": 1.1690527200698853,
"learning_rate": 4.889653340498669e-05,
"loss": 1.8428,
"step": 4278
},
{
"epoch": 0.10795169856699068,
"grad_norm": 1.1106913089752197,
"learning_rate": 4.8871723057776664e-05,
"loss": 1.839,
"step": 4309
},
{
"epoch": 0.10872832949193306,
"grad_norm": 1.1590090990066528,
"learning_rate": 4.8846643321416476e-05,
"loss": 1.8658,
"step": 4340
},
{
"epoch": 0.10950496041687544,
"grad_norm": 1.1315875053405762,
"learning_rate": 4.882129447892753e-05,
"loss": 1.8282,
"step": 4371
},
{
"epoch": 0.11028159134181782,
"grad_norm": 1.6055009365081787,
"learning_rate": 4.8795676816368076e-05,
"loss": 1.8294,
"step": 4402
},
{
"epoch": 0.1110582222667602,
"grad_norm": 1.1236201524734497,
"learning_rate": 4.876979062282995e-05,
"loss": 1.8325,
"step": 4433
},
{
"epoch": 0.11183485319170258,
"grad_norm": 1.0830161571502686,
"learning_rate": 4.8743636190435325e-05,
"loss": 1.8451,
"step": 4464
},
{
"epoch": 0.11261148411664496,
"grad_norm": 1.1663005352020264,
"learning_rate": 4.871721381433344e-05,
"loss": 1.8349,
"step": 4495
},
{
"epoch": 0.11338811504158733,
"grad_norm": 1.2522406578063965,
"learning_rate": 4.869052379269719e-05,
"loss": 1.8162,
"step": 4526
},
{
"epoch": 0.11416474596652972,
"grad_norm": 1.1271142959594727,
"learning_rate": 4.866356642671985e-05,
"loss": 1.8348,
"step": 4557
},
{
"epoch": 0.11494137689147209,
"grad_norm": 1.053753137588501,
"learning_rate": 4.8636342020611634e-05,
"loss": 1.8376,
"step": 4588
},
{
"epoch": 0.11571800781641448,
"grad_norm": 1.1469056606292725,
"learning_rate": 4.860885088159626e-05,
"loss": 1.8248,
"step": 4619
},
{
"epoch": 0.11649463874135685,
"grad_norm": 1.116128921508789,
"learning_rate": 4.858109331990751e-05,
"loss": 1.8454,
"step": 4650
},
{
"epoch": 0.11727126966629922,
"grad_norm": 1.0742651224136353,
"learning_rate": 4.855306964878567e-05,
"loss": 1.8231,
"step": 4681
},
{
"epoch": 0.11804790059124161,
"grad_norm": 1.101657748222351,
"learning_rate": 4.8524780184474084e-05,
"loss": 1.8421,
"step": 4712
},
{
"epoch": 0.11882453151618398,
"grad_norm": 1.0891473293304443,
"learning_rate": 4.8496225246215496e-05,
"loss": 1.8053,
"step": 4743
},
{
"epoch": 0.11960116244112637,
"grad_norm": 1.1044034957885742,
"learning_rate": 4.8467405156248505e-05,
"loss": 1.8184,
"step": 4774
},
{
"epoch": 0.12037779336606874,
"grad_norm": 1.0644783973693848,
"learning_rate": 4.843832023980392e-05,
"loss": 1.8237,
"step": 4805
},
{
"epoch": 0.12115442429101113,
"grad_norm": 1.0427252054214478,
"learning_rate": 4.840897082510106e-05,
"loss": 1.8189,
"step": 4836
},
{
"epoch": 0.1219310552159535,
"grad_norm": 1.0806516408920288,
"learning_rate": 4.8379357243344084e-05,
"loss": 1.7958,
"step": 4867
},
{
"epoch": 0.12270768614089587,
"grad_norm": 1.092581868171692,
"learning_rate": 4.8349479828718236e-05,
"loss": 1.8229,
"step": 4898
},
{
"epoch": 0.12348431706583826,
"grad_norm": 1.0858657360076904,
"learning_rate": 4.8319338918386075e-05,
"loss": 1.8165,
"step": 4929
},
{
"epoch": 0.12426094799078063,
"grad_norm": 1.073818325996399,
"learning_rate": 4.828893485248369e-05,
"loss": 1.8352,
"step": 4960
},
{
"epoch": 0.125037578915723,
"grad_norm": 1.069009780883789,
"learning_rate": 4.825826797411682e-05,
"loss": 1.8,
"step": 4991
},
{
"epoch": 0.1258142098406654,
"grad_norm": 1.0416266918182373,
"learning_rate": 4.822733862935702e-05,
"loss": 1.8224,
"step": 5022
},
{
"epoch": 0.12659084076560778,
"grad_norm": 1.093551516532898,
"learning_rate": 4.819614716723775e-05,
"loss": 1.8005,
"step": 5053
},
{
"epoch": 0.12736747169055015,
"grad_norm": 1.026353359222412,
"learning_rate": 4.8164693939750425e-05,
"loss": 1.844,
"step": 5084
},
{
"epoch": 0.12814410261549253,
"grad_norm": 1.102858304977417,
"learning_rate": 4.813297930184042e-05,
"loss": 1.8045,
"step": 5115
},
{
"epoch": 0.1289207335404349,
"grad_norm": 1.0537272691726685,
"learning_rate": 4.810100361140314e-05,
"loss": 1.8054,
"step": 5146
},
{
"epoch": 0.1296973644653773,
"grad_norm": 1.0389387607574463,
"learning_rate": 4.8068767229279885e-05,
"loss": 1.7734,
"step": 5177
},
{
"epoch": 0.13047399539031967,
"grad_norm": 1.0274319648742676,
"learning_rate": 4.8036270519253854e-05,
"loss": 1.788,
"step": 5208
},
{
"epoch": 0.13125062631526205,
"grad_norm": 1.0932128429412842,
"learning_rate": 4.8003513848046e-05,
"loss": 1.7813,
"step": 5239
},
{
"epoch": 0.13202725724020442,
"grad_norm": 1.0294640064239502,
"learning_rate": 4.79704975853109e-05,
"loss": 1.81,
"step": 5270
},
{
"epoch": 0.13280388816514682,
"grad_norm": 1.0617210865020752,
"learning_rate": 4.793722210363262e-05,
"loss": 1.8017,
"step": 5301
},
{
"epoch": 0.1335805190900892,
"grad_norm": 1.0657798051834106,
"learning_rate": 4.7903687778520414e-05,
"loss": 1.7884,
"step": 5332
},
{
"epoch": 0.13435715001503157,
"grad_norm": 0.9954230189323425,
"learning_rate": 4.7869894988404593e-05,
"loss": 1.7848,
"step": 5363
},
{
"epoch": 0.13513378093997394,
"grad_norm": 1.117066740989685,
"learning_rate": 4.783584411463221e-05,
"loss": 1.7984,
"step": 5394
},
{
"epoch": 0.13591041186491631,
"grad_norm": 1.0706870555877686,
"learning_rate": 4.780153554146274e-05,
"loss": 1.8008,
"step": 5425
},
{
"epoch": 0.13668704278985871,
"grad_norm": 1.063238501548767,
"learning_rate": 4.7766969656063766e-05,
"loss": 1.7754,
"step": 5456
},
{
"epoch": 0.1374636737148011,
"grad_norm": 1.0702152252197266,
"learning_rate": 4.773214684850662e-05,
"loss": 1.8037,
"step": 5487
},
{
"epoch": 0.13824030463974346,
"grad_norm": 1.0318788290023804,
"learning_rate": 4.769706751176193e-05,
"loss": 1.8174,
"step": 5518
},
{
"epoch": 0.13901693556468583,
"grad_norm": 1.0097755193710327,
"learning_rate": 4.7661732041695264e-05,
"loss": 1.7648,
"step": 5549
},
{
"epoch": 0.1397935664896282,
"grad_norm": 1.063873291015625,
"learning_rate": 4.762614083706258e-05,
"loss": 1.809,
"step": 5580
},
{
"epoch": 0.1405701974145706,
"grad_norm": 1.0409623384475708,
"learning_rate": 4.759029429950581e-05,
"loss": 1.7853,
"step": 5611
},
{
"epoch": 0.14134682833951298,
"grad_norm": 1.0131454467773438,
"learning_rate": 4.7554192833548235e-05,
"loss": 1.7759,
"step": 5642
},
{
"epoch": 0.14212345926445535,
"grad_norm": 1.1056182384490967,
"learning_rate": 4.751783684659e-05,
"loss": 1.7959,
"step": 5673
},
{
"epoch": 0.14290009018939773,
"grad_norm": 1.0572400093078613,
"learning_rate": 4.748122674890348e-05,
"loss": 1.7597,
"step": 5704
},
{
"epoch": 0.14367672111434013,
"grad_norm": 1.0375362634658813,
"learning_rate": 4.7444362953628654e-05,
"loss": 1.7719,
"step": 5735
},
{
"epoch": 0.1444533520392825,
"grad_norm": 1.0238806009292603,
"learning_rate": 4.7407245876768424e-05,
"loss": 1.7966,
"step": 5766
},
{
"epoch": 0.14522998296422487,
"grad_norm": 1.0335805416107178,
"learning_rate": 4.736987593718397e-05,
"loss": 1.7805,
"step": 5797
},
{
"epoch": 0.14600661388916725,
"grad_norm": 1.08907151222229,
"learning_rate": 4.733225355658999e-05,
"loss": 1.7905,
"step": 5828
},
{
"epoch": 0.14678324481410962,
"grad_norm": 1.0151680707931519,
"learning_rate": 4.7294379159549926e-05,
"loss": 1.7766,
"step": 5859
},
{
"epoch": 0.14755987573905202,
"grad_norm": 0.9963664412498474,
"learning_rate": 4.725625317347119e-05,
"loss": 1.7708,
"step": 5890
},
{
"epoch": 0.1483365066639944,
"grad_norm": 1.0762684345245361,
"learning_rate": 4.7217876028600374e-05,
"loss": 1.7998,
"step": 5921
},
{
"epoch": 0.14911313758893677,
"grad_norm": 1.1151126623153687,
"learning_rate": 4.717924815801832e-05,
"loss": 1.782,
"step": 5952
},
{
"epoch": 0.14988976851387914,
"grad_norm": 0.9970519542694092,
"learning_rate": 4.714036999763532e-05,
"loss": 1.7779,
"step": 5983
},
{
"epoch": 0.1506663994388215,
"grad_norm": 1.0146524906158447,
"learning_rate": 4.7101241986186116e-05,
"loss": 1.7631,
"step": 6014
},
{
"epoch": 0.15144303036376391,
"grad_norm": 1.0554558038711548,
"learning_rate": 4.7061864565225e-05,
"loss": 1.7729,
"step": 6045
},
{
"epoch": 0.1522196612887063,
"grad_norm": 1.0413624048233032,
"learning_rate": 4.702223817912081e-05,
"loss": 1.79,
"step": 6076
},
{
"epoch": 0.15299629221364866,
"grad_norm": 1.07778000831604,
"learning_rate": 4.698236327505195e-05,
"loss": 1.787,
"step": 6107
},
{
"epoch": 0.15377292313859103,
"grad_norm": 1.0084989070892334,
"learning_rate": 4.694224030300127e-05,
"loss": 1.7695,
"step": 6138
},
{
"epoch": 0.1545495540635334,
"grad_norm": 1.0067577362060547,
"learning_rate": 4.690186971575107e-05,
"loss": 1.7611,
"step": 6169
},
{
"epoch": 0.1553261849884758,
"grad_norm": 1.0149894952774048,
"learning_rate": 4.6861251968877916e-05,
"loss": 1.7734,
"step": 6200
},
{
"epoch": 0.15610281591341818,
"grad_norm": 0.996447741985321,
"learning_rate": 4.68203875207476e-05,
"loss": 1.7609,
"step": 6231
},
{
"epoch": 0.15687944683836055,
"grad_norm": 1.0334714651107788,
"learning_rate": 4.677927683250983e-05,
"loss": 1.7725,
"step": 6262
},
{
"epoch": 0.15765607776330293,
"grad_norm": 0.949685275554657,
"learning_rate": 4.6737920368093156e-05,
"loss": 1.7597,
"step": 6293
},
{
"epoch": 0.15843270868824533,
"grad_norm": 1.0946801900863647,
"learning_rate": 4.669631859419965e-05,
"loss": 1.755,
"step": 6324
},
{
"epoch": 0.1592093396131877,
"grad_norm": 1.0390102863311768,
"learning_rate": 4.6654471980299676e-05,
"loss": 1.7755,
"step": 6355
},
{
"epoch": 0.15998597053813007,
"grad_norm": 0.974733293056488,
"learning_rate": 4.661238099862658e-05,
"loss": 1.7619,
"step": 6386
},
{
"epoch": 0.16076260146307245,
"grad_norm": 1.0192877054214478,
"learning_rate": 4.657004612417138e-05,
"loss": 1.7676,
"step": 6417
},
{
"epoch": 0.16153923238801482,
"grad_norm": 1.0147426128387451,
"learning_rate": 4.6527467834677374e-05,
"loss": 1.7473,
"step": 6448
},
{
"epoch": 0.16231586331295722,
"grad_norm": 1.0231157541275024,
"learning_rate": 4.648464661063478e-05,
"loss": 1.766,
"step": 6479
},
{
"epoch": 0.1630924942378996,
"grad_norm": 0.9995896816253662,
"learning_rate": 4.6441582935275264e-05,
"loss": 1.7594,
"step": 6510
},
{
"epoch": 0.16386912516284197,
"grad_norm": 0.9674972295761108,
"learning_rate": 4.6398277294566586e-05,
"loss": 1.7347,
"step": 6541
},
{
"epoch": 0.16464575608778434,
"grad_norm": 0.9876355528831482,
"learning_rate": 4.6354730177207e-05,
"loss": 1.748,
"step": 6572
},
{
"epoch": 0.1654223870127267,
"grad_norm": 1.0107293128967285,
"learning_rate": 4.6310942074619787e-05,
"loss": 1.7743,
"step": 6603
},
{
"epoch": 0.1661990179376691,
"grad_norm": 0.9731600284576416,
"learning_rate": 4.626691348094777e-05,
"loss": 1.7515,
"step": 6634
},
{
"epoch": 0.1669756488626115,
"grad_norm": 0.9843081831932068,
"learning_rate": 4.622264489304762e-05,
"loss": 1.7498,
"step": 6665
},
{
"epoch": 0.16775227978755386,
"grad_norm": 0.9850636720657349,
"learning_rate": 4.617813681048434e-05,
"loss": 1.7563,
"step": 6696
},
{
"epoch": 0.16852891071249623,
"grad_norm": 0.964599609375,
"learning_rate": 4.61333897355256e-05,
"loss": 1.7699,
"step": 6727
},
{
"epoch": 0.16930554163743863,
"grad_norm": 0.9415149688720703,
"learning_rate": 4.608840417313604e-05,
"loss": 1.745,
"step": 6758
},
{
"epoch": 0.170082172562381,
"grad_norm": 0.9608363509178162,
"learning_rate": 4.6043180630971646e-05,
"loss": 1.7519,
"step": 6789
},
{
"epoch": 0.17085880348732338,
"grad_norm": 0.9133123159408569,
"learning_rate": 4.599771961937391e-05,
"loss": 1.7432,
"step": 6820
},
{
"epoch": 0.17163543441226575,
"grad_norm": 0.943702220916748,
"learning_rate": 4.5952021651364204e-05,
"loss": 1.7231,
"step": 6851
},
{
"epoch": 0.17241206533720813,
"grad_norm": 0.9676947593688965,
"learning_rate": 4.590608724263786e-05,
"loss": 1.7563,
"step": 6882
},
{
"epoch": 0.17318869626215053,
"grad_norm": 0.964959442615509,
"learning_rate": 4.585991691155845e-05,
"loss": 1.7514,
"step": 6913
},
{
"epoch": 0.1739653271870929,
"grad_norm": 1.0078297853469849,
"learning_rate": 4.581351117915188e-05,
"loss": 1.7538,
"step": 6944
},
{
"epoch": 0.17474195811203527,
"grad_norm": 0.9631285071372986,
"learning_rate": 4.5766870569100534e-05,
"loss": 1.7363,
"step": 6975
},
{
"epoch": 0.17551858903697765,
"grad_norm": 0.9542151093482971,
"learning_rate": 4.571999560773736e-05,
"loss": 1.753,
"step": 7006
},
{
"epoch": 0.17629521996192002,
"grad_norm": 0.9770635366439819,
"learning_rate": 4.5672886824039915e-05,
"loss": 1.7435,
"step": 7037
},
{
"epoch": 0.17707185088686242,
"grad_norm": 0.979009211063385,
"learning_rate": 4.5625544749624435e-05,
"loss": 1.7591,
"step": 7068
},
{
"epoch": 0.1778484818118048,
"grad_norm": 0.9663967490196228,
"learning_rate": 4.5577969918739794e-05,
"loss": 1.7129,
"step": 7099
},
{
"epoch": 0.17862511273674717,
"grad_norm": 1.0181857347488403,
"learning_rate": 4.5530162868261486e-05,
"loss": 1.7396,
"step": 7130
},
{
"epoch": 0.17940174366168954,
"grad_norm": 1.0090553760528564,
"learning_rate": 4.548212413768558e-05,
"loss": 1.7488,
"step": 7161
},
{
"epoch": 0.18017837458663194,
"grad_norm": 1.0095964670181274,
"learning_rate": 4.543385426912261e-05,
"loss": 1.7342,
"step": 7192
},
{
"epoch": 0.1809550055115743,
"grad_norm": 0.994108259677887,
"learning_rate": 4.53853538072915e-05,
"loss": 1.7445,
"step": 7223
},
{
"epoch": 0.18173163643651669,
"grad_norm": 0.9566254019737244,
"learning_rate": 4.533662329951336e-05,
"loss": 1.7272,
"step": 7254
},
{
"epoch": 0.18250826736145906,
"grad_norm": 0.9655300974845886,
"learning_rate": 4.528766329570536e-05,
"loss": 1.7112,
"step": 7285
},
{
"epoch": 0.18328489828640143,
"grad_norm": 0.9693655371665955,
"learning_rate": 4.523847434837447e-05,
"loss": 1.74,
"step": 7316
},
{
"epoch": 0.18406152921134383,
"grad_norm": 0.9668208956718445,
"learning_rate": 4.518905701261128e-05,
"loss": 1.7239,
"step": 7347
},
{
"epoch": 0.1848381601362862,
"grad_norm": 1.0193735361099243,
"learning_rate": 4.5139411846083715e-05,
"loss": 1.7273,
"step": 7378
},
{
"epoch": 0.18561479106122858,
"grad_norm": 0.9525153636932373,
"learning_rate": 4.508953940903073e-05,
"loss": 1.7356,
"step": 7409
},
{
"epoch": 0.18639142198617095,
"grad_norm": 0.9590099453926086,
"learning_rate": 4.5039440264255994e-05,
"loss": 1.7296,
"step": 7440
},
{
"epoch": 0.18716805291111333,
"grad_norm": 0.9519559741020203,
"learning_rate": 4.498911497712155e-05,
"loss": 1.7183,
"step": 7471
},
{
"epoch": 0.18794468383605573,
"grad_norm": 0.9692674279212952,
"learning_rate": 4.493856411554142e-05,
"loss": 1.7329,
"step": 7502
},
{
"epoch": 0.1887213147609981,
"grad_norm": 0.9676815271377563,
"learning_rate": 4.4887788249975206e-05,
"loss": 1.7194,
"step": 7533
},
{
"epoch": 0.18949794568594047,
"grad_norm": 0.9612167477607727,
"learning_rate": 4.4836787953421656e-05,
"loss": 1.7297,
"step": 7564
},
{
"epoch": 0.19027457661088285,
"grad_norm": 0.9661709666252136,
"learning_rate": 4.478556380141218e-05,
"loss": 1.728,
"step": 7595
},
{
"epoch": 0.19105120753582525,
"grad_norm": 0.95663982629776,
"learning_rate": 4.4734116372004375e-05,
"loss": 1.7387,
"step": 7626
},
{
"epoch": 0.19182783846076762,
"grad_norm": 0.9469471573829651,
"learning_rate": 4.4682446245775477e-05,
"loss": 1.7034,
"step": 7657
},
{
"epoch": 0.19260446938571,
"grad_norm": 1.03744637966156,
"learning_rate": 4.463055400581586e-05,
"loss": 1.7394,
"step": 7688
},
{
"epoch": 0.19338110031065237,
"grad_norm": 0.932919979095459,
"learning_rate": 4.4578440237722374e-05,
"loss": 1.7113,
"step": 7719
},
{
"epoch": 0.19415773123559474,
"grad_norm": 0.9527629017829895,
"learning_rate": 4.452610552959183e-05,
"loss": 1.7228,
"step": 7750
},
{
"epoch": 0.19493436216053714,
"grad_norm": 0.9649306535720825,
"learning_rate": 4.447355047201428e-05,
"loss": 1.7434,
"step": 7781
},
{
"epoch": 0.1957109930854795,
"grad_norm": 0.9027903079986572,
"learning_rate": 4.4420775658066414e-05,
"loss": 1.7022,
"step": 7812
},
{
"epoch": 0.19648762401042189,
"grad_norm": 0.9486429691314697,
"learning_rate": 4.436778168330484e-05,
"loss": 1.7277,
"step": 7843
},
{
"epoch": 0.19726425493536426,
"grad_norm": 0.9951580762863159,
"learning_rate": 4.4314569145759353e-05,
"loss": 1.7354,
"step": 7874
},
{
"epoch": 0.19804088586030663,
"grad_norm": 0.984990656375885,
"learning_rate": 4.42611386459262e-05,
"loss": 1.7247,
"step": 7905
},
{
"epoch": 0.19881751678524903,
"grad_norm": 1.0146019458770752,
"learning_rate": 4.420749078676133e-05,
"loss": 1.7048,
"step": 7936
},
{
"epoch": 0.1995941477101914,
"grad_norm": 0.9386274814605713,
"learning_rate": 4.4153626173673516e-05,
"loss": 1.7035,
"step": 7967
},
{
"epoch": 0.20037077863513378,
"grad_norm": 0.9656661152839661,
"learning_rate": 4.409954541451762e-05,
"loss": 1.7138,
"step": 7998
},
{
"epoch": 0.20114740956007615,
"grad_norm": 0.9721409678459167,
"learning_rate": 4.404524911958764e-05,
"loss": 1.7174,
"step": 8029
},
{
"epoch": 0.20192404048501855,
"grad_norm": 0.9685981869697571,
"learning_rate": 4.399073790160989e-05,
"loss": 1.7162,
"step": 8060
},
{
"epoch": 0.20270067140996093,
"grad_norm": 0.9472204446792603,
"learning_rate": 4.393601237573607e-05,
"loss": 1.7068,
"step": 8091
},
{
"epoch": 0.2034773023349033,
"grad_norm": 0.9472411870956421,
"learning_rate": 4.388107315953628e-05,
"loss": 1.7244,
"step": 8122
},
{
"epoch": 0.20425393325984567,
"grad_norm": 0.9380378723144531,
"learning_rate": 4.382592087299212e-05,
"loss": 1.7169,
"step": 8153
},
{
"epoch": 0.20503056418478804,
"grad_norm": 0.9613087177276611,
"learning_rate": 4.377055613848964e-05,
"loss": 1.7108,
"step": 8184
},
{
"epoch": 0.20580719510973045,
"grad_norm": 0.9507735967636108,
"learning_rate": 4.3714979580812355e-05,
"loss": 1.7147,
"step": 8215
},
{
"epoch": 0.20658382603467282,
"grad_norm": 0.9669816493988037,
"learning_rate": 4.365919182713416e-05,
"loss": 1.7058,
"step": 8246
},
{
"epoch": 0.2073604569596152,
"grad_norm": 0.8880124688148499,
"learning_rate": 4.360319350701226e-05,
"loss": 1.7049,
"step": 8277
},
{
"epoch": 0.20813708788455756,
"grad_norm": 4.802671909332275,
"learning_rate": 4.3546985252380115e-05,
"loss": 1.7349,
"step": 8308
},
{
"epoch": 0.20891371880949994,
"grad_norm": 0.9394221901893616,
"learning_rate": 4.349056769754021e-05,
"loss": 1.6932,
"step": 8339
},
{
"epoch": 0.20969034973444234,
"grad_norm": 0.9521833062171936,
"learning_rate": 4.3433941479156994e-05,
"loss": 1.695,
"step": 8370
},
{
"epoch": 0.2104669806593847,
"grad_norm": 0.9627339839935303,
"learning_rate": 4.3377107236249647e-05,
"loss": 1.7329,
"step": 8401
},
{
"epoch": 0.21124361158432708,
"grad_norm": 0.9096025228500366,
"learning_rate": 4.332006561018488e-05,
"loss": 1.6885,
"step": 8432
},
{
"epoch": 0.21202024250926946,
"grad_norm": 0.9301928877830505,
"learning_rate": 4.3262817244669683e-05,
"loss": 1.7112,
"step": 8463
},
{
"epoch": 0.21279687343421186,
"grad_norm": 1.0362213850021362,
"learning_rate": 4.3205362785744083e-05,
"loss": 1.7061,
"step": 8494
},
{
"epoch": 0.21357350435915423,
"grad_norm": 1.986802339553833,
"learning_rate": 4.314770288177384e-05,
"loss": 1.7095,
"step": 8525
},
{
"epoch": 0.2143501352840966,
"grad_norm": 0.9070325493812561,
"learning_rate": 4.308983818344313e-05,
"loss": 1.7087,
"step": 8556
},
{
"epoch": 0.21512676620903898,
"grad_norm": 0.985433042049408,
"learning_rate": 4.3031769343747206e-05,
"loss": 1.7176,
"step": 8587
},
{
"epoch": 0.21590339713398135,
"grad_norm": 0.9339021444320679,
"learning_rate": 4.297349701798505e-05,
"loss": 1.6823,
"step": 8618
},
{
"epoch": 0.21668002805892375,
"grad_norm": 0.9413474798202515,
"learning_rate": 4.2915021863751916e-05,
"loss": 1.7205,
"step": 8649
},
{
"epoch": 0.21745665898386612,
"grad_norm": 1.023187279701233,
"learning_rate": 4.285634454093198e-05,
"loss": 1.6943,
"step": 8680
},
{
"epoch": 0.2182332899088085,
"grad_norm": 0.9431496262550354,
"learning_rate": 4.279746571169086e-05,
"loss": 1.7012,
"step": 8711
},
{
"epoch": 0.21900992083375087,
"grad_norm": 0.9386807680130005,
"learning_rate": 4.2738386040468136e-05,
"loss": 1.7069,
"step": 8742
},
{
"epoch": 0.21978655175869324,
"grad_norm": 0.9307882785797119,
"learning_rate": 4.2679106193969866e-05,
"loss": 1.6914,
"step": 8773
},
{
"epoch": 0.22056318268363564,
"grad_norm": 1.0440980195999146,
"learning_rate": 4.261962684116106e-05,
"loss": 1.6892,
"step": 8804
},
{
"epoch": 0.22133981360857802,
"grad_norm": 0.9271411895751953,
"learning_rate": 4.2559948653258145e-05,
"loss": 1.7117,
"step": 8835
},
{
"epoch": 0.2221164445335204,
"grad_norm": 0.9182149767875671,
"learning_rate": 4.250007230372134e-05,
"loss": 1.7092,
"step": 8866
},
{
"epoch": 0.22289307545846276,
"grad_norm": 0.9707378149032593,
"learning_rate": 4.2439998468247126e-05,
"loss": 1.7055,
"step": 8897
},
{
"epoch": 0.22366970638340516,
"grad_norm": 0.9413692951202393,
"learning_rate": 4.2379727824760566e-05,
"loss": 1.6968,
"step": 8928
},
{
"epoch": 0.22444633730834754,
"grad_norm": 1.1943906545639038,
"learning_rate": 4.231926105340768e-05,
"loss": 1.7174,
"step": 8959
},
{
"epoch": 0.2252229682332899,
"grad_norm": 0.9508203268051147,
"learning_rate": 4.225859883654776e-05,
"loss": 1.6941,
"step": 8990
},
{
"epoch": 0.22599959915823228,
"grad_norm": 0.9774237275123596,
"learning_rate": 4.219774185874569e-05,
"loss": 1.6996,
"step": 9021
},
{
"epoch": 0.22677623008317466,
"grad_norm": 0.9906953573226929,
"learning_rate": 4.213669080676418e-05,
"loss": 1.7186,
"step": 9052
},
{
"epoch": 0.22755286100811706,
"grad_norm": 0.966076135635376,
"learning_rate": 4.2075446369556056e-05,
"loss": 1.6828,
"step": 9083
},
{
"epoch": 0.22832949193305943,
"grad_norm": 0.9266401529312134,
"learning_rate": 4.201400923825648e-05,
"loss": 1.6969,
"step": 9114
},
{
"epoch": 0.2291061228580018,
"grad_norm": 0.8797683119773865,
"learning_rate": 4.195238010617511e-05,
"loss": 1.6835,
"step": 9145
},
{
"epoch": 0.22988275378294418,
"grad_norm": 0.92194664478302,
"learning_rate": 4.1890559668788344e-05,
"loss": 1.7028,
"step": 9176
},
{
"epoch": 0.23065938470788655,
"grad_norm": 0.9278755187988281,
"learning_rate": 4.1828548623731405e-05,
"loss": 1.6853,
"step": 9207
},
{
"epoch": 0.23143601563282895,
"grad_norm": 0.9345482587814331,
"learning_rate": 4.1766347670790506e-05,
"loss": 1.6879,
"step": 9238
},
{
"epoch": 0.23221264655777132,
"grad_norm": 0.9217596054077148,
"learning_rate": 4.170395751189495e-05,
"loss": 1.679,
"step": 9269
},
{
"epoch": 0.2329892774827137,
"grad_norm": 0.9330849647521973,
"learning_rate": 4.164137885110921e-05,
"loss": 1.6891,
"step": 9300
},
{
"epoch": 0.23376590840765607,
"grad_norm": 0.923973023891449,
"learning_rate": 4.157861239462495e-05,
"loss": 1.6929,
"step": 9331
},
{
"epoch": 0.23454253933259844,
"grad_norm": 0.9245807528495789,
"learning_rate": 4.1515658850753114e-05,
"loss": 1.7162,
"step": 9362
},
{
"epoch": 0.23531917025754084,
"grad_norm": 0.9212767481803894,
"learning_rate": 4.145251892991588e-05,
"loss": 1.7169,
"step": 9393
},
{
"epoch": 0.23609580118248322,
"grad_norm": 0.9439975023269653,
"learning_rate": 4.138919334463868e-05,
"loss": 1.6869,
"step": 9424
},
{
"epoch": 0.2368724321074256,
"grad_norm": 0.9876207709312439,
"learning_rate": 4.1325682809542124e-05,
"loss": 1.6989,
"step": 9455
},
{
"epoch": 0.23764906303236796,
"grad_norm": 0.8833792805671692,
"learning_rate": 4.126198804133398e-05,
"loss": 1.6917,
"step": 9486
},
{
"epoch": 0.23842569395731036,
"grad_norm": 0.9133450388908386,
"learning_rate": 4.1198109758801055e-05,
"loss": 1.7038,
"step": 9517
},
{
"epoch": 0.23920232488225274,
"grad_norm": 0.8729329705238342,
"learning_rate": 4.113404868280107e-05,
"loss": 1.672,
"step": 9548
},
{
"epoch": 0.2399789558071951,
"grad_norm": 0.892375111579895,
"learning_rate": 4.106980553625457e-05,
"loss": 1.6829,
"step": 9579
},
{
"epoch": 0.24075558673213748,
"grad_norm": 0.8939517140388489,
"learning_rate": 4.100538104413674e-05,
"loss": 1.6965,
"step": 9610
},
{
"epoch": 0.24153221765707986,
"grad_norm": 0.8656035661697388,
"learning_rate": 4.09407759334692e-05,
"loss": 1.7063,
"step": 9641
},
{
"epoch": 0.24230884858202226,
"grad_norm": 0.8904181122779846,
"learning_rate": 4.087599093331186e-05,
"loss": 1.6719,
"step": 9672
},
{
"epoch": 0.24308547950696463,
"grad_norm": 0.9176281690597534,
"learning_rate": 4.081102677475462e-05,
"loss": 1.6648,
"step": 9703
},
{
"epoch": 0.243862110431907,
"grad_norm": 0.9038643836975098,
"learning_rate": 4.0745884190909194e-05,
"loss": 1.6756,
"step": 9734
},
{
"epoch": 0.24463874135684938,
"grad_norm": 0.9255871772766113,
"learning_rate": 4.0680563916900796e-05,
"loss": 1.6752,
"step": 9765
},
{
"epoch": 0.24541537228179175,
"grad_norm": 0.9124768972396851,
"learning_rate": 4.0615066689859815e-05,
"loss": 1.6982,
"step": 9796
},
{
"epoch": 0.24619200320673415,
"grad_norm": 0.9149096608161926,
"learning_rate": 4.0549393248913584e-05,
"loss": 1.6692,
"step": 9827
},
{
"epoch": 0.24696863413167652,
"grad_norm": 0.9468510150909424,
"learning_rate": 4.048354433517794e-05,
"loss": 1.6775,
"step": 9858
},
{
"epoch": 0.2477452650566189,
"grad_norm": 1.2210197448730469,
"learning_rate": 4.0417520691748916e-05,
"loss": 1.692,
"step": 9889
},
{
"epoch": 0.24852189598156127,
"grad_norm": 0.9717695116996765,
"learning_rate": 4.035132306369438e-05,
"loss": 1.6583,
"step": 9920
},
{
"epoch": 0.24929852690650367,
"grad_norm": 0.9524528980255127,
"learning_rate": 4.028495219804555e-05,
"loss": 1.6606,
"step": 9951
},
{
"epoch": 0.250075157831446,
"grad_norm": 0.937414824962616,
"learning_rate": 4.021840884378864e-05,
"loss": 1.7063,
"step": 9982
},
{
"epoch": 0.2508517887563884,
"grad_norm": 0.9392083883285522,
"learning_rate": 4.015169375185633e-05,
"loss": 1.6813,
"step": 10013
},
{
"epoch": 0.2516284196813308,
"grad_norm": 0.9055363535881042,
"learning_rate": 4.0084807675119396e-05,
"loss": 1.6622,
"step": 10044
},
{
"epoch": 0.25240505060627316,
"grad_norm": 0.8967134356498718,
"learning_rate": 4.0017751368378106e-05,
"loss": 1.7052,
"step": 10075
},
{
"epoch": 0.25318168153121556,
"grad_norm": 0.9304991364479065,
"learning_rate": 3.995052558835377e-05,
"loss": 1.6779,
"step": 10106
},
{
"epoch": 0.2539583124561579,
"grad_norm": 0.9946010708808899,
"learning_rate": 3.988313109368017e-05,
"loss": 1.6727,
"step": 10137
},
{
"epoch": 0.2547349433811003,
"grad_norm": 0.936072051525116,
"learning_rate": 3.981556864489504e-05,
"loss": 1.6861,
"step": 10168
},
{
"epoch": 0.2555115743060427,
"grad_norm": 0.9157513380050659,
"learning_rate": 3.974783900443142e-05,
"loss": 1.6871,
"step": 10199
},
{
"epoch": 0.25628820523098506,
"grad_norm": 0.8980918526649475,
"learning_rate": 3.9679942936609095e-05,
"loss": 1.674,
"step": 10230
},
{
"epoch": 0.25706483615592746,
"grad_norm": 0.9256001114845276,
"learning_rate": 3.961188120762596e-05,
"loss": 1.6858,
"step": 10261
},
{
"epoch": 0.2578414670808698,
"grad_norm": 0.8827298879623413,
"learning_rate": 3.954365458554938e-05,
"loss": 1.6734,
"step": 10292
},
{
"epoch": 0.2586180980058122,
"grad_norm": 0.9297879338264465,
"learning_rate": 3.947526384030751e-05,
"loss": 1.6648,
"step": 10323
},
{
"epoch": 0.2593947289307546,
"grad_norm": 0.91744065284729,
"learning_rate": 3.9406709743680624e-05,
"loss": 1.6713,
"step": 10354
},
{
"epoch": 0.26017135985569695,
"grad_norm": 0.949213445186615,
"learning_rate": 3.9337993069292366e-05,
"loss": 1.6633,
"step": 10385
},
{
"epoch": 0.26094799078063935,
"grad_norm": 0.9231769442558289,
"learning_rate": 3.926911459260109e-05,
"loss": 1.6815,
"step": 10416
},
{
"epoch": 0.2617246217055817,
"grad_norm": 0.9055944681167603,
"learning_rate": 3.920007509089102e-05,
"loss": 1.6621,
"step": 10447
},
{
"epoch": 0.2625012526305241,
"grad_norm": 1.7065776586532593,
"learning_rate": 3.913087534326357e-05,
"loss": 1.6832,
"step": 10478
},
{
"epoch": 0.2632778835554665,
"grad_norm": 0.9098711013793945,
"learning_rate": 3.9061516130628475e-05,
"loss": 1.7106,
"step": 10509
},
{
"epoch": 0.26405451448040884,
"grad_norm": 0.9403277635574341,
"learning_rate": 3.8991998235695025e-05,
"loss": 1.708,
"step": 10540
},
{
"epoch": 0.26483114540535124,
"grad_norm": 0.9076827168464661,
"learning_rate": 3.8922322442963224e-05,
"loss": 1.6571,
"step": 10571
},
{
"epoch": 0.26560777633029364,
"grad_norm": 0.9140092730522156,
"learning_rate": 3.885248953871491e-05,
"loss": 1.6626,
"step": 10602
},
{
"epoch": 0.266384407255236,
"grad_norm": 0.9277907013893127,
"learning_rate": 3.8782500311004915e-05,
"loss": 1.6625,
"step": 10633
},
{
"epoch": 0.2671610381801784,
"grad_norm": 0.9199085831642151,
"learning_rate": 3.871235554965218e-05,
"loss": 1.663,
"step": 10664
},
{
"epoch": 0.26793766910512073,
"grad_norm": 0.9358407855033875,
"learning_rate": 3.864205604623078e-05,
"loss": 1.6743,
"step": 10695
},
{
"epoch": 0.26871430003006314,
"grad_norm": 0.8855702877044678,
"learning_rate": 3.857160259406107e-05,
"loss": 1.6619,
"step": 10726
},
{
"epoch": 0.26949093095500554,
"grad_norm": 0.9157874584197998,
"learning_rate": 3.8500995988200674e-05,
"loss": 1.6772,
"step": 10757
},
{
"epoch": 0.2702675618799479,
"grad_norm": 0.9023963809013367,
"learning_rate": 3.843023702543556e-05,
"loss": 1.6559,
"step": 10788
},
{
"epoch": 0.2710441928048903,
"grad_norm": 0.9174203872680664,
"learning_rate": 3.8359326504270984e-05,
"loss": 1.6393,
"step": 10819
},
{
"epoch": 0.27182082372983263,
"grad_norm": 0.9019688963890076,
"learning_rate": 3.828826522492255e-05,
"loss": 1.6763,
"step": 10850
},
{
"epoch": 0.27259745465477503,
"grad_norm": 0.8955567479133606,
"learning_rate": 3.821705398930713e-05,
"loss": 1.6679,
"step": 10881
},
{
"epoch": 0.27337408557971743,
"grad_norm": 0.9326659440994263,
"learning_rate": 3.814569360103385e-05,
"loss": 1.6765,
"step": 10912
},
{
"epoch": 0.2741507165046598,
"grad_norm": 0.8885530233383179,
"learning_rate": 3.807418486539499e-05,
"loss": 1.674,
"step": 10943
},
{
"epoch": 0.2749273474296022,
"grad_norm": 0.909267783164978,
"learning_rate": 3.80025285893569e-05,
"loss": 1.6602,
"step": 10974
},
{
"epoch": 0.2757039783545445,
"grad_norm": 0.9005350470542908,
"learning_rate": 3.793072558155093e-05,
"loss": 1.6628,
"step": 11005
},
{
"epoch": 0.2764806092794869,
"grad_norm": 0.919683575630188,
"learning_rate": 3.785877665226426e-05,
"loss": 1.6647,
"step": 11036
},
{
"epoch": 0.2772572402044293,
"grad_norm": 0.8954905867576599,
"learning_rate": 3.778668261343079e-05,
"loss": 1.6763,
"step": 11067
},
{
"epoch": 0.27803387112937167,
"grad_norm": 0.9120632410049438,
"learning_rate": 3.771444427862192e-05,
"loss": 1.6552,
"step": 11098
},
{
"epoch": 0.27881050205431407,
"grad_norm": 0.9041403532028198,
"learning_rate": 3.7642062463037465e-05,
"loss": 1.6604,
"step": 11129
},
{
"epoch": 0.2795871329792564,
"grad_norm": 0.9149596095085144,
"learning_rate": 3.7569537983496373e-05,
"loss": 1.6666,
"step": 11160
},
{
"epoch": 0.2803637639041988,
"grad_norm": 0.9379541277885437,
"learning_rate": 3.749687165842753e-05,
"loss": 1.654,
"step": 11191
},
{
"epoch": 0.2811403948291412,
"grad_norm": 0.8942285180091858,
"learning_rate": 3.7424064307860536e-05,
"loss": 1.6558,
"step": 11222
},
{
"epoch": 0.28191702575408356,
"grad_norm": 0.9407718181610107,
"learning_rate": 3.735111675341645e-05,
"loss": 1.6574,
"step": 11253
},
{
"epoch": 0.28269365667902596,
"grad_norm": 0.9167685508728027,
"learning_rate": 3.7278029818298524e-05,
"loss": 1.6767,
"step": 11284
},
{
"epoch": 0.2834702876039683,
"grad_norm": 0.8957439661026001,
"learning_rate": 3.720480432728287e-05,
"loss": 1.6846,
"step": 11315
},
{
"epoch": 0.2842469185289107,
"grad_norm": 0.9044497609138489,
"learning_rate": 3.71314411067092e-05,
"loss": 1.6323,
"step": 11346
},
{
"epoch": 0.2850235494538531,
"grad_norm": 0.9024799466133118,
"learning_rate": 3.70579409844715e-05,
"loss": 1.682,
"step": 11377
},
{
"epoch": 0.28580018037879545,
"grad_norm": 0.9126555323600769,
"learning_rate": 3.698430479000865e-05,
"loss": 1.6802,
"step": 11408
},
{
"epoch": 0.28657681130373786,
"grad_norm": 0.924310564994812,
"learning_rate": 3.691053335429509e-05,
"loss": 1.6568,
"step": 11439
},
{
"epoch": 0.28735344222868026,
"grad_norm": 0.8566572666168213,
"learning_rate": 3.683662750983147e-05,
"loss": 1.6682,
"step": 11470
},
{
"epoch": 0.2881300731536226,
"grad_norm": 0.8287222981452942,
"learning_rate": 3.676258809063518e-05,
"loss": 1.6484,
"step": 11501
},
{
"epoch": 0.288906704078565,
"grad_norm": 1.0163201093673706,
"learning_rate": 3.6688415932231004e-05,
"loss": 1.635,
"step": 11532
},
{
"epoch": 0.28968333500350735,
"grad_norm": 0.8813551068305969,
"learning_rate": 3.661411187164166e-05,
"loss": 1.6654,
"step": 11563
},
{
"epoch": 0.29045996592844975,
"grad_norm": 0.8688541650772095,
"learning_rate": 3.65396767473784e-05,
"loss": 1.6787,
"step": 11594
},
{
"epoch": 0.29123659685339215,
"grad_norm": 0.8820333480834961,
"learning_rate": 3.6465111399431465e-05,
"loss": 1.6418,
"step": 11625
},
{
"epoch": 0.2920132277783345,
"grad_norm": 0.9284527897834778,
"learning_rate": 3.6390416669260674e-05,
"loss": 1.6727,
"step": 11656
},
{
"epoch": 0.2927898587032769,
"grad_norm": 0.8665077090263367,
"learning_rate": 3.63155933997859e-05,
"loss": 1.6377,
"step": 11687
},
{
"epoch": 0.29356648962821924,
"grad_norm": 0.9927760362625122,
"learning_rate": 3.624064243537758e-05,
"loss": 1.6628,
"step": 11718
},
{
"epoch": 0.29434312055316164,
"grad_norm": 0.9006340503692627,
"learning_rate": 3.616556462184716e-05,
"loss": 1.6589,
"step": 11749
},
{
"epoch": 0.29511975147810404,
"grad_norm": 0.9278168082237244,
"learning_rate": 3.609036080643755e-05,
"loss": 1.6519,
"step": 11780
},
{
"epoch": 0.2958963824030464,
"grad_norm": 0.9146036505699158,
"learning_rate": 3.60150318378136e-05,
"loss": 1.6541,
"step": 11811
},
{
"epoch": 0.2966730133279888,
"grad_norm": 0.8698804378509521,
"learning_rate": 3.5939578566052465e-05,
"loss": 1.6722,
"step": 11842
},
{
"epoch": 0.29744964425293113,
"grad_norm": 0.9304215312004089,
"learning_rate": 3.586400184263408e-05,
"loss": 1.6392,
"step": 11873
},
{
"epoch": 0.29822627517787353,
"grad_norm": 0.8917405605316162,
"learning_rate": 3.578830252043148e-05,
"loss": 1.6373,
"step": 11904
},
{
"epoch": 0.29900290610281594,
"grad_norm": 0.9075546860694885,
"learning_rate": 3.571248145370125e-05,
"loss": 1.6401,
"step": 11935
},
{
"epoch": 0.2997795370277583,
"grad_norm": 0.9235963225364685,
"learning_rate": 3.5636539498073794e-05,
"loss": 1.667,
"step": 11966
},
{
"epoch": 0.3005561679527007,
"grad_norm": 0.9514668583869934,
"learning_rate": 3.556047751054378e-05,
"loss": 1.6654,
"step": 11997
},
{
"epoch": 0.301332798877643,
"grad_norm": 0.8843172192573547,
"learning_rate": 3.548429634946039e-05,
"loss": 1.6609,
"step": 12028
},
{
"epoch": 0.3021094298025854,
"grad_norm": 0.9325070977210999,
"learning_rate": 3.540799687451768e-05,
"loss": 1.6392,
"step": 12059
},
{
"epoch": 0.30288606072752783,
"grad_norm": 0.8855581879615784,
"learning_rate": 3.533157994674485e-05,
"loss": 1.6607,
"step": 12090
},
{
"epoch": 0.3036626916524702,
"grad_norm": 0.9044367671012878,
"learning_rate": 3.5255046428496546e-05,
"loss": 1.6417,
"step": 12121
},
{
"epoch": 0.3044393225774126,
"grad_norm": 0.9446975588798523,
"learning_rate": 3.517839718344311e-05,
"loss": 1.6279,
"step": 12152
},
{
"epoch": 0.3052159535023549,
"grad_norm": 0.8764945864677429,
"learning_rate": 3.510163307656086e-05,
"loss": 1.6297,
"step": 12183
},
{
"epoch": 0.3059925844272973,
"grad_norm": 0.8633154034614563,
"learning_rate": 3.5024754974122324e-05,
"loss": 1.648,
"step": 12214
},
{
"epoch": 0.3067692153522397,
"grad_norm": 0.8891461491584778,
"learning_rate": 3.494776374368643e-05,
"loss": 1.6492,
"step": 12245
},
{
"epoch": 0.30754584627718207,
"grad_norm": 0.8888112306594849,
"learning_rate": 3.4870660254088724e-05,
"loss": 1.6507,
"step": 12276
},
{
"epoch": 0.30832247720212447,
"grad_norm": 0.8980075120925903,
"learning_rate": 3.479344537543164e-05,
"loss": 1.6571,
"step": 12307
},
{
"epoch": 0.3090991081270668,
"grad_norm": 0.86537104845047,
"learning_rate": 3.4716119979074565e-05,
"loss": 1.6331,
"step": 12338
},
{
"epoch": 0.3098757390520092,
"grad_norm": 0.9001527428627014,
"learning_rate": 3.463868493762412e-05,
"loss": 1.6453,
"step": 12369
},
{
"epoch": 0.3106523699769516,
"grad_norm": 0.8868582248687744,
"learning_rate": 3.456114112492418e-05,
"loss": 1.6623,
"step": 12400
},
{
"epoch": 0.31142900090189396,
"grad_norm": 0.899174690246582,
"learning_rate": 3.4483489416046164e-05,
"loss": 1.6451,
"step": 12431
},
{
"epoch": 0.31220563182683636,
"grad_norm": 0.9003099203109741,
"learning_rate": 3.440573068727905e-05,
"loss": 1.6544,
"step": 12462
},
{
"epoch": 0.31298226275177876,
"grad_norm": 0.8912012577056885,
"learning_rate": 3.4327865816119495e-05,
"loss": 1.6579,
"step": 12493
},
{
"epoch": 0.3137588936767211,
"grad_norm": 0.8915206789970398,
"learning_rate": 3.4249895681262025e-05,
"loss": 1.65,
"step": 12524
},
{
"epoch": 0.3145355246016635,
"grad_norm": 0.9975365400314331,
"learning_rate": 3.417182116258899e-05,
"loss": 1.6506,
"step": 12555
},
{
"epoch": 0.31531215552660585,
"grad_norm": 0.8904858827590942,
"learning_rate": 3.409364314116074e-05,
"loss": 1.6402,
"step": 12586
},
{
"epoch": 0.31608878645154825,
"grad_norm": 0.8596284985542297,
"learning_rate": 3.401536249920559e-05,
"loss": 1.6473,
"step": 12617
},
{
"epoch": 0.31686541737649065,
"grad_norm": 0.8970181345939636,
"learning_rate": 3.393698012010998e-05,
"loss": 1.6347,
"step": 12648
},
{
"epoch": 0.317642048301433,
"grad_norm": 0.890839695930481,
"learning_rate": 3.385849688840839e-05,
"loss": 1.6323,
"step": 12679
},
{
"epoch": 0.3184186792263754,
"grad_norm": 0.914010226726532,
"learning_rate": 3.3779913689773414e-05,
"loss": 1.6396,
"step": 12710
},
{
"epoch": 0.31919531015131775,
"grad_norm": 0.8762277960777283,
"learning_rate": 3.370123141100578e-05,
"loss": 1.6057,
"step": 12741
},
{
"epoch": 0.31997194107626015,
"grad_norm": 0.9001442193984985,
"learning_rate": 3.3622450940024305e-05,
"loss": 1.647,
"step": 12772
},
{
"epoch": 0.32074857200120255,
"grad_norm": 0.8792625069618225,
"learning_rate": 3.35435731658559e-05,
"loss": 1.6441,
"step": 12803
},
{
"epoch": 0.3215252029261449,
"grad_norm": 0.889738142490387,
"learning_rate": 3.346459897862552e-05,
"loss": 1.6433,
"step": 12834
},
{
"epoch": 0.3223018338510873,
"grad_norm": 0.9831507205963135,
"learning_rate": 3.338552926954613e-05,
"loss": 1.6635,
"step": 12865
},
{
"epoch": 0.32307846477602964,
"grad_norm": 0.8929276466369629,
"learning_rate": 3.330636493090868e-05,
"loss": 1.6439,
"step": 12896
},
{
"epoch": 0.32385509570097204,
"grad_norm": 0.8983376026153564,
"learning_rate": 3.322710685607193e-05,
"loss": 1.641,
"step": 12927
},
{
"epoch": 0.32463172662591444,
"grad_norm": 0.8629191517829895,
"learning_rate": 3.314775593945251e-05,
"loss": 1.6336,
"step": 12958
},
{
"epoch": 0.3254083575508568,
"grad_norm": 0.9323459267616272,
"learning_rate": 3.3068313076514714e-05,
"loss": 1.6304,
"step": 12989
},
{
"epoch": 0.3261849884757992,
"grad_norm": 0.8520583510398865,
"learning_rate": 3.298877916376047e-05,
"loss": 1.6277,
"step": 13020
},
{
"epoch": 0.32696161940074153,
"grad_norm": 0.8893365263938904,
"learning_rate": 3.290915509871915e-05,
"loss": 1.6508,
"step": 13051
},
{
"epoch": 0.32773825032568393,
"grad_norm": 0.8612290024757385,
"learning_rate": 3.282944177993753e-05,
"loss": 1.6319,
"step": 13082
},
{
"epoch": 0.32851488125062633,
"grad_norm": 0.9378547072410583,
"learning_rate": 3.274964010696957e-05,
"loss": 1.637,
"step": 13113
},
{
"epoch": 0.3292915121755687,
"grad_norm": 0.8851338028907776,
"learning_rate": 3.266975098036629e-05,
"loss": 1.6326,
"step": 13144
},
{
"epoch": 0.3300681431005111,
"grad_norm": 0.9052163362503052,
"learning_rate": 3.258977530166562e-05,
"loss": 1.6507,
"step": 13175
},
{
"epoch": 0.3308447740254534,
"grad_norm": 0.9593476057052612,
"learning_rate": 3.250971397338227e-05,
"loss": 1.6497,
"step": 13206
},
{
"epoch": 0.3316214049503958,
"grad_norm": 0.8861746788024902,
"learning_rate": 3.2429567898997404e-05,
"loss": 1.6487,
"step": 13237
},
{
"epoch": 0.3323980358753382,
"grad_norm": 0.8835756778717041,
"learning_rate": 3.234933798294859e-05,
"loss": 1.6328,
"step": 13268
},
{
"epoch": 0.33317466680028057,
"grad_norm": 0.9133350849151611,
"learning_rate": 3.2269025130619535e-05,
"loss": 1.6326,
"step": 13299
},
{
"epoch": 0.333951297725223,
"grad_norm": 0.8866952657699585,
"learning_rate": 3.218863024832985e-05,
"loss": 1.6418,
"step": 13330
},
{
"epoch": 0.3347279286501654,
"grad_norm": 0.8961243033409119,
"learning_rate": 3.2108154243324864e-05,
"loss": 1.6436,
"step": 13361
},
{
"epoch": 0.3355045595751077,
"grad_norm": 0.871108889579773,
"learning_rate": 3.2027598023765345e-05,
"loss": 1.6205,
"step": 13392
},
{
"epoch": 0.3362811905000501,
"grad_norm": 0.8666812777519226,
"learning_rate": 3.194696249871729e-05,
"loss": 1.6478,
"step": 13423
},
{
"epoch": 0.33705782142499247,
"grad_norm": 0.8760045766830444,
"learning_rate": 3.186624857814164e-05,
"loss": 1.63,
"step": 13454
},
{
"epoch": 0.33783445234993487,
"grad_norm": 0.9416362047195435,
"learning_rate": 3.178545717288401e-05,
"loss": 1.6268,
"step": 13485
},
{
"epoch": 0.33861108327487727,
"grad_norm": 0.8591843843460083,
"learning_rate": 3.170458919466444e-05,
"loss": 1.6206,
"step": 13516
},
{
"epoch": 0.3393877141998196,
"grad_norm": 0.9452426433563232,
"learning_rate": 3.1623645556067063e-05,
"loss": 1.6351,
"step": 13547
},
{
"epoch": 0.340164345124762,
"grad_norm": 0.8680084943771362,
"learning_rate": 3.154262717052985e-05,
"loss": 1.6389,
"step": 13578
},
{
"epoch": 0.34094097604970436,
"grad_norm": 0.8631142973899841,
"learning_rate": 3.146153495233426e-05,
"loss": 1.6198,
"step": 13609
},
{
"epoch": 0.34171760697464676,
"grad_norm": 0.905168354511261,
"learning_rate": 3.1380369816594944e-05,
"loss": 1.6395,
"step": 13640
},
{
"epoch": 0.34249423789958916,
"grad_norm": 0.8644616007804871,
"learning_rate": 3.129913267924946e-05,
"loss": 1.6282,
"step": 13671
},
{
"epoch": 0.3432708688245315,
"grad_norm": 0.8989911675453186,
"learning_rate": 3.121782445704782e-05,
"loss": 1.6378,
"step": 13702
},
{
"epoch": 0.3440474997494739,
"grad_norm": 0.908371090888977,
"learning_rate": 3.11364460675423e-05,
"loss": 1.6152,
"step": 13733
},
{
"epoch": 0.34482413067441625,
"grad_norm": 0.876471221446991,
"learning_rate": 3.1054998429076934e-05,
"loss": 1.6533,
"step": 13764
},
{
"epoch": 0.34560076159935865,
"grad_norm": 0.879152238368988,
"learning_rate": 3.097348246077728e-05,
"loss": 1.6196,
"step": 13795
},
{
"epoch": 0.34637739252430105,
"grad_norm": 0.8737320303916931,
"learning_rate": 3.0891899082539924e-05,
"loss": 1.6122,
"step": 13826
},
{
"epoch": 0.3471540234492434,
"grad_norm": 0.894914448261261,
"learning_rate": 3.0810249215022233e-05,
"loss": 1.6231,
"step": 13857
},
{
"epoch": 0.3479306543741858,
"grad_norm": 0.8699216842651367,
"learning_rate": 3.0728533779631865e-05,
"loss": 1.6374,
"step": 13888
},
{
"epoch": 0.34870728529912814,
"grad_norm": 0.8811262249946594,
"learning_rate": 3.064675369851637e-05,
"loss": 1.635,
"step": 13919
},
{
"epoch": 0.34948391622407055,
"grad_norm": 0.9017398953437805,
"learning_rate": 3.056490989455289e-05,
"loss": 1.6377,
"step": 13950
},
{
"epoch": 0.35026054714901295,
"grad_norm": 0.8944412469863892,
"learning_rate": 3.0483003291337596e-05,
"loss": 1.6401,
"step": 13981
},
{
"epoch": 0.3510371780739553,
"grad_norm": 0.9429489970207214,
"learning_rate": 3.040103481317539e-05,
"loss": 1.6313,
"step": 14012
},
{
"epoch": 0.3518138089988977,
"grad_norm": 0.8750442862510681,
"learning_rate": 3.03190053850694e-05,
"loss": 1.6305,
"step": 14043
},
{
"epoch": 0.35259043992384004,
"grad_norm": 0.8219168782234192,
"learning_rate": 3.0236915932710573e-05,
"loss": 1.6289,
"step": 14074
},
{
"epoch": 0.35336707084878244,
"grad_norm": 0.9778533577919006,
"learning_rate": 3.0154767382467232e-05,
"loss": 1.6221,
"step": 14105
},
{
"epoch": 0.35414370177372484,
"grad_norm": 0.8655532598495483,
"learning_rate": 3.0072560661374582e-05,
"loss": 1.6352,
"step": 14136
},
{
"epoch": 0.3549203326986672,
"grad_norm": 0.8914876580238342,
"learning_rate": 2.999029669712431e-05,
"loss": 1.6283,
"step": 14167
},
{
"epoch": 0.3556969636236096,
"grad_norm": 0.9823268055915833,
"learning_rate": 2.990797641805408e-05,
"loss": 1.6037,
"step": 14198
},
{
"epoch": 0.356473594548552,
"grad_norm": 0.8770225048065186,
"learning_rate": 2.982560075313704e-05,
"loss": 1.6199,
"step": 14229
},
{
"epoch": 0.35725022547349433,
"grad_norm": 0.8294642567634583,
"learning_rate": 2.9743170631971368e-05,
"loss": 1.587,
"step": 14260
},
{
"epoch": 0.35802685639843673,
"grad_norm": 0.8678600192070007,
"learning_rate": 2.9660686984769792e-05,
"loss": 1.6006,
"step": 14291
},
{
"epoch": 0.3588034873233791,
"grad_norm": 0.8829056620597839,
"learning_rate": 2.9578150742349047e-05,
"loss": 1.6354,
"step": 14322
},
{
"epoch": 0.3595801182483215,
"grad_norm": 0.8785590529441833,
"learning_rate": 2.949556283611942e-05,
"loss": 1.6293,
"step": 14353
},
{
"epoch": 0.3603567491732639,
"grad_norm": 0.8895497918128967,
"learning_rate": 2.9412924198074206e-05,
"loss": 1.6121,
"step": 14384
},
{
"epoch": 0.3611333800982062,
"grad_norm": 0.8808982372283936,
"learning_rate": 2.9330235760779208e-05,
"loss": 1.6251,
"step": 14415
},
{
"epoch": 0.3619100110231486,
"grad_norm": 0.8854517936706543,
"learning_rate": 2.9247498457362188e-05,
"loss": 1.617,
"step": 14446
},
{
"epoch": 0.36268664194809097,
"grad_norm": 0.9169306755065918,
"learning_rate": 2.9164713221502373e-05,
"loss": 1.632,
"step": 14477
},
{
"epoch": 0.36346327287303337,
"grad_norm": 0.9230756163597107,
"learning_rate": 2.9081880987419912e-05,
"loss": 1.6227,
"step": 14508
},
{
"epoch": 0.3642399037979758,
"grad_norm": 0.8728978037834167,
"learning_rate": 2.8999002689865296e-05,
"loss": 1.6175,
"step": 14539
},
{
"epoch": 0.3650165347229181,
"grad_norm": 0.9083303213119507,
"learning_rate": 2.8916079264108852e-05,
"loss": 1.6147,
"step": 14570
},
{
"epoch": 0.3657931656478605,
"grad_norm": 0.8577157855033875,
"learning_rate": 2.883311164593017e-05,
"loss": 1.6357,
"step": 14601
},
{
"epoch": 0.36656979657280286,
"grad_norm": 0.9495333433151245,
"learning_rate": 2.875010077160754e-05,
"loss": 1.6192,
"step": 14632
},
{
"epoch": 0.36734642749774526,
"grad_norm": 0.9099650979042053,
"learning_rate": 2.866704757790741e-05,
"loss": 1.6417,
"step": 14663
},
{
"epoch": 0.36812305842268767,
"grad_norm": 0.8684884905815125,
"learning_rate": 2.858395300207376e-05,
"loss": 1.6171,
"step": 14694
},
{
"epoch": 0.36889968934763,
"grad_norm": 0.8524884581565857,
"learning_rate": 2.8500817981817607e-05,
"loss": 1.6153,
"step": 14725
},
{
"epoch": 0.3696763202725724,
"grad_norm": 0.9122047424316406,
"learning_rate": 2.8417643455306336e-05,
"loss": 1.6223,
"step": 14756
},
{
"epoch": 0.37045295119751476,
"grad_norm": 0.875884473323822,
"learning_rate": 2.8334430361153185e-05,
"loss": 1.6364,
"step": 14787
},
{
"epoch": 0.37122958212245716,
"grad_norm": 0.8689435720443726,
"learning_rate": 2.8251179638406612e-05,
"loss": 1.639,
"step": 14818
},
{
"epoch": 0.37200621304739956,
"grad_norm": 0.8948261141777039,
"learning_rate": 2.8167892226539704e-05,
"loss": 1.6091,
"step": 14849
},
{
"epoch": 0.3727828439723419,
"grad_norm": 0.8971095085144043,
"learning_rate": 2.8084569065439588e-05,
"loss": 1.6527,
"step": 14880
},
{
"epoch": 0.3735594748972843,
"grad_norm": 0.9517078995704651,
"learning_rate": 2.8001211095396807e-05,
"loss": 1.6201,
"step": 14911
},
{
"epoch": 0.37433610582222665,
"grad_norm": 0.8965676426887512,
"learning_rate": 2.791781925709473e-05,
"loss": 1.6104,
"step": 14942
},
{
"epoch": 0.37511273674716905,
"grad_norm": 0.9771518707275391,
"learning_rate": 2.7834394491598908e-05,
"loss": 1.6329,
"step": 14973
},
{
"epoch": 0.37588936767211145,
"grad_norm": 0.8746607899665833,
"learning_rate": 2.7750937740346485e-05,
"loss": 1.6135,
"step": 15004
},
{
"epoch": 0.3766659985970538,
"grad_norm": 0.860730767250061,
"learning_rate": 2.7667449945135564e-05,
"loss": 1.6163,
"step": 15035
},
{
"epoch": 0.3774426295219962,
"grad_norm": 0.8879541158676147,
"learning_rate": 2.7583932048114557e-05,
"loss": 1.6036,
"step": 15066
},
{
"epoch": 0.37821926044693854,
"grad_norm": 0.936347246170044,
"learning_rate": 2.7500384991771587e-05,
"loss": 1.5914,
"step": 15097
},
{
"epoch": 0.37899589137188094,
"grad_norm": 0.8772497773170471,
"learning_rate": 2.7416809718923825e-05,
"loss": 1.5994,
"step": 15128
},
{
"epoch": 0.37977252229682334,
"grad_norm": 0.8785439729690552,
"learning_rate": 2.7333207172706864e-05,
"loss": 1.6057,
"step": 15159
},
{
"epoch": 0.3805491532217657,
"grad_norm": 1.1660430431365967,
"learning_rate": 2.7249578296564088e-05,
"loss": 1.6257,
"step": 15190
},
{
"epoch": 0.3813257841467081,
"grad_norm": 0.8710660934448242,
"learning_rate": 2.7165924034235973e-05,
"loss": 1.6223,
"step": 15221
},
{
"epoch": 0.3821024150716505,
"grad_norm": 0.9040597081184387,
"learning_rate": 2.708224532974953e-05,
"loss": 1.6054,
"step": 15252
},
{
"epoch": 0.38287904599659284,
"grad_norm": 0.8942875862121582,
"learning_rate": 2.6998543127407538e-05,
"loss": 1.6307,
"step": 15283
},
{
"epoch": 0.38365567692153524,
"grad_norm": 0.8733716011047363,
"learning_rate": 2.6914818371777988e-05,
"loss": 1.6298,
"step": 15314
},
{
"epoch": 0.3844323078464776,
"grad_norm": 0.9183559417724609,
"learning_rate": 2.6831072007683373e-05,
"loss": 1.6056,
"step": 15345
},
{
"epoch": 0.38520893877142,
"grad_norm": 0.8594227433204651,
"learning_rate": 2.6747304980190018e-05,
"loss": 1.6092,
"step": 15376
},
{
"epoch": 0.3859855696963624,
"grad_norm": 0.8866741061210632,
"learning_rate": 2.6663518234597453e-05,
"loss": 1.6013,
"step": 15407
},
{
"epoch": 0.38676220062130473,
"grad_norm": 0.8940442204475403,
"learning_rate": 2.6579712716427696e-05,
"loss": 1.6203,
"step": 15438
},
{
"epoch": 0.38753883154624713,
"grad_norm": 0.8408336639404297,
"learning_rate": 2.6495889371414652e-05,
"loss": 1.5997,
"step": 15469
},
{
"epoch": 0.3883154624711895,
"grad_norm": 0.8949300646781921,
"learning_rate": 2.6412049145493367e-05,
"loss": 1.6062,
"step": 15500
},
{
"epoch": 0.3890920933961319,
"grad_norm": 0.8919069766998291,
"learning_rate": 2.632819298478939e-05,
"loss": 1.6197,
"step": 15531
},
{
"epoch": 0.3898687243210743,
"grad_norm": 0.9069515466690063,
"learning_rate": 2.6244321835608105e-05,
"loss": 1.6147,
"step": 15562
},
{
"epoch": 0.3906453552460166,
"grad_norm": 0.909631073474884,
"learning_rate": 2.6160436644424024e-05,
"loss": 1.6182,
"step": 15593
},
{
"epoch": 0.391421986170959,
"grad_norm": 0.8957046866416931,
"learning_rate": 2.6076538357870133e-05,
"loss": 1.5959,
"step": 15624
},
{
"epoch": 0.39219861709590137,
"grad_norm": 0.8709155321121216,
"learning_rate": 2.5992627922727196e-05,
"loss": 1.6211,
"step": 15655
},
{
"epoch": 0.39297524802084377,
"grad_norm": 0.8602601885795593,
"learning_rate": 2.5908706285913066e-05,
"loss": 1.6032,
"step": 15686
},
{
"epoch": 0.39375187894578617,
"grad_norm": 0.853210985660553,
"learning_rate": 2.5824774394472008e-05,
"loss": 1.5994,
"step": 15717
},
{
"epoch": 0.3945285098707285,
"grad_norm": 0.8873618245124817,
"learning_rate": 2.5740833195563996e-05,
"loss": 1.5949,
"step": 15748
},
{
"epoch": 0.3953051407956709,
"grad_norm": 0.8576481342315674,
"learning_rate": 2.5656883636454067e-05,
"loss": 1.6029,
"step": 15779
},
{
"epoch": 0.39608177172061326,
"grad_norm": 0.8468286395072937,
"learning_rate": 2.557292666450159e-05,
"loss": 1.6108,
"step": 15810
},
{
"epoch": 0.39685840264555566,
"grad_norm": 0.9147023558616638,
"learning_rate": 2.5488963227149566e-05,
"loss": 1.6175,
"step": 15841
},
{
"epoch": 0.39763503357049806,
"grad_norm": 0.8911499977111816,
"learning_rate": 2.5404994271913983e-05,
"loss": 1.6232,
"step": 15872
},
{
"epoch": 0.3984116644954404,
"grad_norm": 0.8820773363113403,
"learning_rate": 2.5321020746373085e-05,
"loss": 1.6064,
"step": 15903
},
{
"epoch": 0.3991882954203828,
"grad_norm": 0.8574590086936951,
"learning_rate": 2.52370435981567e-05,
"loss": 1.6107,
"step": 15934
},
{
"epoch": 0.39996492634532516,
"grad_norm": 0.8728383183479309,
"learning_rate": 2.5153063774935533e-05,
"loss": 1.5737,
"step": 15965
},
{
"epoch": 0.40074155727026756,
"grad_norm": 0.9165883660316467,
"learning_rate": 2.506908222441045e-05,
"loss": 1.601,
"step": 15996
},
{
"epoch": 0.40151818819520996,
"grad_norm": 0.8354183435440063,
"learning_rate": 2.498509989430187e-05,
"loss": 1.5931,
"step": 16027
},
{
"epoch": 0.4022948191201523,
"grad_norm": 0.881255567073822,
"learning_rate": 2.4901117732338958e-05,
"loss": 1.6065,
"step": 16058
},
{
"epoch": 0.4030714500450947,
"grad_norm": 0.8858495950698853,
"learning_rate": 2.481713668624899e-05,
"loss": 1.6148,
"step": 16089
},
{
"epoch": 0.4038480809700371,
"grad_norm": 0.8628421425819397,
"learning_rate": 2.4733157703746663e-05,
"loss": 1.6098,
"step": 16120
},
{
"epoch": 0.40462471189497945,
"grad_norm": 0.8520597815513611,
"learning_rate": 2.4649181732523392e-05,
"loss": 1.6127,
"step": 16151
},
{
"epoch": 0.40540134281992185,
"grad_norm": 0.8664495348930359,
"learning_rate": 2.4565209720236582e-05,
"loss": 1.6039,
"step": 16182
},
{
"epoch": 0.4061779737448642,
"grad_norm": 0.8572731614112854,
"learning_rate": 2.4481242614498975e-05,
"loss": 1.6023,
"step": 16213
},
{
"epoch": 0.4069546046698066,
"grad_norm": 0.8819786906242371,
"learning_rate": 2.439728136286796e-05,
"loss": 1.6105,
"step": 16244
},
{
"epoch": 0.407731235594749,
"grad_norm": 0.8858864903450012,
"learning_rate": 2.4313326912834852e-05,
"loss": 1.5995,
"step": 16275
},
{
"epoch": 0.40850786651969134,
"grad_norm": 0.8981665372848511,
"learning_rate": 2.4229380211814206e-05,
"loss": 1.6041,
"step": 16306
},
{
"epoch": 0.40928449744463374,
"grad_norm": 0.8813167214393616,
"learning_rate": 2.4145442207133124e-05,
"loss": 1.5976,
"step": 16337
},
{
"epoch": 0.4100611283695761,
"grad_norm": 0.8797750473022461,
"learning_rate": 2.406151384602059e-05,
"loss": 1.602,
"step": 16368
},
{
"epoch": 0.4108377592945185,
"grad_norm": 0.8421767950057983,
"learning_rate": 2.3977596075596747e-05,
"loss": 1.5781,
"step": 16399
},
{
"epoch": 0.4116143902194609,
"grad_norm": 0.856469452381134,
"learning_rate": 2.3893689842862223e-05,
"loss": 1.588,
"step": 16430
},
{
"epoch": 0.41239102114440324,
"grad_norm": 0.8485891819000244,
"learning_rate": 2.3809796094687475e-05,
"loss": 1.589,
"step": 16461
},
{
"epoch": 0.41316765206934564,
"grad_norm": 0.8968520760536194,
"learning_rate": 2.372591577780202e-05,
"loss": 1.6012,
"step": 16492
},
{
"epoch": 0.413944282994288,
"grad_norm": 0.875661313533783,
"learning_rate": 2.3642049838783838e-05,
"loss": 1.6067,
"step": 16523
},
{
"epoch": 0.4147209139192304,
"grad_norm": 0.9094383716583252,
"learning_rate": 2.3558199224048666e-05,
"loss": 1.5836,
"step": 16554
},
{
"epoch": 0.4154975448441728,
"grad_norm": 0.8707971572875977,
"learning_rate": 2.347436487983929e-05,
"loss": 1.6115,
"step": 16585
},
{
"epoch": 0.41627417576911513,
"grad_norm": 0.8473120331764221,
"learning_rate": 2.3390547752214888e-05,
"loss": 1.6227,
"step": 16616
},
{
"epoch": 0.41705080669405753,
"grad_norm": 0.9022457003593445,
"learning_rate": 2.330674878704035e-05,
"loss": 1.5926,
"step": 16647
},
{
"epoch": 0.4178274376189999,
"grad_norm": 0.8699711561203003,
"learning_rate": 2.322296892997561e-05,
"loss": 1.6033,
"step": 16678
},
{
"epoch": 0.4186040685439423,
"grad_norm": 0.8574140667915344,
"learning_rate": 2.313920912646497e-05,
"loss": 1.5861,
"step": 16709
},
{
"epoch": 0.4193806994688847,
"grad_norm": 0.9694622159004211,
"learning_rate": 2.305547032172643e-05,
"loss": 1.5801,
"step": 16740
},
{
"epoch": 0.420157330393827,
"grad_norm": 0.8770648837089539,
"learning_rate": 2.2971753460741014e-05,
"loss": 1.5877,
"step": 16771
},
{
"epoch": 0.4209339613187694,
"grad_norm": 0.8674799799919128,
"learning_rate": 2.288805948824212e-05,
"loss": 1.6172,
"step": 16802
},
{
"epoch": 0.42171059224371177,
"grad_norm": 0.9614238142967224,
"learning_rate": 2.2804389348704858e-05,
"loss": 1.6031,
"step": 16833
},
{
"epoch": 0.42248722316865417,
"grad_norm": 0.8952891826629639,
"learning_rate": 2.2720743986335374e-05,
"loss": 1.5921,
"step": 16864
},
{
"epoch": 0.42326385409359657,
"grad_norm": 0.8816654682159424,
"learning_rate": 2.2637124345060233e-05,
"loss": 1.5838,
"step": 16895
},
{
"epoch": 0.4240404850185389,
"grad_norm": 0.8586190938949585,
"learning_rate": 2.2553531368515695e-05,
"loss": 1.5965,
"step": 16926
},
{
"epoch": 0.4248171159434813,
"grad_norm": 0.9344453811645508,
"learning_rate": 2.2469966000037144e-05,
"loss": 1.6041,
"step": 16957
},
{
"epoch": 0.4255937468684237,
"grad_norm": 0.924696683883667,
"learning_rate": 2.2386429182648417e-05,
"loss": 1.5944,
"step": 16988
},
{
"epoch": 0.42637037779336606,
"grad_norm": 0.8722828030586243,
"learning_rate": 2.230292185905114e-05,
"loss": 1.6213,
"step": 17019
},
{
"epoch": 0.42714700871830846,
"grad_norm": 0.8846574425697327,
"learning_rate": 2.2219444971614116e-05,
"loss": 1.6087,
"step": 17050
},
{
"epoch": 0.4279236396432508,
"grad_norm": 0.8602490425109863,
"learning_rate": 2.2135999462362655e-05,
"loss": 1.5915,
"step": 17081
},
{
"epoch": 0.4287002705681932,
"grad_norm": 0.8821462988853455,
"learning_rate": 2.2052586272968003e-05,
"loss": 1.6104,
"step": 17112
},
{
"epoch": 0.4294769014931356,
"grad_norm": 0.8978081345558167,
"learning_rate": 2.196920634473666e-05,
"loss": 1.5932,
"step": 17143
},
{
"epoch": 0.43025353241807796,
"grad_norm": 0.9008092284202576,
"learning_rate": 2.1885860618599787e-05,
"loss": 1.5915,
"step": 17174
},
{
"epoch": 0.43103016334302036,
"grad_norm": 0.9174037575721741,
"learning_rate": 2.1802550035102577e-05,
"loss": 1.6062,
"step": 17205
},
{
"epoch": 0.4318067942679627,
"grad_norm": 0.8990745544433594,
"learning_rate": 2.171927553439363e-05,
"loss": 1.6019,
"step": 17236
},
{
"epoch": 0.4325834251929051,
"grad_norm": 0.867262601852417,
"learning_rate": 2.1636038056214376e-05,
"loss": 1.5862,
"step": 17267
},
{
"epoch": 0.4333600561178475,
"grad_norm": 0.8440841436386108,
"learning_rate": 2.155283853988844e-05,
"loss": 1.5698,
"step": 17298
},
{
"epoch": 0.43413668704278985,
"grad_norm": 0.9019381999969482,
"learning_rate": 2.146967792431106e-05,
"loss": 1.6102,
"step": 17329
},
{
"epoch": 0.43491331796773225,
"grad_norm": 0.8731846213340759,
"learning_rate": 2.138655714793849e-05,
"loss": 1.5867,
"step": 17360
},
{
"epoch": 0.4356899488926746,
"grad_norm": 0.8628113269805908,
"learning_rate": 2.1303477148777367e-05,
"loss": 1.5992,
"step": 17391
},
{
"epoch": 0.436466579817617,
"grad_norm": 0.8760020732879639,
"learning_rate": 2.122043886437421e-05,
"loss": 1.5897,
"step": 17422
},
{
"epoch": 0.4372432107425594,
"grad_norm": 0.8777160048484802,
"learning_rate": 2.1137443231804765e-05,
"loss": 1.5809,
"step": 17453
},
{
"epoch": 0.43801984166750174,
"grad_norm": 0.8788965344429016,
"learning_rate": 2.105449118766347e-05,
"loss": 1.6006,
"step": 17484
},
{
"epoch": 0.43879647259244414,
"grad_norm": 0.9438194036483765,
"learning_rate": 2.097158366805287e-05,
"loss": 1.5916,
"step": 17515
},
{
"epoch": 0.4395731035173865,
"grad_norm": 0.9142041206359863,
"learning_rate": 2.0888721608573047e-05,
"loss": 1.5874,
"step": 17546
},
{
"epoch": 0.4403497344423289,
"grad_norm": 0.8893091082572937,
"learning_rate": 2.0805905944311087e-05,
"loss": 1.6032,
"step": 17577
},
{
"epoch": 0.4411263653672713,
"grad_norm": 0.8822838068008423,
"learning_rate": 2.0723137609830497e-05,
"loss": 1.6018,
"step": 17608
},
{
"epoch": 0.44190299629221363,
"grad_norm": 0.8874905109405518,
"learning_rate": 2.0640417539160686e-05,
"loss": 1.5894,
"step": 17639
},
{
"epoch": 0.44267962721715604,
"grad_norm": 0.8892821669578552,
"learning_rate": 2.0557746665786427e-05,
"loss": 1.5873,
"step": 17670
},
{
"epoch": 0.4434562581420984,
"grad_norm": 0.8921502232551575,
"learning_rate": 2.0475125922637256e-05,
"loss": 1.6111,
"step": 17701
},
{
"epoch": 0.4442328890670408,
"grad_norm": 0.908928394317627,
"learning_rate": 2.0392556242077047e-05,
"loss": 1.5754,
"step": 17732
},
{
"epoch": 0.4450095199919832,
"grad_norm": 0.9088190793991089,
"learning_rate": 2.031003855589343e-05,
"loss": 1.589,
"step": 17763
},
{
"epoch": 0.4457861509169255,
"grad_norm": 0.9038158655166626,
"learning_rate": 2.022757379528727e-05,
"loss": 1.6003,
"step": 17794
},
{
"epoch": 0.44656278184186793,
"grad_norm": 0.9049093127250671,
"learning_rate": 2.0145162890862184e-05,
"loss": 1.5752,
"step": 17825
},
{
"epoch": 0.44733941276681033,
"grad_norm": 0.827835202217102,
"learning_rate": 2.0062806772614022e-05,
"loss": 1.584,
"step": 17856
},
{
"epoch": 0.4481160436917527,
"grad_norm": 0.951187789440155,
"learning_rate": 1.9980506369920392e-05,
"loss": 1.5783,
"step": 17887
},
{
"epoch": 0.4488926746166951,
"grad_norm": 0.8784689903259277,
"learning_rate": 1.989826261153015e-05,
"loss": 1.5744,
"step": 17918
},
{
"epoch": 0.4496693055416374,
"grad_norm": 0.8900132179260254,
"learning_rate": 1.9816076425552923e-05,
"loss": 1.5695,
"step": 17949
},
{
"epoch": 0.4504459364665798,
"grad_norm": 0.8698625564575195,
"learning_rate": 1.9733948739448676e-05,
"loss": 1.5884,
"step": 17980
},
{
"epoch": 0.4512225673915222,
"grad_norm": 0.8933484554290771,
"learning_rate": 1.9651880480017155e-05,
"loss": 1.5761,
"step": 18011
},
{
"epoch": 0.45199919831646457,
"grad_norm": 0.8772753477096558,
"learning_rate": 1.9569872573387516e-05,
"loss": 1.5765,
"step": 18042
},
{
"epoch": 0.45277582924140697,
"grad_norm": 0.8958901762962341,
"learning_rate": 1.9487925945007854e-05,
"loss": 1.5815,
"step": 18073
},
{
"epoch": 0.4535524601663493,
"grad_norm": 0.8793728947639465,
"learning_rate": 1.9406041519634726e-05,
"loss": 1.5662,
"step": 18104
},
{
"epoch": 0.4543290910912917,
"grad_norm": 0.8501383066177368,
"learning_rate": 1.932422022132275e-05,
"loss": 1.5755,
"step": 18135
},
{
"epoch": 0.4551057220162341,
"grad_norm": 0.8817576766014099,
"learning_rate": 1.924246297341414e-05,
"loss": 1.6053,
"step": 18166
},
{
"epoch": 0.45588235294117646,
"grad_norm": 0.8871498703956604,
"learning_rate": 1.9160770698528338e-05,
"loss": 1.5758,
"step": 18197
},
{
"epoch": 0.45665898386611886,
"grad_norm": 0.8694120049476624,
"learning_rate": 1.907914431855156e-05,
"loss": 1.5969,
"step": 18228
},
{
"epoch": 0.4574356147910612,
"grad_norm": 0.8998878598213196,
"learning_rate": 1.8997584754626412e-05,
"loss": 1.5888,
"step": 18259
},
{
"epoch": 0.4582122457160036,
"grad_norm": 0.90102618932724,
"learning_rate": 1.8916092927141486e-05,
"loss": 1.589,
"step": 18290
},
{
"epoch": 0.458988876640946,
"grad_norm": 0.9209203720092773,
"learning_rate": 1.883466975572098e-05,
"loss": 1.569,
"step": 18321
},
{
"epoch": 0.45976550756588835,
"grad_norm": 0.906152069568634,
"learning_rate": 1.8753316159214312e-05,
"loss": 1.5681,
"step": 18352
},
{
"epoch": 0.46054213849083075,
"grad_norm": 0.8538137078285217,
"learning_rate": 1.8672033055685766e-05,
"loss": 1.5675,
"step": 18383
},
{
"epoch": 0.4613187694157731,
"grad_norm": 0.8828029036521912,
"learning_rate": 1.8590821362404116e-05,
"loss": 1.5978,
"step": 18414
},
{
"epoch": 0.4620954003407155,
"grad_norm": 0.9327139854431152,
"learning_rate": 1.8509681995832294e-05,
"loss": 1.5828,
"step": 18445
},
{
"epoch": 0.4628720312656579,
"grad_norm": 0.8733485341072083,
"learning_rate": 1.8428615871617004e-05,
"loss": 1.6059,
"step": 18476
},
{
"epoch": 0.46364866219060025,
"grad_norm": 0.8775501847267151,
"learning_rate": 1.8347623904578448e-05,
"loss": 1.5851,
"step": 18507
},
{
"epoch": 0.46442529311554265,
"grad_norm": 0.9166297316551208,
"learning_rate": 1.8266707008699975e-05,
"loss": 1.5957,
"step": 18538
},
{
"epoch": 0.465201924040485,
"grad_norm": 0.8798413276672363,
"learning_rate": 1.818586609711774e-05,
"loss": 1.5733,
"step": 18569
},
{
"epoch": 0.4659785549654274,
"grad_norm": 0.8816075921058655,
"learning_rate": 1.8105102082110462e-05,
"loss": 1.5783,
"step": 18600
},
{
"epoch": 0.4667551858903698,
"grad_norm": 0.897884726524353,
"learning_rate": 1.8024415875089058e-05,
"loss": 1.5814,
"step": 18631
},
{
"epoch": 0.46753181681531214,
"grad_norm": 0.9144574999809265,
"learning_rate": 1.7943808386586407e-05,
"loss": 1.57,
"step": 18662
},
{
"epoch": 0.46830844774025454,
"grad_norm": 0.9229109883308411,
"learning_rate": 1.7863280526247073e-05,
"loss": 1.5678,
"step": 18693
},
{
"epoch": 0.4690850786651969,
"grad_norm": 0.893386721611023,
"learning_rate": 1.7782833202817003e-05,
"loss": 1.5851,
"step": 18724
},
{
"epoch": 0.4698617095901393,
"grad_norm": 0.8726303577423096,
"learning_rate": 1.7702467324133327e-05,
"loss": 1.5823,
"step": 18755
},
{
"epoch": 0.4706383405150817,
"grad_norm": 0.8656278252601624,
"learning_rate": 1.7622183797114042e-05,
"loss": 1.5632,
"step": 18786
},
{
"epoch": 0.47141497144002403,
"grad_norm": 0.8968300819396973,
"learning_rate": 1.7541983527747838e-05,
"loss": 1.5617,
"step": 18817
},
{
"epoch": 0.47219160236496643,
"grad_norm": 0.8675340414047241,
"learning_rate": 1.746186742108387e-05,
"loss": 1.5846,
"step": 18848
},
{
"epoch": 0.47296823328990883,
"grad_norm": 0.9016551971435547,
"learning_rate": 1.73818363812215e-05,
"loss": 1.5821,
"step": 18879
},
{
"epoch": 0.4737448642148512,
"grad_norm": 0.9066888093948364,
"learning_rate": 1.7301891311300153e-05,
"loss": 1.5856,
"step": 18910
},
{
"epoch": 0.4745214951397936,
"grad_norm": 0.8297645449638367,
"learning_rate": 1.7222033113489055e-05,
"loss": 1.5766,
"step": 18941
},
{
"epoch": 0.4752981260647359,
"grad_norm": 0.8680350184440613,
"learning_rate": 1.7142262688977127e-05,
"loss": 1.6083,
"step": 18972
},
{
"epoch": 0.4760747569896783,
"grad_norm": 0.8944342136383057,
"learning_rate": 1.7062580937962764e-05,
"loss": 1.5802,
"step": 19003
},
{
"epoch": 0.47685138791462073,
"grad_norm": 0.8671095967292786,
"learning_rate": 1.698298875964369e-05,
"loss": 1.5858,
"step": 19034
},
{
"epoch": 0.4776280188395631,
"grad_norm": 0.8976803421974182,
"learning_rate": 1.690348705220684e-05,
"loss": 1.5546,
"step": 19065
},
{
"epoch": 0.4784046497645055,
"grad_norm": 0.8853389024734497,
"learning_rate": 1.6824076712818156e-05,
"loss": 1.5767,
"step": 19096
},
{
"epoch": 0.4791812806894478,
"grad_norm": 0.8673645853996277,
"learning_rate": 1.6744758637612533e-05,
"loss": 1.5868,
"step": 19127
},
{
"epoch": 0.4799579116143902,
"grad_norm": 0.8737816214561462,
"learning_rate": 1.6665533721683664e-05,
"loss": 1.5826,
"step": 19158
},
{
"epoch": 0.4807345425393326,
"grad_norm": 0.8424599170684814,
"learning_rate": 1.6586402859073974e-05,
"loss": 1.5718,
"step": 19189
},
{
"epoch": 0.48151117346427497,
"grad_norm": 0.8880824446678162,
"learning_rate": 1.6507366942764463e-05,
"loss": 1.5749,
"step": 19220
},
{
"epoch": 0.48228780438921737,
"grad_norm": 0.8780614137649536,
"learning_rate": 1.6428426864664732e-05,
"loss": 1.5848,
"step": 19251
},
{
"epoch": 0.4830644353141597,
"grad_norm": 0.8731892108917236,
"learning_rate": 1.6349583515602816e-05,
"loss": 1.5735,
"step": 19282
},
{
"epoch": 0.4838410662391021,
"grad_norm": 0.9195248484611511,
"learning_rate": 1.6270837785315208e-05,
"loss": 1.6181,
"step": 19313
},
{
"epoch": 0.4846176971640445,
"grad_norm": 0.8955127000808716,
"learning_rate": 1.619219056243676e-05,
"loss": 1.5962,
"step": 19344
},
{
"epoch": 0.48539432808898686,
"grad_norm": 0.864900529384613,
"learning_rate": 1.6113642734490698e-05,
"loss": 1.5664,
"step": 19375
},
{
"epoch": 0.48617095901392926,
"grad_norm": 0.8904951214790344,
"learning_rate": 1.6035195187878577e-05,
"loss": 1.5695,
"step": 19406
},
{
"epoch": 0.4869475899388716,
"grad_norm": 0.9016303420066833,
"learning_rate": 1.5956848807870305e-05,
"loss": 1.5536,
"step": 19437
},
{
"epoch": 0.487724220863814,
"grad_norm": 0.8600819706916809,
"learning_rate": 1.587860447859413e-05,
"loss": 1.5722,
"step": 19468
},
{
"epoch": 0.4885008517887564,
"grad_norm": 0.8715513348579407,
"learning_rate": 1.5800463083026686e-05,
"loss": 1.5776,
"step": 19499
},
{
"epoch": 0.48927748271369875,
"grad_norm": 0.8642942905426025,
"learning_rate": 1.572242550298298e-05,
"loss": 1.5711,
"step": 19530
},
{
"epoch": 0.49005411363864115,
"grad_norm": 0.877073347568512,
"learning_rate": 1.56444926191065e-05,
"loss": 1.549,
"step": 19561
},
{
"epoch": 0.4908307445635835,
"grad_norm": 0.8812013268470764,
"learning_rate": 1.5566665310859257e-05,
"loss": 1.5701,
"step": 19592
},
{
"epoch": 0.4916073754885259,
"grad_norm": 0.9232684373855591,
"learning_rate": 1.5488944456511846e-05,
"loss": 1.5736,
"step": 19623
},
{
"epoch": 0.4923840064134683,
"grad_norm": 0.9008351564407349,
"learning_rate": 1.5411330933133546e-05,
"loss": 1.5806,
"step": 19654
},
{
"epoch": 0.49316063733841065,
"grad_norm": 0.909685492515564,
"learning_rate": 1.533382561658241e-05,
"loss": 1.5643,
"step": 19685
},
{
"epoch": 0.49393726826335305,
"grad_norm": 0.8817586898803711,
"learning_rate": 1.525642938149541e-05,
"loss": 1.5751,
"step": 19716
},
{
"epoch": 0.49471389918829545,
"grad_norm": 0.853233814239502,
"learning_rate": 1.5179143101278536e-05,
"loss": 1.5655,
"step": 19747
},
{
"epoch": 0.4954905301132378,
"grad_norm": 0.9284148812294006,
"learning_rate": 1.5101967648096955e-05,
"loss": 1.5781,
"step": 19778
},
{
"epoch": 0.4962671610381802,
"grad_norm": 0.9339998364448547,
"learning_rate": 1.5024903892865172e-05,
"loss": 1.5732,
"step": 19809
},
{
"epoch": 0.49704379196312254,
"grad_norm": 0.9051497578620911,
"learning_rate": 1.4947952705237184e-05,
"loss": 1.5475,
"step": 19840
},
{
"epoch": 0.49782042288806494,
"grad_norm": 0.8712489008903503,
"learning_rate": 1.4871114953596682e-05,
"loss": 1.559,
"step": 19871
},
{
"epoch": 0.49859705381300734,
"grad_norm": 0.9316422939300537,
"learning_rate": 1.4794391505047256e-05,
"loss": 1.5615,
"step": 19902
},
{
"epoch": 0.4993736847379497,
"grad_norm": 0.8911099433898926,
"learning_rate": 1.4717783225402596e-05,
"loss": 1.5776,
"step": 19933
},
{
"epoch": 0.500150315662892,
"grad_norm": 0.8842763900756836,
"learning_rate": 1.4641290979176735e-05,
"loss": 1.5799,
"step": 19964
},
{
"epoch": 0.5009269465878344,
"grad_norm": 0.9785636067390442,
"learning_rate": 1.4564915629574246e-05,
"loss": 1.5841,
"step": 19995
},
{
"epoch": 0.5017035775127768,
"grad_norm": 0.8933955430984497,
"learning_rate": 1.4488658038480601e-05,
"loss": 1.5825,
"step": 20026
},
{
"epoch": 0.5024802084377192,
"grad_norm": 0.9056789875030518,
"learning_rate": 1.4412519066452323e-05,
"loss": 1.549,
"step": 20057
},
{
"epoch": 0.5032568393626616,
"grad_norm": 0.9087361693382263,
"learning_rate": 1.4336499572707373e-05,
"loss": 1.5448,
"step": 20088
},
{
"epoch": 0.5040334702876039,
"grad_norm": 0.8783806562423706,
"learning_rate": 1.4260600415115433e-05,
"loss": 1.5512,
"step": 20119
},
{
"epoch": 0.5048101012125463,
"grad_norm": 0.872285008430481,
"learning_rate": 1.4184822450188137e-05,
"loss": 1.5703,
"step": 20150
},
{
"epoch": 0.5055867321374887,
"grad_norm": 0.8820748329162598,
"learning_rate": 1.410916653306954e-05,
"loss": 1.5595,
"step": 20181
},
{
"epoch": 0.5063633630624311,
"grad_norm": 0.912257194519043,
"learning_rate": 1.403363351752639e-05,
"loss": 1.5795,
"step": 20212
},
{
"epoch": 0.5071399939873735,
"grad_norm": 0.9199615716934204,
"learning_rate": 1.3958224255938485e-05,
"loss": 1.5686,
"step": 20243
},
{
"epoch": 0.5079166249123158,
"grad_norm": 0.895979106426239,
"learning_rate": 1.388293959928911e-05,
"loss": 1.5777,
"step": 20274
},
{
"epoch": 0.5086932558372582,
"grad_norm": 0.9053602814674377,
"learning_rate": 1.3807780397155379e-05,
"loss": 1.5762,
"step": 20305
},
{
"epoch": 0.5094698867622006,
"grad_norm": 0.9254106283187866,
"learning_rate": 1.3732747497698655e-05,
"loss": 1.5789,
"step": 20336
},
{
"epoch": 0.510246517687143,
"grad_norm": 0.867800235748291,
"learning_rate": 1.3657841747655038e-05,
"loss": 1.5648,
"step": 20367
},
{
"epoch": 0.5110231486120854,
"grad_norm": 0.9261404871940613,
"learning_rate": 1.3583063992325706e-05,
"loss": 1.5754,
"step": 20398
},
{
"epoch": 0.5117997795370277,
"grad_norm": 0.8558968901634216,
"learning_rate": 1.3508415075567496e-05,
"loss": 1.5706,
"step": 20429
},
{
"epoch": 0.5125764104619701,
"grad_norm": 0.8890355229377747,
"learning_rate": 1.343389583978327e-05,
"loss": 1.5528,
"step": 20460
},
{
"epoch": 0.5133530413869125,
"grad_norm": 0.8697656989097595,
"learning_rate": 1.3359507125912468e-05,
"loss": 1.5647,
"step": 20491
},
{
"epoch": 0.5141296723118549,
"grad_norm": 0.8828274011611938,
"learning_rate": 1.3285249773421627e-05,
"loss": 1.5748,
"step": 20522
},
{
"epoch": 0.5149063032367973,
"grad_norm": 0.9029169678688049,
"learning_rate": 1.3211124620294884e-05,
"loss": 1.5452,
"step": 20553
},
{
"epoch": 0.5156829341617396,
"grad_norm": 2.4660046100616455,
"learning_rate": 1.313713250302451e-05,
"loss": 1.5479,
"step": 20584
},
{
"epoch": 0.516459565086682,
"grad_norm": 0.9502973556518555,
"learning_rate": 1.3063274256601479e-05,
"loss": 1.5751,
"step": 20615
},
{
"epoch": 0.5172361960116244,
"grad_norm": 0.8808090090751648,
"learning_rate": 1.2989550714506086e-05,
"loss": 1.5822,
"step": 20646
},
{
"epoch": 0.5180128269365668,
"grad_norm": 0.8710846304893494,
"learning_rate": 1.291596270869846e-05,
"loss": 1.559,
"step": 20677
},
{
"epoch": 0.5187894578615092,
"grad_norm": 0.9257534742355347,
"learning_rate": 1.284251106960927e-05,
"loss": 1.579,
"step": 20708
},
{
"epoch": 0.5195660887864515,
"grad_norm": 0.9311069250106812,
"learning_rate": 1.2769196626130263e-05,
"loss": 1.5672,
"step": 20739
},
{
"epoch": 0.5203427197113939,
"grad_norm": 0.8663937449455261,
"learning_rate": 1.2696020205604969e-05,
"loss": 1.5601,
"step": 20770
},
{
"epoch": 0.5211193506363363,
"grad_norm": 0.8988776206970215,
"learning_rate": 1.2622982633819359e-05,
"loss": 1.5538,
"step": 20801
},
{
"epoch": 0.5218959815612787,
"grad_norm": 0.8996877670288086,
"learning_rate": 1.2550084734992484e-05,
"loss": 1.5715,
"step": 20832
},
{
"epoch": 0.5226726124862211,
"grad_norm": 0.8598377704620361,
"learning_rate": 1.247732733176724e-05,
"loss": 1.5494,
"step": 20863
},
{
"epoch": 0.5234492434111634,
"grad_norm": 0.8787708282470703,
"learning_rate": 1.2404711245201044e-05,
"loss": 1.5401,
"step": 20894
},
{
"epoch": 0.5242258743361058,
"grad_norm": 0.9126632809638977,
"learning_rate": 1.2332237294756535e-05,
"loss": 1.5567,
"step": 20925
},
{
"epoch": 0.5250025052610482,
"grad_norm": 0.8951886296272278,
"learning_rate": 1.225990629829241e-05,
"loss": 1.5513,
"step": 20956
},
{
"epoch": 0.5257791361859906,
"grad_norm": 0.8855202794075012,
"learning_rate": 1.2187719072054136e-05,
"loss": 1.5408,
"step": 20987
},
{
"epoch": 0.526555767110933,
"grad_norm": 0.8703266978263855,
"learning_rate": 1.2115676430664735e-05,
"loss": 1.5481,
"step": 21018
},
{
"epoch": 0.5273323980358753,
"grad_norm": 0.8899471163749695,
"learning_rate": 1.2043779187115647e-05,
"loss": 1.5589,
"step": 21049
},
{
"epoch": 0.5281090289608177,
"grad_norm": 0.946517825126648,
"learning_rate": 1.1972028152757476e-05,
"loss": 1.56,
"step": 21080
},
{
"epoch": 0.5288856598857601,
"grad_norm": 0.9399780631065369,
"learning_rate": 1.1900424137290889e-05,
"loss": 1.5744,
"step": 21111
},
{
"epoch": 0.5296622908107025,
"grad_norm": 0.9172708988189697,
"learning_rate": 1.1828967948757482e-05,
"loss": 1.5737,
"step": 21142
},
{
"epoch": 0.5304389217356449,
"grad_norm": 0.9447731375694275,
"learning_rate": 1.175766039353062e-05,
"loss": 1.5644,
"step": 21173
},
{
"epoch": 0.5312155526605873,
"grad_norm": 0.8953673243522644,
"learning_rate": 1.1686502276306382e-05,
"loss": 1.5763,
"step": 21204
},
{
"epoch": 0.5319921835855296,
"grad_norm": 0.9340639114379883,
"learning_rate": 1.1615494400094445e-05,
"loss": 1.5508,
"step": 21235
},
{
"epoch": 0.532768814510472,
"grad_norm": 0.8861269950866699,
"learning_rate": 1.1544637566209029e-05,
"loss": 1.5558,
"step": 21266
},
{
"epoch": 0.5335454454354144,
"grad_norm": 0.8895665407180786,
"learning_rate": 1.1473932574259886e-05,
"loss": 1.5614,
"step": 21297
},
{
"epoch": 0.5343220763603568,
"grad_norm": 0.8842455148696899,
"learning_rate": 1.1403380222143247e-05,
"loss": 1.5585,
"step": 21328
},
{
"epoch": 0.5350987072852992,
"grad_norm": 0.8868069052696228,
"learning_rate": 1.1332981306032808e-05,
"loss": 1.535,
"step": 21359
},
{
"epoch": 0.5358753382102415,
"grad_norm": 0.9249818921089172,
"learning_rate": 1.1262736620370762e-05,
"loss": 1.5528,
"step": 21390
},
{
"epoch": 0.5366519691351839,
"grad_norm": 0.8665826916694641,
"learning_rate": 1.1192646957858854e-05,
"loss": 1.5656,
"step": 21421
},
{
"epoch": 0.5374286000601263,
"grad_norm": 0.9058834910392761,
"learning_rate": 1.1122713109449381e-05,
"loss": 1.5498,
"step": 21452
},
{
"epoch": 0.5382052309850687,
"grad_norm": 0.8894379138946533,
"learning_rate": 1.105293586433634e-05,
"loss": 1.5667,
"step": 21483
},
{
"epoch": 0.5389818619100111,
"grad_norm": 0.9372698068618774,
"learning_rate": 1.0983316009946446e-05,
"loss": 1.5463,
"step": 21514
},
{
"epoch": 0.5397584928349534,
"grad_norm": 0.9113041758537292,
"learning_rate": 1.0913854331930282e-05,
"loss": 1.5809,
"step": 21545
},
{
"epoch": 0.5405351237598958,
"grad_norm": 0.9373642802238464,
"learning_rate": 1.0844551614153456e-05,
"loss": 1.5664,
"step": 21576
},
{
"epoch": 0.5413117546848382,
"grad_norm": 0.8933825492858887,
"learning_rate": 1.0775408638687725e-05,
"loss": 1.555,
"step": 21607
},
{
"epoch": 0.5420883856097806,
"grad_norm": 0.9607747793197632,
"learning_rate": 1.0706426185802165e-05,
"loss": 1.5545,
"step": 21638
},
{
"epoch": 0.542865016534723,
"grad_norm": 0.8767135143280029,
"learning_rate": 1.0637605033954371e-05,
"loss": 1.5552,
"step": 21669
},
{
"epoch": 0.5436416474596653,
"grad_norm": 0.9311366081237793,
"learning_rate": 1.05689459597817e-05,
"loss": 1.5664,
"step": 21700
},
{
"epoch": 0.5444182783846077,
"grad_norm": 0.9141988158226013,
"learning_rate": 1.050044973809246e-05,
"loss": 1.5567,
"step": 21731
},
{
"epoch": 0.5451949093095501,
"grad_norm": 0.899506151676178,
"learning_rate": 1.043211714185722e-05,
"loss": 1.5702,
"step": 21762
},
{
"epoch": 0.5459715402344925,
"grad_norm": 0.8849896788597107,
"learning_rate": 1.036394894220003e-05,
"loss": 1.5609,
"step": 21793
},
{
"epoch": 0.5467481711594349,
"grad_norm": 0.904323399066925,
"learning_rate": 1.0295945908389751e-05,
"loss": 1.5607,
"step": 21824
},
{
"epoch": 0.5475248020843771,
"grad_norm": 0.9160985350608826,
"learning_rate": 1.0228108807831393e-05,
"loss": 1.5752,
"step": 21855
},
{
"epoch": 0.5483014330093195,
"grad_norm": 0.8923236131668091,
"learning_rate": 1.01604384060574e-05,
"loss": 1.5468,
"step": 21886
},
{
"epoch": 0.549078063934262,
"grad_norm": 0.8595706820487976,
"learning_rate": 1.009293546671907e-05,
"loss": 1.5394,
"step": 21917
},
{
"epoch": 0.5498546948592044,
"grad_norm": 0.9517648220062256,
"learning_rate": 1.002560075157791e-05,
"loss": 1.548,
"step": 21948
},
{
"epoch": 0.5506313257841468,
"grad_norm": 0.86150723695755,
"learning_rate": 9.958435020496995e-06,
"loss": 1.5662,
"step": 21979
},
{
"epoch": 0.551407956709089,
"grad_norm": 0.9183287620544434,
"learning_rate": 9.89143903143249e-06,
"loss": 1.5598,
"step": 22010
},
{
"epoch": 0.5521845876340314,
"grad_norm": 0.8485851287841797,
"learning_rate": 9.824613540425038e-06,
"loss": 1.5521,
"step": 22041
},
{
"epoch": 0.5529612185589738,
"grad_norm": 0.8906897306442261,
"learning_rate": 9.757959301591197e-06,
"loss": 1.5744,
"step": 22072
},
{
"epoch": 0.5537378494839162,
"grad_norm": 0.9588461518287659,
"learning_rate": 9.691477067115017e-06,
"loss": 1.557,
"step": 22103
},
{
"epoch": 0.5545144804088586,
"grad_norm": 0.9231610298156738,
"learning_rate": 9.625167587239467e-06,
"loss": 1.5609,
"step": 22134
},
{
"epoch": 0.5552911113338009,
"grad_norm": 0.9148405194282532,
"learning_rate": 9.559031610258007e-06,
"loss": 1.5533,
"step": 22165
},
{
"epoch": 0.5560677422587433,
"grad_norm": 0.9222816824913025,
"learning_rate": 9.493069882506164e-06,
"loss": 1.5558,
"step": 22196
},
{
"epoch": 0.5568443731836857,
"grad_norm": 0.8628998398780823,
"learning_rate": 9.427283148353056e-06,
"loss": 1.5721,
"step": 22227
},
{
"epoch": 0.5576210041086281,
"grad_norm": 0.8785997033119202,
"learning_rate": 9.361672150193052e-06,
"loss": 1.5538,
"step": 22258
},
{
"epoch": 0.5583976350335705,
"grad_norm": 0.9264510273933411,
"learning_rate": 9.29623762843734e-06,
"loss": 1.5606,
"step": 22289
},
{
"epoch": 0.5591742659585128,
"grad_norm": 0.9573788642883301,
"learning_rate": 9.230980321505594e-06,
"loss": 1.5596,
"step": 22320
},
{
"epoch": 0.5599508968834552,
"grad_norm": 0.8998466730117798,
"learning_rate": 9.165900965817668e-06,
"loss": 1.5421,
"step": 22351
},
{
"epoch": 0.5607275278083976,
"grad_norm": 0.9093721508979797,
"learning_rate": 9.101000295785245e-06,
"loss": 1.5575,
"step": 22382
},
{
"epoch": 0.56150415873334,
"grad_norm": 0.8681372404098511,
"learning_rate": 9.036279043803565e-06,
"loss": 1.5501,
"step": 22413
},
{
"epoch": 0.5622807896582824,
"grad_norm": 0.8595815896987915,
"learning_rate": 8.971737940243147e-06,
"loss": 1.5607,
"step": 22444
},
{
"epoch": 0.5630574205832247,
"grad_norm": 0.894629955291748,
"learning_rate": 8.907377713441592e-06,
"loss": 1.5441,
"step": 22475
},
{
"epoch": 0.5638340515081671,
"grad_norm": 0.8969973921775818,
"learning_rate": 8.843199089695293e-06,
"loss": 1.5566,
"step": 22506
},
{
"epoch": 0.5646106824331095,
"grad_norm": 0.8613621592521667,
"learning_rate": 8.779202793251311e-06,
"loss": 1.5548,
"step": 22537
},
{
"epoch": 0.5653873133580519,
"grad_norm": 0.942423403263092,
"learning_rate": 8.715389546299149e-06,
"loss": 1.5715,
"step": 22568
},
{
"epoch": 0.5661639442829943,
"grad_norm": 0.8348737955093384,
"learning_rate": 8.651760068962617e-06,
"loss": 1.5554,
"step": 22599
},
{
"epoch": 0.5669405752079366,
"grad_norm": 0.9314144849777222,
"learning_rate": 8.588315079291733e-06,
"loss": 1.5738,
"step": 22630
},
{
"epoch": 0.567717206132879,
"grad_norm": 0.9175193309783936,
"learning_rate": 8.52505529325457e-06,
"loss": 1.5369,
"step": 22661
},
{
"epoch": 0.5684938370578214,
"grad_norm": 0.9038921594619751,
"learning_rate": 8.461981424729216e-06,
"loss": 1.5518,
"step": 22692
},
{
"epoch": 0.5692704679827638,
"grad_norm": 0.9067641496658325,
"learning_rate": 8.399094185495725e-06,
"loss": 1.5554,
"step": 22723
},
{
"epoch": 0.5700470989077062,
"grad_norm": 0.8826406002044678,
"learning_rate": 8.336394285228017e-06,
"loss": 1.5648,
"step": 22754
},
{
"epoch": 0.5708237298326485,
"grad_norm": 0.8925307393074036,
"learning_rate": 8.273882431485952e-06,
"loss": 1.5504,
"step": 22785
},
{
"epoch": 0.5716003607575909,
"grad_norm": 0.9345889091491699,
"learning_rate": 8.211559329707316e-06,
"loss": 1.5451,
"step": 22816
},
{
"epoch": 0.5723769916825333,
"grad_norm": 0.913657546043396,
"learning_rate": 8.149425683199823e-06,
"loss": 1.569,
"step": 22847
},
{
"epoch": 0.5731536226074757,
"grad_norm": 0.8716319799423218,
"learning_rate": 8.08748219313325e-06,
"loss": 1.5714,
"step": 22878
},
{
"epoch": 0.5739302535324181,
"grad_norm": 0.869755744934082,
"learning_rate": 8.025729558531453e-06,
"loss": 1.5588,
"step": 22909
},
{
"epoch": 0.5747068844573605,
"grad_norm": 0.8963233232498169,
"learning_rate": 7.964168476264508e-06,
"loss": 1.5429,
"step": 22940
},
{
"epoch": 0.5754835153823028,
"grad_norm": 0.8966054916381836,
"learning_rate": 7.902799641040884e-06,
"loss": 1.5607,
"step": 22971
},
{
"epoch": 0.5762601463072452,
"grad_norm": 0.9066189527511597,
"learning_rate": 7.841623745399523e-06,
"loss": 1.5386,
"step": 23002
},
{
"epoch": 0.5770367772321876,
"grad_norm": 0.9433830976486206,
"learning_rate": 7.780641479702114e-06,
"loss": 1.5594,
"step": 23033
},
{
"epoch": 0.57781340815713,
"grad_norm": 0.9131210446357727,
"learning_rate": 7.719853532125227e-06,
"loss": 1.5359,
"step": 23064
},
{
"epoch": 0.5785900390820724,
"grad_norm": 0.880100429058075,
"learning_rate": 7.65926058865258e-06,
"loss": 1.547,
"step": 23095
},
{
"epoch": 0.5793666700070147,
"grad_norm": 0.960096538066864,
"learning_rate": 7.598863333067313e-06,
"loss": 1.5449,
"step": 23126
},
{
"epoch": 0.5801433009319571,
"grad_norm": 0.9209834337234497,
"learning_rate": 7.538662446944253e-06,
"loss": 1.5648,
"step": 23157
},
{
"epoch": 0.5809199318568995,
"grad_norm": 0.9270380139350891,
"learning_rate": 7.478658609642211e-06,
"loss": 1.5551,
"step": 23188
},
{
"epoch": 0.5816965627818419,
"grad_norm": 0.9106172919273376,
"learning_rate": 7.418852498296327e-06,
"loss": 1.5482,
"step": 23219
},
{
"epoch": 0.5824731937067843,
"grad_norm": 0.9246364831924438,
"learning_rate": 7.359244787810457e-06,
"loss": 1.5383,
"step": 23250
},
{
"epoch": 0.5832498246317266,
"grad_norm": 0.945947527885437,
"learning_rate": 7.299836150849493e-06,
"loss": 1.5497,
"step": 23281
},
{
"epoch": 0.584026455556669,
"grad_norm": 0.9162856340408325,
"learning_rate": 7.240627257831847e-06,
"loss": 1.5606,
"step": 23312
},
{
"epoch": 0.5848030864816114,
"grad_norm": 0.898065447807312,
"learning_rate": 7.1816187769218195e-06,
"loss": 1.5512,
"step": 23343
},
{
"epoch": 0.5855797174065538,
"grad_norm": 0.928261935710907,
"learning_rate": 7.1228113740220895e-06,
"loss": 1.5689,
"step": 23374
},
{
"epoch": 0.5863563483314962,
"grad_norm": 0.8689351677894592,
"learning_rate": 7.064205712766226e-06,
"loss": 1.5365,
"step": 23405
},
{
"epoch": 0.5871329792564385,
"grad_norm": 0.9178231954574585,
"learning_rate": 7.005802454511129e-06,
"loss": 1.5358,
"step": 23436
},
{
"epoch": 0.5879096101813809,
"grad_norm": 0.892164409160614,
"learning_rate": 6.947602258329639e-06,
"loss": 1.5617,
"step": 23467
},
{
"epoch": 0.5886862411063233,
"grad_norm": 0.8901685476303101,
"learning_rate": 6.889605781003078e-06,
"loss": 1.5314,
"step": 23498
},
{
"epoch": 0.5894628720312657,
"grad_norm": 0.904805064201355,
"learning_rate": 6.831813677013776e-06,
"loss": 1.5589,
"step": 23529
},
{
"epoch": 0.5902395029562081,
"grad_norm": 0.8700739145278931,
"learning_rate": 6.774226598537792e-06,
"loss": 1.5414,
"step": 23560
},
{
"epoch": 0.5910161338811504,
"grad_norm": 0.9062610864639282,
"learning_rate": 6.716845195437482e-06,
"loss": 1.5579,
"step": 23591
},
{
"epoch": 0.5917927648060928,
"grad_norm": 0.9072427749633789,
"learning_rate": 6.659670115254168e-06,
"loss": 1.5498,
"step": 23622
},
{
"epoch": 0.5925693957310352,
"grad_norm": 0.8828601837158203,
"learning_rate": 6.602702003200872e-06,
"loss": 1.5546,
"step": 23653
},
{
"epoch": 0.5933460266559776,
"grad_norm": 0.8832286596298218,
"learning_rate": 6.545941502154992e-06,
"loss": 1.5277,
"step": 23684
},
{
"epoch": 0.59412265758092,
"grad_norm": 0.9346180558204651,
"learning_rate": 6.489389252651057e-06,
"loss": 1.5555,
"step": 23715
},
{
"epoch": 0.5948992885058623,
"grad_norm": 0.9047152400016785,
"learning_rate": 6.4330458928735325e-06,
"loss": 1.5593,
"step": 23746
},
{
"epoch": 0.5956759194308047,
"grad_norm": 0.9097047448158264,
"learning_rate": 6.376912058649559e-06,
"loss": 1.5481,
"step": 23777
},
{
"epoch": 0.5964525503557471,
"grad_norm": 0.9316630959510803,
"learning_rate": 6.320988383441845e-06,
"loss": 1.5605,
"step": 23808
},
{
"epoch": 0.5972291812806895,
"grad_norm": 0.9213578701019287,
"learning_rate": 6.265275498341452e-06,
"loss": 1.5532,
"step": 23839
},
{
"epoch": 0.5980058122056319,
"grad_norm": 0.8936622738838196,
"learning_rate": 6.209774032060714e-06,
"loss": 1.576,
"step": 23870
},
{
"epoch": 0.5987824431305742,
"grad_norm": 0.907272458076477,
"learning_rate": 6.1544846109261365e-06,
"loss": 1.5403,
"step": 23901
},
{
"epoch": 0.5995590740555166,
"grad_norm": 0.943996012210846,
"learning_rate": 6.099407858871342e-06,
"loss": 1.5353,
"step": 23932
},
{
"epoch": 0.600335704980459,
"grad_norm": 0.8588092923164368,
"learning_rate": 6.044544397429958e-06,
"loss": 1.5432,
"step": 23963
},
{
"epoch": 0.6011123359054014,
"grad_norm": 0.8964900374412537,
"learning_rate": 5.989894845728708e-06,
"loss": 1.5383,
"step": 23994
},
{
"epoch": 0.6018889668303438,
"grad_norm": 0.9199158549308777,
"learning_rate": 5.9354598204803605e-06,
"loss": 1.5479,
"step": 24025
},
{
"epoch": 0.602665597755286,
"grad_norm": 0.9112527370452881,
"learning_rate": 5.881239935976762e-06,
"loss": 1.5592,
"step": 24056
},
{
"epoch": 0.6034422286802285,
"grad_norm": 0.8948724865913391,
"learning_rate": 5.827235804081954e-06,
"loss": 1.5442,
"step": 24087
},
{
"epoch": 0.6042188596051709,
"grad_norm": 0.8647116422653198,
"learning_rate": 5.773448034225221e-06,
"loss": 1.5399,
"step": 24118
},
{
"epoch": 0.6049954905301133,
"grad_norm": 0.9131345748901367,
"learning_rate": 5.719877233394228e-06,
"loss": 1.5597,
"step": 24149
},
{
"epoch": 0.6057721214550557,
"grad_norm": 0.9051525592803955,
"learning_rate": 5.666524006128191e-06,
"loss": 1.5403,
"step": 24180
},
{
"epoch": 0.606548752379998,
"grad_norm": 0.9300383925437927,
"learning_rate": 5.613388954511015e-06,
"loss": 1.5424,
"step": 24211
},
{
"epoch": 0.6073253833049403,
"grad_norm": 2.001582384109497,
"learning_rate": 5.560472678164552e-06,
"loss": 1.5169,
"step": 24242
},
{
"epoch": 0.6081020142298827,
"grad_norm": 0.9331688284873962,
"learning_rate": 5.507775774241775e-06,
"loss": 1.5582,
"step": 24273
},
{
"epoch": 0.6088786451548251,
"grad_norm": 0.8840691447257996,
"learning_rate": 5.4552988374200945e-06,
"loss": 1.5441,
"step": 24304
},
{
"epoch": 0.6096552760797675,
"grad_norm": 0.8922248482704163,
"learning_rate": 5.403042459894597e-06,
"loss": 1.5109,
"step": 24335
},
{
"epoch": 0.6104319070047098,
"grad_norm": 0.9129870533943176,
"learning_rate": 5.3510072313714135e-06,
"loss": 1.5311,
"step": 24366
},
{
"epoch": 0.6112085379296522,
"grad_norm": 0.9153522849082947,
"learning_rate": 5.2991937390610205e-06,
"loss": 1.5589,
"step": 24397
},
{
"epoch": 0.6119851688545946,
"grad_norm": 0.926564633846283,
"learning_rate": 5.247602567671625e-06,
"loss": 1.5615,
"step": 24428
},
{
"epoch": 0.612761799779537,
"grad_norm": 0.8949169516563416,
"learning_rate": 5.196234299402603e-06,
"loss": 1.5302,
"step": 24459
},
{
"epoch": 0.6135384307044794,
"grad_norm": 0.9150370955467224,
"learning_rate": 5.145089513937865e-06,
"loss": 1.5299,
"step": 24490
},
{
"epoch": 0.6143150616294217,
"grad_norm": 0.9064559936523438,
"learning_rate": 5.094168788439369e-06,
"loss": 1.5589,
"step": 24521
},
{
"epoch": 0.6150916925543641,
"grad_norm": 0.9159832000732422,
"learning_rate": 5.043472697540594e-06,
"loss": 1.5586,
"step": 24552
},
{
"epoch": 0.6158683234793065,
"grad_norm": 0.920007586479187,
"learning_rate": 4.993001813340012e-06,
"loss": 1.5293,
"step": 24583
},
{
"epoch": 0.6166449544042489,
"grad_norm": 0.8829004764556885,
"learning_rate": 4.942756705394702e-06,
"loss": 1.5531,
"step": 24614
},
{
"epoch": 0.6174215853291913,
"grad_norm": 0.8751837611198425,
"learning_rate": 4.892737940713884e-06,
"loss": 1.538,
"step": 24645
},
{
"epoch": 0.6181982162541336,
"grad_norm": 0.9076789617538452,
"learning_rate": 4.842946083752511e-06,
"loss": 1.5282,
"step": 24676
},
{
"epoch": 0.618974847179076,
"grad_norm": 0.9124575853347778,
"learning_rate": 4.79338169640493e-06,
"loss": 1.5309,
"step": 24707
},
{
"epoch": 0.6197514781040184,
"grad_norm": 0.8740583062171936,
"learning_rate": 4.74404533799851e-06,
"loss": 1.5308,
"step": 24738
},
{
"epoch": 0.6205281090289608,
"grad_norm": 0.8814982771873474,
"learning_rate": 4.694937565287344e-06,
"loss": 1.5122,
"step": 24769
},
{
"epoch": 0.6213047399539032,
"grad_norm": 0.9013394117355347,
"learning_rate": 4.646058932445985e-06,
"loss": 1.5372,
"step": 24800
},
{
"epoch": 0.6220813708788456,
"grad_norm": 0.9024440050125122,
"learning_rate": 4.597409991063148e-06,
"loss": 1.5449,
"step": 24831
},
{
"epoch": 0.6228580018037879,
"grad_norm": 0.9109915494918823,
"learning_rate": 4.5489912901355375e-06,
"loss": 1.5525,
"step": 24862
},
{
"epoch": 0.6236346327287303,
"grad_norm": 0.8715227246284485,
"learning_rate": 4.500803376061608e-06,
"loss": 1.562,
"step": 24893
},
{
"epoch": 0.6244112636536727,
"grad_norm": 0.924454391002655,
"learning_rate": 4.45284679263541e-06,
"loss": 1.5415,
"step": 24924
},
{
"epoch": 0.6251878945786151,
"grad_norm": 0.9070348739624023,
"learning_rate": 4.4051220810404775e-06,
"loss": 1.5204,
"step": 24955
},
{
"epoch": 0.6259645255035575,
"grad_norm": 0.8900454640388489,
"learning_rate": 4.3576297798437025e-06,
"loss": 1.5172,
"step": 24986
},
{
"epoch": 0.6267411564284998,
"grad_norm": 0.9387392997741699,
"learning_rate": 4.3103704249892436e-06,
"loss": 1.5577,
"step": 25017
},
{
"epoch": 0.6275177873534422,
"grad_norm": 0.9040491580963135,
"learning_rate": 4.263344549792487e-06,
"loss": 1.5575,
"step": 25048
},
{
"epoch": 0.6282944182783846,
"grad_norm": 0.9115748405456543,
"learning_rate": 4.216552684934056e-06,
"loss": 1.5373,
"step": 25079
},
{
"epoch": 0.629071049203327,
"grad_norm": 0.8901536464691162,
"learning_rate": 4.169995358453777e-06,
"loss": 1.5433,
"step": 25110
},
{
"epoch": 0.6298476801282694,
"grad_norm": 0.9255604147911072,
"learning_rate": 4.123673095744757e-06,
"loss": 1.5328,
"step": 25141
},
{
"epoch": 0.6306243110532117,
"grad_norm": 0.915124237537384,
"learning_rate": 4.077586419547435e-06,
"loss": 1.5506,
"step": 25172
},
{
"epoch": 0.6314009419781541,
"grad_norm": 0.8989410996437073,
"learning_rate": 4.03173584994368e-06,
"loss": 1.5426,
"step": 25203
},
{
"epoch": 0.6321775729030965,
"grad_norm": 0.8552297353744507,
"learning_rate": 3.986121904350948e-06,
"loss": 1.5271,
"step": 25234
},
{
"epoch": 0.6329542038280389,
"grad_norm": 0.956108033657074,
"learning_rate": 3.940745097516407e-06,
"loss": 1.5405,
"step": 25265
},
{
"epoch": 0.6337308347529813,
"grad_norm": 0.8881359696388245,
"learning_rate": 3.89560594151116e-06,
"loss": 1.535,
"step": 25296
},
{
"epoch": 0.6345074656779236,
"grad_norm": 0.8930270075798035,
"learning_rate": 3.850704945724456e-06,
"loss": 1.5405,
"step": 25327
},
{
"epoch": 0.635284096602866,
"grad_norm": 0.862516462802887,
"learning_rate": 3.8060426168579077e-06,
"loss": 1.5396,
"step": 25358
},
{
"epoch": 0.6360607275278084,
"grad_norm": 0.9361773729324341,
"learning_rate": 3.7616194589198407e-06,
"loss": 1.5546,
"step": 25389
},
{
"epoch": 0.6368373584527508,
"grad_norm": 0.897813081741333,
"learning_rate": 3.7174359732195574e-06,
"loss": 1.5479,
"step": 25420
},
{
"epoch": 0.6376139893776932,
"grad_norm": 0.9219157099723816,
"learning_rate": 3.673492658361677e-06,
"loss": 1.5404,
"step": 25451
},
{
"epoch": 0.6383906203026355,
"grad_norm": 0.9218947291374207,
"learning_rate": 3.6297900102405467e-06,
"loss": 1.5637,
"step": 25482
},
{
"epoch": 0.6391672512275779,
"grad_norm": 0.9385016560554504,
"learning_rate": 3.586328522034607e-06,
"loss": 1.5455,
"step": 25513
},
{
"epoch": 0.6399438821525203,
"grad_norm": 0.9211886525154114,
"learning_rate": 3.543108684200838e-06,
"loss": 1.5471,
"step": 25544
},
{
"epoch": 0.6407205130774627,
"grad_norm": 0.9449164867401123,
"learning_rate": 3.5001309844692464e-06,
"loss": 1.5341,
"step": 25575
},
{
"epoch": 0.6414971440024051,
"grad_norm": 0.9099578261375427,
"learning_rate": 3.4573959078373215e-06,
"loss": 1.5688,
"step": 25606
},
{
"epoch": 0.6422737749273474,
"grad_norm": 0.8888375759124756,
"learning_rate": 3.4149039365646063e-06,
"loss": 1.5408,
"step": 25637
},
{
"epoch": 0.6430504058522898,
"grad_norm": 0.9400342106819153,
"learning_rate": 3.3726555501672143e-06,
"loss": 1.5442,
"step": 25668
},
{
"epoch": 0.6438270367772322,
"grad_norm": 0.9119865298271179,
"learning_rate": 3.33065122541244e-06,
"loss": 1.542,
"step": 25699
},
{
"epoch": 0.6446036677021746,
"grad_norm": 0.9019779562950134,
"learning_rate": 3.288891436313385e-06,
"loss": 1.5281,
"step": 25730
},
{
"epoch": 0.645380298627117,
"grad_norm": 0.8899328708648682,
"learning_rate": 3.2473766541235963e-06,
"loss": 1.5382,
"step": 25761
},
{
"epoch": 0.6461569295520593,
"grad_norm": 0.9054051637649536,
"learning_rate": 3.2061073473317466e-06,
"loss": 1.5604,
"step": 25792
},
{
"epoch": 0.6469335604770017,
"grad_norm": 0.863438606262207,
"learning_rate": 3.1650839816563444e-06,
"loss": 1.5296,
"step": 25823
},
{
"epoch": 0.6477101914019441,
"grad_norm": 0.8931278586387634,
"learning_rate": 3.1243070200405093e-06,
"loss": 1.5461,
"step": 25854
},
{
"epoch": 0.6484868223268865,
"grad_norm": 0.9182883501052856,
"learning_rate": 3.0837769226467e-06,
"loss": 1.5296,
"step": 25885
},
{
"epoch": 0.6492634532518289,
"grad_norm": 0.9026657938957214,
"learning_rate": 3.0434941468515666e-06,
"loss": 1.5465,
"step": 25916
},
{
"epoch": 0.6500400841767712,
"grad_norm": 0.8940238952636719,
"learning_rate": 3.003459147240753e-06,
"loss": 1.5197,
"step": 25947
},
{
"epoch": 0.6508167151017136,
"grad_norm": 0.8756072521209717,
"learning_rate": 2.9636723756037875e-06,
"loss": 1.5416,
"step": 25978
},
{
"epoch": 0.651593346026656,
"grad_norm": 0.9100384712219238,
"learning_rate": 2.9241342809289833e-06,
"loss": 1.5403,
"step": 26009
},
{
"epoch": 0.6523699769515984,
"grad_norm": 0.9010562300682068,
"learning_rate": 2.8848453093983594e-06,
"loss": 1.5407,
"step": 26040
},
{
"epoch": 0.6531466078765408,
"grad_norm": 0.8857883214950562,
"learning_rate": 2.8458059043826257e-06,
"loss": 1.5366,
"step": 26071
},
{
"epoch": 0.6539232388014831,
"grad_norm": 0.8999698758125305,
"learning_rate": 2.807016506436172e-06,
"loss": 1.529,
"step": 26102
},
{
"epoch": 0.6546998697264255,
"grad_norm": 0.9420500993728638,
"learning_rate": 2.7684775532920566e-06,
"loss": 1.5195,
"step": 26133
},
{
"epoch": 0.6554765006513679,
"grad_norm": 0.9227805733680725,
"learning_rate": 2.7301894798571425e-06,
"loss": 1.5563,
"step": 26164
},
{
"epoch": 0.6562531315763103,
"grad_norm": 0.9569822549819946,
"learning_rate": 2.6921527182071386e-06,
"loss": 1.5239,
"step": 26195
},
{
"epoch": 0.6570297625012527,
"grad_norm": 0.9220489859580994,
"learning_rate": 2.654367697581725e-06,
"loss": 1.5485,
"step": 26226
},
{
"epoch": 0.657806393426195,
"grad_norm": 0.9232506155967712,
"learning_rate": 2.6168348443797175e-06,
"loss": 1.5451,
"step": 26257
},
{
"epoch": 0.6585830243511374,
"grad_norm": 0.9334014058113098,
"learning_rate": 2.5795545821542757e-06,
"loss": 1.5228,
"step": 26288
},
{
"epoch": 0.6593596552760798,
"grad_norm": 0.9631486535072327,
"learning_rate": 2.54252733160808e-06,
"loss": 1.5595,
"step": 26319
},
{
"epoch": 0.6601362862010222,
"grad_norm": 0.9698452353477478,
"learning_rate": 2.5057535105886294e-06,
"loss": 1.5273,
"step": 26350
},
{
"epoch": 0.6609129171259646,
"grad_norm": 0.8972797989845276,
"learning_rate": 2.4692335340834953e-06,
"loss": 1.5373,
"step": 26381
},
{
"epoch": 0.6616895480509069,
"grad_norm": 0.8990224599838257,
"learning_rate": 2.432967814215639e-06,
"loss": 1.541,
"step": 26412
},
{
"epoch": 0.6624661789758493,
"grad_norm": 0.9032983183860779,
"learning_rate": 2.396956760238794e-06,
"loss": 1.5189,
"step": 26443
},
{
"epoch": 0.6632428099007917,
"grad_norm": 0.9151228666305542,
"learning_rate": 2.361200778532796e-06,
"loss": 1.5277,
"step": 26474
},
{
"epoch": 0.664019440825734,
"grad_norm": 0.8908092975616455,
"learning_rate": 2.325700272599049e-06,
"loss": 1.5431,
"step": 26505
},
{
"epoch": 0.6647960717506765,
"grad_norm": 0.9372129440307617,
"learning_rate": 2.2904556430559415e-06,
"loss": 1.5373,
"step": 26536
},
{
"epoch": 0.6655727026756189,
"grad_norm": 0.9178755879402161,
"learning_rate": 2.2554672876343106e-06,
"loss": 1.551,
"step": 26567
},
{
"epoch": 0.6663493336005611,
"grad_norm": 0.9076238870620728,
"learning_rate": 2.220735601173002e-06,
"loss": 1.5471,
"step": 26598
},
{
"epoch": 0.6671259645255035,
"grad_norm": 0.8949811458587646,
"learning_rate": 2.186260975614382e-06,
"loss": 1.5498,
"step": 26629
},
{
"epoch": 0.667902595450446,
"grad_norm": 0.9235818386077881,
"learning_rate": 2.1520437999999034e-06,
"loss": 1.5445,
"step": 26660
},
{
"epoch": 0.6686792263753883,
"grad_norm": 0.884668231010437,
"learning_rate": 2.1180844604657526e-06,
"loss": 1.5344,
"step": 26691
},
{
"epoch": 0.6694558573003307,
"grad_norm": 0.8701885938644409,
"learning_rate": 2.084383340238455e-06,
"loss": 1.5268,
"step": 26722
},
{
"epoch": 0.670232488225273,
"grad_norm": 0.9150304198265076,
"learning_rate": 2.0509408196305704e-06,
"loss": 1.5445,
"step": 26753
},
{
"epoch": 0.6710091191502154,
"grad_norm": 0.9240284562110901,
"learning_rate": 2.017757276036403e-06,
"loss": 1.5694,
"step": 26784
},
{
"epoch": 0.6717857500751578,
"grad_norm": 0.9412962794303894,
"learning_rate": 1.984833083927726e-06,
"loss": 1.5657,
"step": 26815
},
{
"epoch": 0.6725623810001002,
"grad_norm": 0.9341692924499512,
"learning_rate": 1.952168614849581e-06,
"loss": 1.5396,
"step": 26846
},
{
"epoch": 0.6733390119250426,
"grad_norm": 0.9173117876052856,
"learning_rate": 1.919764237416058e-06,
"loss": 1.5496,
"step": 26877
},
{
"epoch": 0.6741156428499849,
"grad_norm": 0.8984019160270691,
"learning_rate": 1.8876203173061463e-06,
"loss": 1.5385,
"step": 26908
},
{
"epoch": 0.6748922737749273,
"grad_norm": 0.9131044745445251,
"learning_rate": 1.8557372172596206e-06,
"loss": 1.5551,
"step": 26939
},
{
"epoch": 0.6756689046998697,
"grad_norm": 0.934843897819519,
"learning_rate": 1.8241152970729341e-06,
"loss": 1.5497,
"step": 26970
},
{
"epoch": 0.6764455356248121,
"grad_norm": 0.878828227519989,
"learning_rate": 1.7927549135951572e-06,
"loss": 1.5587,
"step": 27001
},
{
"epoch": 0.6772221665497545,
"grad_norm": 0.9447210431098938,
"learning_rate": 1.7616564207239477e-06,
"loss": 1.5403,
"step": 27032
},
{
"epoch": 0.6779987974746968,
"grad_norm": 0.9129860997200012,
"learning_rate": 1.730820169401584e-06,
"loss": 1.5343,
"step": 27063
},
{
"epoch": 0.6787754283996392,
"grad_norm": 0.9327501058578491,
"learning_rate": 1.7002465076109558e-06,
"loss": 1.5279,
"step": 27094
},
{
"epoch": 0.6795520593245816,
"grad_norm": 0.9292470216751099,
"learning_rate": 1.6699357803716898e-06,
"loss": 1.5316,
"step": 27125
},
{
"epoch": 0.680328690249524,
"grad_norm": 0.9054301977157593,
"learning_rate": 1.6398883297362305e-06,
"loss": 1.5257,
"step": 27156
},
{
"epoch": 0.6811053211744664,
"grad_norm": 0.8920509815216064,
"learning_rate": 1.6101044947859606e-06,
"loss": 1.5451,
"step": 27187
},
{
"epoch": 0.6818819520994087,
"grad_norm": 0.8931530714035034,
"learning_rate": 1.5805846116274114e-06,
"loss": 1.5316,
"step": 27218
},
{
"epoch": 0.6826585830243511,
"grad_norm": 0.9331194758415222,
"learning_rate": 1.5513290133884611e-06,
"loss": 1.5352,
"step": 27249
},
{
"epoch": 0.6834352139492935,
"grad_norm": 0.9129217267036438,
"learning_rate": 1.5223380302145512e-06,
"loss": 1.536,
"step": 27280
},
{
"epoch": 0.6842118448742359,
"grad_norm": 0.9494413137435913,
"learning_rate": 1.4936119892649925e-06,
"loss": 1.519,
"step": 27311
},
{
"epoch": 0.6849884757991783,
"grad_norm": 0.9152242541313171,
"learning_rate": 1.4651512147092482e-06,
"loss": 1.5227,
"step": 27342
},
{
"epoch": 0.6857651067241206,
"grad_norm": 0.9150066375732422,
"learning_rate": 1.4369560277232908e-06,
"loss": 1.5323,
"step": 27373
},
{
"epoch": 0.686541737649063,
"grad_norm": 0.9378058910369873,
"learning_rate": 1.409026746485978e-06,
"loss": 1.5414,
"step": 27404
},
{
"epoch": 0.6873183685740054,
"grad_norm": 0.9014245867729187,
"learning_rate": 1.3813636861754464e-06,
"loss": 1.5541,
"step": 27435
},
{
"epoch": 0.6880949994989478,
"grad_norm": 0.9115684628486633,
"learning_rate": 1.3539671589655773e-06,
"loss": 1.5481,
"step": 27466
},
{
"epoch": 0.6888716304238902,
"grad_norm": 0.9748349785804749,
"learning_rate": 1.3268374740224548e-06,
"loss": 1.5266,
"step": 27497
},
{
"epoch": 0.6896482613488325,
"grad_norm": 0.9160287976264954,
"learning_rate": 1.2999749375008807e-06,
"loss": 1.5253,
"step": 27528
},
{
"epoch": 0.6904248922737749,
"grad_norm": 0.9088099002838135,
"learning_rate": 1.2733798525409346e-06,
"loss": 1.5405,
"step": 27559
},
{
"epoch": 0.6912015231987173,
"grad_norm": 0.9275810718536377,
"learning_rate": 1.2470525192645383e-06,
"loss": 1.538,
"step": 27590
},
{
"epoch": 0.6919781541236597,
"grad_norm": 0.8710309863090515,
"learning_rate": 1.2209932347720666e-06,
"loss": 1.5489,
"step": 27621
},
{
"epoch": 0.6927547850486021,
"grad_norm": 0.9207929968833923,
"learning_rate": 1.1952022931389972e-06,
"loss": 1.5322,
"step": 27652
},
{
"epoch": 0.6935314159735444,
"grad_norm": 0.9486070871353149,
"learning_rate": 1.1696799854126083e-06,
"loss": 1.5305,
"step": 27683
},
{
"epoch": 0.6943080468984868,
"grad_norm": 0.8901826739311218,
"learning_rate": 1.1444265996086694e-06,
"loss": 1.5498,
"step": 27714
},
{
"epoch": 0.6950846778234292,
"grad_norm": 0.9303523302078247,
"learning_rate": 1.119442420708211e-06,
"loss": 1.5485,
"step": 27745
},
{
"epoch": 0.6958613087483716,
"grad_norm": 1.046007513999939,
"learning_rate": 1.0947277306542964e-06,
"loss": 1.5512,
"step": 27776
},
{
"epoch": 0.696637939673314,
"grad_norm": 0.9494882225990295,
"learning_rate": 1.0702828083488353e-06,
"loss": 1.5154,
"step": 27807
},
{
"epoch": 0.6974145705982563,
"grad_norm": 0.9443824291229248,
"learning_rate": 1.0461079296494647e-06,
"loss": 1.5367,
"step": 27838
},
{
"epoch": 0.6981912015231987,
"grad_norm": 0.9392052292823792,
"learning_rate": 1.0222033673663978e-06,
"loss": 1.5383,
"step": 27869
},
{
"epoch": 0.6989678324481411,
"grad_norm": 0.8942481279373169,
"learning_rate": 9.985693912593713e-07,
"loss": 1.5513,
"step": 27900
},
{
"epoch": 0.6997444633730835,
"grad_norm": 0.9549274444580078,
"learning_rate": 9.752062680346035e-07,
"loss": 1.5589,
"step": 27931
},
{
"epoch": 0.7005210942980259,
"grad_norm": 0.8858839869499207,
"learning_rate": 9.521142613417494e-07,
"loss": 1.527,
"step": 27962
},
{
"epoch": 0.7012977252229682,
"grad_norm": 0.9377241730690002,
"learning_rate": 9.292936317709722e-07,
"loss": 1.5354,
"step": 27993
},
{
"epoch": 0.7020743561479106,
"grad_norm": 0.9183421730995178,
"learning_rate": 9.067446368499793e-07,
"loss": 1.5365,
"step": 28024
},
{
"epoch": 0.702850987072853,
"grad_norm": 0.9407469630241394,
"learning_rate": 8.844675310411055e-07,
"loss": 1.5392,
"step": 28055
},
{
"epoch": 0.7036276179977954,
"grad_norm": 0.9024103879928589,
"learning_rate": 8.6246256573847e-07,
"loss": 1.5275,
"step": 28086
},
{
"epoch": 0.7044042489227378,
"grad_norm": 0.9077733755111694,
"learning_rate": 8.407299892651127e-07,
"loss": 1.5285,
"step": 28117
},
{
"epoch": 0.7051808798476801,
"grad_norm": 0.8907241821289062,
"learning_rate": 8.19270046870202e-07,
"loss": 1.5283,
"step": 28148
},
{
"epoch": 0.7059575107726225,
"grad_norm": 0.9342151284217834,
"learning_rate": 7.980829807262752e-07,
"loss": 1.5493,
"step": 28179
},
{
"epoch": 0.7067341416975649,
"grad_norm": 0.9555601477622986,
"learning_rate": 7.771690299264889e-07,
"loss": 1.5452,
"step": 28210
},
{
"epoch": 0.7075107726225073,
"grad_norm": 0.9291635155677795,
"learning_rate": 7.565284304819426e-07,
"loss": 1.5286,
"step": 28241
},
{
"epoch": 0.7082874035474497,
"grad_norm": 0.9101645946502686,
"learning_rate": 7.361614153189922e-07,
"loss": 1.5085,
"step": 28272
},
{
"epoch": 0.709064034472392,
"grad_norm": 0.9269605278968811,
"learning_rate": 7.160682142766328e-07,
"loss": 1.5366,
"step": 28303
},
{
"epoch": 0.7098406653973344,
"grad_norm": 0.9426531791687012,
"learning_rate": 6.962490541039091e-07,
"loss": 1.5571,
"step": 28334
},
{
"epoch": 0.7106172963222768,
"grad_norm": 0.9087902903556824,
"learning_rate": 6.767041584573531e-07,
"loss": 1.5253,
"step": 28365
},
{
"epoch": 0.7113939272472192,
"grad_norm": 0.9450654983520508,
"learning_rate": 6.574337478984532e-07,
"loss": 1.5361,
"step": 28396
},
{
"epoch": 0.7121705581721616,
"grad_norm": 0.9181643128395081,
"learning_rate": 6.384380398911732e-07,
"loss": 1.5346,
"step": 28427
},
{
"epoch": 0.712947189097104,
"grad_norm": 0.9480751752853394,
"learning_rate": 6.197172487994951e-07,
"loss": 1.5218,
"step": 28458
},
{
"epoch": 0.7137238200220463,
"grad_norm": 0.947593629360199,
"learning_rate": 6.012715858850021e-07,
"loss": 1.5611,
"step": 28489
},
{
"epoch": 0.7145004509469887,
"grad_norm": 0.9011168479919434,
"learning_rate": 5.831012593044971e-07,
"loss": 1.5437,
"step": 28520
},
{
"epoch": 0.7152770818719311,
"grad_norm": 0.8987276554107666,
"learning_rate": 5.652064741076435e-07,
"loss": 1.5348,
"step": 28551
},
{
"epoch": 0.7160537127968735,
"grad_norm": 0.9200072288513184,
"learning_rate": 5.475874322346558e-07,
"loss": 1.5614,
"step": 28582
},
{
"epoch": 0.7168303437218159,
"grad_norm": 0.9210062026977539,
"learning_rate": 5.30244332514035e-07,
"loss": 1.5238,
"step": 28613
},
{
"epoch": 0.7176069746467582,
"grad_norm": 0.9360731244087219,
"learning_rate": 5.131773706602977e-07,
"loss": 1.544,
"step": 28644
},
{
"epoch": 0.7183836055717006,
"grad_norm": 0.9032440185546875,
"learning_rate": 4.963867392717897e-07,
"loss": 1.5359,
"step": 28675
},
{
"epoch": 0.719160236496643,
"grad_norm": 0.910786509513855,
"learning_rate": 4.798726278285093e-07,
"loss": 1.5312,
"step": 28706
},
{
"epoch": 0.7199368674215854,
"grad_norm": 0.9738684296607971,
"learning_rate": 4.6363522268995097e-07,
"loss": 1.5173,
"step": 28737
},
{
"epoch": 0.7207134983465278,
"grad_norm": 0.9202192425727844,
"learning_rate": 4.4767470709302927e-07,
"loss": 1.5445,
"step": 28768
},
{
"epoch": 0.72149012927147,
"grad_norm": 0.9537370800971985,
"learning_rate": 4.319912611499971e-07,
"loss": 1.5318,
"step": 28799
},
{
"epoch": 0.7222667601964124,
"grad_norm": 0.9033458232879639,
"learning_rate": 4.1658506184640564e-07,
"loss": 1.5215,
"step": 28830
},
{
"epoch": 0.7230433911213549,
"grad_norm": 0.897890567779541,
"learning_rate": 4.0145628303911996e-07,
"loss": 1.532,
"step": 28861
},
{
"epoch": 0.7238200220462973,
"grad_norm": 0.9329019784927368,
"learning_rate": 3.866050954543565e-07,
"loss": 1.5374,
"step": 28892
},
{
"epoch": 0.7245966529712397,
"grad_norm": 0.9132122993469238,
"learning_rate": 3.720316666857432e-07,
"loss": 1.5292,
"step": 28923
},
{
"epoch": 0.7253732838961819,
"grad_norm": 0.9273849129676819,
"learning_rate": 3.5773616119244845e-07,
"loss": 1.5229,
"step": 28954
},
{
"epoch": 0.7261499148211243,
"grad_norm": 0.8980498909950256,
"learning_rate": 3.437187402973052e-07,
"loss": 1.5167,
"step": 28985
},
{
"epoch": 0.7269265457460667,
"grad_norm": 0.9546352028846741,
"learning_rate": 3.2997956218500104e-07,
"loss": 1.5321,
"step": 29016
},
{
"epoch": 0.7277031766710091,
"grad_norm": 0.87969970703125,
"learning_rate": 3.165187819003018e-07,
"loss": 1.5308,
"step": 29047
},
{
"epoch": 0.7284798075959515,
"grad_norm": 0.9020631313323975,
"learning_rate": 3.033365513462755e-07,
"loss": 1.5387,
"step": 29078
},
{
"epoch": 0.7292564385208938,
"grad_norm": 0.9020014405250549,
"learning_rate": 2.9043301928260437e-07,
"loss": 1.5278,
"step": 29109
},
{
"epoch": 0.7300330694458362,
"grad_norm": 0.8693344593048096,
"learning_rate": 2.7780833132389773e-07,
"loss": 1.5211,
"step": 29140
},
{
"epoch": 0.7308097003707786,
"grad_norm": 0.9200028777122498,
"learning_rate": 2.6546262993803473e-07,
"loss": 1.5627,
"step": 29171
},
{
"epoch": 0.731586331295721,
"grad_norm": 0.8627704977989197,
"learning_rate": 2.533960544445879e-07,
"loss": 1.553,
"step": 29202
},
{
"epoch": 0.7323629622206634,
"grad_norm": 0.9087816476821899,
"learning_rate": 2.416087410132134e-07,
"loss": 1.5218,
"step": 29233
},
{
"epoch": 0.7331395931456057,
"grad_norm": 0.9687427282333374,
"learning_rate": 2.301008226621465e-07,
"loss": 1.5471,
"step": 29264
},
{
"epoch": 0.7339162240705481,
"grad_norm": 0.9520723223686218,
"learning_rate": 2.1887242925668073e-07,
"loss": 1.5318,
"step": 29295
},
{
"epoch": 0.7346928549954905,
"grad_norm": 0.916723906993866,
"learning_rate": 2.0792368750770785e-07,
"loss": 1.5413,
"step": 29326
},
{
"epoch": 0.7354694859204329,
"grad_norm": 0.9351438283920288,
"learning_rate": 1.9725472097028851e-07,
"loss": 1.5393,
"step": 29357
},
{
"epoch": 0.7362461168453753,
"grad_norm": 0.9236180186271667,
"learning_rate": 1.8686565004226718e-07,
"loss": 1.5314,
"step": 29388
},
{
"epoch": 0.7370227477703176,
"grad_norm": 0.921168327331543,
"learning_rate": 1.7675659196288995e-07,
"loss": 1.5469,
"step": 29419
},
{
"epoch": 0.73779937869526,
"grad_norm": 0.8722018599510193,
"learning_rate": 1.6692766081150556e-07,
"loss": 1.5183,
"step": 29450
},
{
"epoch": 0.7385760096202024,
"grad_norm": 0.9068271517753601,
"learning_rate": 1.5737896750626647e-07,
"loss": 1.5467,
"step": 29481
},
{
"epoch": 0.7393526405451448,
"grad_norm": 0.8897390961647034,
"learning_rate": 1.4811061980287976e-07,
"loss": 1.5517,
"step": 29512
},
{
"epoch": 0.7401292714700872,
"grad_norm": 0.8943518996238708,
"learning_rate": 1.3912272229338886e-07,
"loss": 1.5143,
"step": 29543
},
{
"epoch": 0.7409059023950295,
"grad_norm": 0.9387959241867065,
"learning_rate": 1.3041537640499645e-07,
"loss": 1.5579,
"step": 29574
},
{
"epoch": 0.7416825333199719,
"grad_norm": 0.9111166000366211,
"learning_rate": 1.2198868039891564e-07,
"loss": 1.5281,
"step": 29605
},
{
"epoch": 0.7424591642449143,
"grad_norm": 0.9361575245857239,
"learning_rate": 1.138427293692651e-07,
"loss": 1.5488,
"step": 29636
},
{
"epoch": 0.7432357951698567,
"grad_norm": 0.9410037398338318,
"learning_rate": 1.0597761524199778e-07,
"loss": 1.5175,
"step": 29667
},
{
"epoch": 0.7440124260947991,
"grad_norm": 0.9299407601356506,
"learning_rate": 9.839342677385455e-08,
"loss": 1.5295,
"step": 29698
},
{
"epoch": 0.7447890570197414,
"grad_norm": 0.9386714100837708,
"learning_rate": 9.109024955137325e-08,
"loss": 1.5498,
"step": 29729
},
{
"epoch": 0.7455656879446838,
"grad_norm": 0.9419563412666321,
"learning_rate": 8.406816598991729e-08,
"loss": 1.5222,
"step": 29760
},
{
"epoch": 0.7463423188696262,
"grad_norm": 0.9137088656425476,
"learning_rate": 7.73272553327431e-08,
"loss": 1.5529,
"step": 29791
},
{
"epoch": 0.7471189497945686,
"grad_norm": 0.9149454236030579,
"learning_rate": 7.086759365011186e-08,
"loss": 1.5208,
"step": 29822
},
{
"epoch": 0.747895580719511,
"grad_norm": 0.9159924387931824,
"learning_rate": 6.468925383842639e-08,
"loss": 1.5515,
"step": 29853
},
{
"epoch": 0.7486722116444533,
"grad_norm": 0.9082822203636169,
"learning_rate": 5.8792305619415067e-08,
"loss": 1.5258,
"step": 29884
},
{
"epoch": 0.7494488425693957,
"grad_norm": 0.9176669120788574,
"learning_rate": 5.317681553933529e-08,
"loss": 1.5567,
"step": 29915
},
{
"epoch": 0.7502254734943381,
"grad_norm": 0.9306588172912598,
"learning_rate": 4.78428469682296e-08,
"loss": 1.5395,
"step": 29946
},
{
"epoch": 0.7510021044192805,
"grad_norm": 0.9428490996360779,
"learning_rate": 4.2790460099206844e-08,
"loss": 1.5498,
"step": 29977
},
{
"epoch": 0.7517787353442229,
"grad_norm": 0.8921723961830139,
"learning_rate": 3.801971194777043e-08,
"loss": 1.5456,
"step": 30008
},
{
"epoch": 0.7525553662691652,
"grad_norm": 0.9158675670623779,
"learning_rate": 3.353065635115782e-08,
"loss": 1.5148,
"step": 30039
},
{
"epoch": 0.7533319971941076,
"grad_norm": 0.9300774335861206,
"learning_rate": 2.93233439677576e-08,
"loss": 1.5298,
"step": 30070
},
{
"epoch": 0.75410862811905,
"grad_norm": 0.8786063194274902,
"learning_rate": 2.539782227651555e-08,
"loss": 1.5023,
"step": 30101
},
{
"epoch": 0.7548852590439924,
"grad_norm": 0.9041824340820312,
"learning_rate": 2.175413557641004e-08,
"loss": 1.5372,
"step": 30132
},
{
"epoch": 0.7556618899689348,
"grad_norm": 0.9231988787651062,
"learning_rate": 1.839232498594967e-08,
"loss": 1.5502,
"step": 30163
},
{
"epoch": 0.7564385208938771,
"grad_norm": 0.9001336097717285,
"learning_rate": 1.5312428442712522e-08,
"loss": 1.5239,
"step": 30194
},
{
"epoch": 0.7572151518188195,
"grad_norm": 0.9294939637184143,
"learning_rate": 1.2514480702913168e-08,
"loss": 1.5117,
"step": 30225
},
{
"epoch": 0.7579917827437619,
"grad_norm": 0.874643087387085,
"learning_rate": 9.998513341005766e-09,
"loss": 1.5402,
"step": 30256
},
{
"epoch": 0.7587684136687043,
"grad_norm": 0.8916767239570618,
"learning_rate": 7.764554749345454e-09,
"loss": 1.5303,
"step": 30287
},
{
"epoch": 0.7595450445936467,
"grad_norm": 0.9058026075363159,
"learning_rate": 5.812630137849717e-09,
"loss": 1.5194,
"step": 30318
},
{
"epoch": 0.7603216755185891,
"grad_norm": 0.898659348487854,
"learning_rate": 4.142761533723616e-09,
"loss": 1.509,
"step": 30349
},
{
"epoch": 0.7610983064435314,
"grad_norm": 0.8996184468269348,
"learning_rate": 2.7549677812044317e-09,
"loss": 1.527,
"step": 30380
},
{
"epoch": 0.7618749373684738,
"grad_norm": 0.9288005232810974,
"learning_rate": 1.6492645413590525e-09,
"loss": 1.5369,
"step": 30411
},
{
"epoch": 0.7626515682934162,
"grad_norm": 0.9407172203063965,
"learning_rate": 8.256642918980096e-10,
"loss": 1.5501,
"step": 30442
},
{
"epoch": 0.7634281992183586,
"grad_norm": 0.905489444732666,
"learning_rate": 2.841763270367004e-10,
"loss": 1.5464,
"step": 30473
},
{
"epoch": 0.764204830143301,
"grad_norm": 0.8837622404098511,
"learning_rate": 2.480675739269245e-11,
"loss": 1.531,
"step": 30504
}
],
"logging_steps": 31,
"max_steps": 30517,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 3052,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.131475882955925e+19,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}