| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.025437201907790145, | |
| "eval_steps": 50, | |
| "global_step": 200, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0001271860095389507, | |
| "grad_norm": 1.9835193157196045, | |
| "learning_rate": 2e-05, | |
| "loss": 5.4485, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0001271860095389507, | |
| "eval_loss": 1.2489664554595947, | |
| "eval_runtime": 60.2564, | |
| "eval_samples_per_second": 54.949, | |
| "eval_steps_per_second": 27.483, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0002543720190779014, | |
| "grad_norm": 1.839722990989685, | |
| "learning_rate": 4e-05, | |
| "loss": 5.0062, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.00038155802861685216, | |
| "grad_norm": 1.906867265701294, | |
| "learning_rate": 6e-05, | |
| "loss": 5.0884, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.0005087440381558028, | |
| "grad_norm": 1.7935138940811157, | |
| "learning_rate": 8e-05, | |
| "loss": 4.9646, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.0006359300476947536, | |
| "grad_norm": 2.1912240982055664, | |
| "learning_rate": 0.0001, | |
| "loss": 5.574, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0007631160572337043, | |
| "grad_norm": 1.704566478729248, | |
| "learning_rate": 0.00012, | |
| "loss": 4.7769, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.000890302066772655, | |
| "grad_norm": 1.8012601137161255, | |
| "learning_rate": 0.00014, | |
| "loss": 4.937, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.0010174880763116057, | |
| "grad_norm": 1.76153564453125, | |
| "learning_rate": 0.00016, | |
| "loss": 4.7734, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.0011446740858505565, | |
| "grad_norm": 1.877334713935852, | |
| "learning_rate": 0.00018, | |
| "loss": 4.9324, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.0012718600953895071, | |
| "grad_norm": 1.7994229793548584, | |
| "learning_rate": 0.0002, | |
| "loss": 4.6046, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0013990461049284578, | |
| "grad_norm": 1.9988417625427246, | |
| "learning_rate": 0.0001999863304992469, | |
| "loss": 4.3125, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.0015262321144674086, | |
| "grad_norm": 2.3464603424072266, | |
| "learning_rate": 0.00019994532573409262, | |
| "loss": 4.363, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.0016534181240063592, | |
| "grad_norm": 2.255720615386963, | |
| "learning_rate": 0.00019987699691483048, | |
| "loss": 4.1383, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.00178060413354531, | |
| "grad_norm": 2.2891433238983154, | |
| "learning_rate": 0.00019978136272187747, | |
| "loss": 4.1131, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.0019077901430842607, | |
| "grad_norm": 2.1453323364257812, | |
| "learning_rate": 0.000199658449300667, | |
| "loss": 4.1299, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.0020349761526232114, | |
| "grad_norm": 1.9900203943252563, | |
| "learning_rate": 0.00019950829025450114, | |
| "loss": 4.0106, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.002162162162162162, | |
| "grad_norm": 1.8799800872802734, | |
| "learning_rate": 0.00019933092663536382, | |
| "loss": 4.0144, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.002289348171701113, | |
| "grad_norm": 2.0695888996124268, | |
| "learning_rate": 0.00019912640693269752, | |
| "loss": 3.7752, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.0024165341812400635, | |
| "grad_norm": 2.120725631713867, | |
| "learning_rate": 0.00019889478706014687, | |
| "loss": 4.0842, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.0025437201907790143, | |
| "grad_norm": 2.8703718185424805, | |
| "learning_rate": 0.00019863613034027224, | |
| "loss": 3.6905, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.002670906200317965, | |
| "grad_norm": 1.7893959283828735, | |
| "learning_rate": 0.00019835050748723824, | |
| "loss": 3.5627, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.0027980922098569156, | |
| "grad_norm": 1.9003410339355469, | |
| "learning_rate": 0.00019803799658748094, | |
| "loss": 3.6049, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.0029252782193958664, | |
| "grad_norm": 2.649674892425537, | |
| "learning_rate": 0.00019769868307835994, | |
| "loss": 3.7806, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.0030524642289348172, | |
| "grad_norm": 1.8168840408325195, | |
| "learning_rate": 0.0001973326597248006, | |
| "loss": 3.4188, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.003179650238473768, | |
| "grad_norm": 1.8296242952346802, | |
| "learning_rate": 0.00019694002659393305, | |
| "loss": 3.6451, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.0033068362480127185, | |
| "grad_norm": 1.8689442873001099, | |
| "learning_rate": 0.00019652089102773488, | |
| "loss": 3.5423, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.0034340222575516693, | |
| "grad_norm": 1.7191494703292847, | |
| "learning_rate": 0.00019607536761368484, | |
| "loss": 3.5463, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.00356120826709062, | |
| "grad_norm": 1.8769090175628662, | |
| "learning_rate": 0.00019560357815343577, | |
| "loss": 3.6548, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.0036883942766295706, | |
| "grad_norm": 1.6859794855117798, | |
| "learning_rate": 0.00019510565162951537, | |
| "loss": 3.3713, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.0038155802861685214, | |
| "grad_norm": 1.5716698169708252, | |
| "learning_rate": 0.00019458172417006347, | |
| "loss": 3.2842, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.003942766295707472, | |
| "grad_norm": 1.8034677505493164, | |
| "learning_rate": 0.00019403193901161613, | |
| "loss": 3.533, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.004069952305246423, | |
| "grad_norm": 1.8093938827514648, | |
| "learning_rate": 0.0001934564464599461, | |
| "loss": 3.408, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.0041971383147853735, | |
| "grad_norm": 1.7909035682678223, | |
| "learning_rate": 0.00019285540384897073, | |
| "loss": 3.2401, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.004324324324324324, | |
| "grad_norm": 1.7447187900543213, | |
| "learning_rate": 0.00019222897549773848, | |
| "loss": 3.2696, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.004451510333863275, | |
| "grad_norm": 1.5972126722335815, | |
| "learning_rate": 0.00019157733266550575, | |
| "loss": 3.0369, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.004578696343402226, | |
| "grad_norm": 1.729097843170166, | |
| "learning_rate": 0.00019090065350491626, | |
| "loss": 3.2384, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.004705882352941176, | |
| "grad_norm": 1.8049160242080688, | |
| "learning_rate": 0.00019019912301329592, | |
| "loss": 3.2673, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.004833068362480127, | |
| "grad_norm": 1.684900164604187, | |
| "learning_rate": 0.00018947293298207635, | |
| "loss": 3.3017, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.004960254372019078, | |
| "grad_norm": 2.184971332550049, | |
| "learning_rate": 0.0001887222819443612, | |
| "loss": 2.946, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.005087440381558029, | |
| "grad_norm": 1.815535306930542, | |
| "learning_rate": 0.0001879473751206489, | |
| "loss": 3.0601, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.0052146263910969794, | |
| "grad_norm": 1.8295841217041016, | |
| "learning_rate": 0.00018714842436272773, | |
| "loss": 3.2226, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.00534181240063593, | |
| "grad_norm": 1.9899147748947144, | |
| "learning_rate": 0.00018632564809575742, | |
| "loss": 3.225, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.005468998410174881, | |
| "grad_norm": 2.021754026412964, | |
| "learning_rate": 0.0001854792712585539, | |
| "loss": 3.0691, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.005596184419713831, | |
| "grad_norm": 1.8898154497146606, | |
| "learning_rate": 0.00018460952524209355, | |
| "loss": 3.2772, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.005723370429252782, | |
| "grad_norm": 2.083461046218872, | |
| "learning_rate": 0.00018371664782625287, | |
| "loss": 3.4458, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.005850556438791733, | |
| "grad_norm": 2.219532012939453, | |
| "learning_rate": 0.00018280088311480201, | |
| "loss": 3.3751, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.005977742448330684, | |
| "grad_norm": 2.285486936569214, | |
| "learning_rate": 0.00018186248146866927, | |
| "loss": 3.3656, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.0061049284578696345, | |
| "grad_norm": 2.1301989555358887, | |
| "learning_rate": 0.00018090169943749476, | |
| "loss": 2.9066, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.006232114467408585, | |
| "grad_norm": 2.079562187194824, | |
| "learning_rate": 0.0001799187996894925, | |
| "loss": 3.0637, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.006359300476947536, | |
| "grad_norm": 2.057091236114502, | |
| "learning_rate": 0.00017891405093963938, | |
| "loss": 3.1446, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.006359300476947536, | |
| "eval_loss": 0.7800452709197998, | |
| "eval_runtime": 57.7119, | |
| "eval_samples_per_second": 57.371, | |
| "eval_steps_per_second": 28.694, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.006486486486486486, | |
| "grad_norm": 1.9653096199035645, | |
| "learning_rate": 0.00017788772787621126, | |
| "loss": 3.393, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.006613672496025437, | |
| "grad_norm": 1.9508553743362427, | |
| "learning_rate": 0.00017684011108568592, | |
| "loss": 3.1593, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.006740858505564388, | |
| "grad_norm": 1.9427250623703003, | |
| "learning_rate": 0.0001757714869760335, | |
| "loss": 3.3619, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.006868044515103339, | |
| "grad_norm": 1.902197241783142, | |
| "learning_rate": 0.0001746821476984154, | |
| "loss": 3.1835, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.0069952305246422895, | |
| "grad_norm": 2.077125310897827, | |
| "learning_rate": 0.00017357239106731317, | |
| "loss": 3.4491, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.00712241653418124, | |
| "grad_norm": 1.8819434642791748, | |
| "learning_rate": 0.00017244252047910892, | |
| "loss": 3.0944, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.00724960254372019, | |
| "grad_norm": 2.3512206077575684, | |
| "learning_rate": 0.00017129284482913972, | |
| "loss": 3.1957, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.007376788553259141, | |
| "grad_norm": 2.125260829925537, | |
| "learning_rate": 0.00017012367842724887, | |
| "loss": 2.8759, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.007503974562798092, | |
| "grad_norm": 2.0126636028289795, | |
| "learning_rate": 0.0001689353409118566, | |
| "loss": 3.2306, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.007631160572337043, | |
| "grad_norm": 2.0234174728393555, | |
| "learning_rate": 0.00016772815716257412, | |
| "loss": 3.1808, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.007758346581875994, | |
| "grad_norm": 2.0953781604766846, | |
| "learning_rate": 0.0001665024572113848, | |
| "loss": 3.0547, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.007885532591414944, | |
| "grad_norm": 2.0005664825439453, | |
| "learning_rate": 0.00016525857615241687, | |
| "loss": 3.1831, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.008012718600953895, | |
| "grad_norm": 2.023897886276245, | |
| "learning_rate": 0.00016399685405033167, | |
| "loss": 3.0127, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.008139904610492845, | |
| "grad_norm": 2.0961215496063232, | |
| "learning_rate": 0.0001627176358473537, | |
| "loss": 3.3373, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.008267090620031796, | |
| "grad_norm": 2.068119525909424, | |
| "learning_rate": 0.0001614212712689668, | |
| "loss": 3.2092, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.008394276629570747, | |
| "grad_norm": 2.0452160835266113, | |
| "learning_rate": 0.00016010811472830252, | |
| "loss": 3.2085, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.008521462639109698, | |
| "grad_norm": 2.156864643096924, | |
| "learning_rate": 0.00015877852522924732, | |
| "loss": 3.0868, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.008648648648648649, | |
| "grad_norm": 2.029207706451416, | |
| "learning_rate": 0.00015743286626829437, | |
| "loss": 2.9679, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.0087758346581876, | |
| "grad_norm": 2.12054181098938, | |
| "learning_rate": 0.0001560715057351673, | |
| "loss": 3.1446, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.00890302066772655, | |
| "grad_norm": 2.1028153896331787, | |
| "learning_rate": 0.00015469481581224272, | |
| "loss": 3.0386, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.009030206677265501, | |
| "grad_norm": 2.2078723907470703, | |
| "learning_rate": 0.0001533031728727994, | |
| "loss": 3.2846, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.009157392686804452, | |
| "grad_norm": 2.080521821975708, | |
| "learning_rate": 0.00015189695737812152, | |
| "loss": 3.046, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.009284578696343403, | |
| "grad_norm": 2.0975699424743652, | |
| "learning_rate": 0.0001504765537734844, | |
| "loss": 2.8228, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.009411764705882352, | |
| "grad_norm": 2.443129062652588, | |
| "learning_rate": 0.00014904235038305083, | |
| "loss": 3.1119, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.009538950715421303, | |
| "grad_norm": 2.1817243099212646, | |
| "learning_rate": 0.00014759473930370736, | |
| "loss": 2.6607, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.009666136724960254, | |
| "grad_norm": 2.2016985416412354, | |
| "learning_rate": 0.0001461341162978688, | |
| "loss": 2.9715, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.009793322734499205, | |
| "grad_norm": 2.2988955974578857, | |
| "learning_rate": 0.00014466088068528068, | |
| "loss": 2.962, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.009920508744038155, | |
| "grad_norm": 2.2614760398864746, | |
| "learning_rate": 0.00014317543523384928, | |
| "loss": 3.1009, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.010047694753577106, | |
| "grad_norm": 2.051297664642334, | |
| "learning_rate": 0.00014167818604952906, | |
| "loss": 3.0087, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.010174880763116057, | |
| "grad_norm": 2.0947837829589844, | |
| "learning_rate": 0.00014016954246529696, | |
| "loss": 3.0739, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.010302066772655008, | |
| "grad_norm": 2.4646780490875244, | |
| "learning_rate": 0.00013864991692924523, | |
| "loss": 3.2066, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.010429252782193959, | |
| "grad_norm": 2.156864881515503, | |
| "learning_rate": 0.00013711972489182208, | |
| "loss": 3.2273, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.01055643879173291, | |
| "grad_norm": 2.163877010345459, | |
| "learning_rate": 0.00013557938469225167, | |
| "loss": 3.2036, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.01068362480127186, | |
| "grad_norm": 2.1124367713928223, | |
| "learning_rate": 0.00013402931744416433, | |
| "loss": 2.9042, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.010810810810810811, | |
| "grad_norm": 2.2118308544158936, | |
| "learning_rate": 0.00013246994692046836, | |
| "loss": 3.2564, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.010937996820349762, | |
| "grad_norm": 2.289252758026123, | |
| "learning_rate": 0.00013090169943749476, | |
| "loss": 3.05, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.011065182829888713, | |
| "grad_norm": 2.044328451156616, | |
| "learning_rate": 0.0001293250037384465, | |
| "loss": 2.8559, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.011192368839427662, | |
| "grad_norm": 2.127413511276245, | |
| "learning_rate": 0.00012774029087618446, | |
| "loss": 3.0749, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.011319554848966613, | |
| "grad_norm": 2.1147000789642334, | |
| "learning_rate": 0.00012614799409538198, | |
| "loss": 3.065, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.011446740858505564, | |
| "grad_norm": 2.2932112216949463, | |
| "learning_rate": 0.00012454854871407994, | |
| "loss": 3.6366, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.011573926868044515, | |
| "grad_norm": 2.405611753463745, | |
| "learning_rate": 0.00012294239200467516, | |
| "loss": 3.1775, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.011701112877583466, | |
| "grad_norm": 2.0750105381011963, | |
| "learning_rate": 0.0001213299630743747, | |
| "loss": 2.9886, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.011828298887122416, | |
| "grad_norm": 2.072420358657837, | |
| "learning_rate": 0.00011971170274514802, | |
| "loss": 3.0914, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.011955484896661367, | |
| "grad_norm": 2.19963002204895, | |
| "learning_rate": 0.000118088053433211, | |
| "loss": 2.9166, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.012082670906200318, | |
| "grad_norm": 2.1508893966674805, | |
| "learning_rate": 0.00011645945902807341, | |
| "loss": 2.8944, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.012209856915739269, | |
| "grad_norm": 2.4836275577545166, | |
| "learning_rate": 0.0001148263647711842, | |
| "loss": 2.9622, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.01233704292527822, | |
| "grad_norm": 2.2222437858581543, | |
| "learning_rate": 0.00011318921713420691, | |
| "loss": 2.8688, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.01246422893481717, | |
| "grad_norm": 2.1132421493530273, | |
| "learning_rate": 0.00011154846369695863, | |
| "loss": 2.8259, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.012591414944356122, | |
| "grad_norm": 2.1007161140441895, | |
| "learning_rate": 0.0001099045530250463, | |
| "loss": 2.7156, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.012718600953895072, | |
| "grad_norm": 2.1596603393554688, | |
| "learning_rate": 0.00010825793454723325, | |
| "loss": 3.0597, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.012718600953895072, | |
| "eval_loss": 0.7423936724662781, | |
| "eval_runtime": 57.8996, | |
| "eval_samples_per_second": 57.185, | |
| "eval_steps_per_second": 28.601, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.012845786963434021, | |
| "grad_norm": 2.3275163173675537, | |
| "learning_rate": 0.00010660905843256994, | |
| "loss": 3.1327, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.012972972972972972, | |
| "grad_norm": 2.0324950218200684, | |
| "learning_rate": 0.00010495837546732224, | |
| "loss": 2.7223, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.013100158982511923, | |
| "grad_norm": 2.420083999633789, | |
| "learning_rate": 0.00010330633693173082, | |
| "loss": 3.1192, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.013227344992050874, | |
| "grad_norm": 2.116345167160034, | |
| "learning_rate": 0.00010165339447663587, | |
| "loss": 3.1318, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.013354531001589825, | |
| "grad_norm": 2.523977756500244, | |
| "learning_rate": 0.0001, | |
| "loss": 2.8798, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.013481717011128776, | |
| "grad_norm": 2.141871213912964, | |
| "learning_rate": 9.834660552336415e-05, | |
| "loss": 3.1058, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.013608903020667727, | |
| "grad_norm": 2.2081823348999023, | |
| "learning_rate": 9.669366306826919e-05, | |
| "loss": 2.9769, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.013736089030206677, | |
| "grad_norm": 2.0940332412719727, | |
| "learning_rate": 9.504162453267777e-05, | |
| "loss": 2.9284, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.013863275039745628, | |
| "grad_norm": 2.2576355934143066, | |
| "learning_rate": 9.339094156743007e-05, | |
| "loss": 3.0978, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.013990461049284579, | |
| "grad_norm": 2.107511043548584, | |
| "learning_rate": 9.174206545276677e-05, | |
| "loss": 2.8895, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.01411764705882353, | |
| "grad_norm": 2.289916515350342, | |
| "learning_rate": 9.009544697495374e-05, | |
| "loss": 3.1167, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.01424483306836248, | |
| "grad_norm": 2.1832778453826904, | |
| "learning_rate": 8.845153630304139e-05, | |
| "loss": 2.8849, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.014372019077901432, | |
| "grad_norm": 2.0638606548309326, | |
| "learning_rate": 8.681078286579311e-05, | |
| "loss": 2.8642, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.01449920508744038, | |
| "grad_norm": 2.2622199058532715, | |
| "learning_rate": 8.517363522881579e-05, | |
| "loss": 3.1066, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.014626391096979332, | |
| "grad_norm": 2.0616939067840576, | |
| "learning_rate": 8.35405409719266e-05, | |
| "loss": 2.9251, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.014753577106518282, | |
| "grad_norm": 2.2342519760131836, | |
| "learning_rate": 8.191194656678904e-05, | |
| "loss": 3.1882, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.014880763116057233, | |
| "grad_norm": 2.0429413318634033, | |
| "learning_rate": 8.028829725485199e-05, | |
| "loss": 3.0048, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.015007949125596184, | |
| "grad_norm": 2.2607035636901855, | |
| "learning_rate": 7.867003692562534e-05, | |
| "loss": 3.0522, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.015135135135135135, | |
| "grad_norm": 2.1233274936676025, | |
| "learning_rate": 7.705760799532485e-05, | |
| "loss": 2.952, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.015262321144674086, | |
| "grad_norm": 2.23429274559021, | |
| "learning_rate": 7.54514512859201e-05, | |
| "loss": 2.9693, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.015389507154213037, | |
| "grad_norm": 2.315385103225708, | |
| "learning_rate": 7.385200590461803e-05, | |
| "loss": 3.2038, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.015516693163751987, | |
| "grad_norm": 2.2843518257141113, | |
| "learning_rate": 7.225970912381556e-05, | |
| "loss": 2.7576, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.015643879173290937, | |
| "grad_norm": 2.316584825515747, | |
| "learning_rate": 7.067499626155354e-05, | |
| "loss": 3.1409, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.015771065182829887, | |
| "grad_norm": 2.1097593307495117, | |
| "learning_rate": 6.909830056250527e-05, | |
| "loss": 3.0405, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.01589825119236884, | |
| "grad_norm": 2.2882628440856934, | |
| "learning_rate": 6.753005307953167e-05, | |
| "loss": 2.9586, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.01602543720190779, | |
| "grad_norm": 2.2434895038604736, | |
| "learning_rate": 6.59706825558357e-05, | |
| "loss": 2.8796, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.01615262321144674, | |
| "grad_norm": 2.2158470153808594, | |
| "learning_rate": 6.442061530774834e-05, | |
| "loss": 3.1726, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.01627980922098569, | |
| "grad_norm": 2.311504364013672, | |
| "learning_rate": 6.28802751081779e-05, | |
| "loss": 3.2291, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.01640699523052464, | |
| "grad_norm": 2.22208833694458, | |
| "learning_rate": 6.135008307075481e-05, | |
| "loss": 3.2121, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.016534181240063592, | |
| "grad_norm": 2.165191650390625, | |
| "learning_rate": 5.983045753470308e-05, | |
| "loss": 3.103, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.016661367249602543, | |
| "grad_norm": 2.0541014671325684, | |
| "learning_rate": 5.832181395047098e-05, | |
| "loss": 2.9611, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.016788553259141494, | |
| "grad_norm": 2.2373898029327393, | |
| "learning_rate": 5.6824564766150726e-05, | |
| "loss": 2.7329, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.016915739268680445, | |
| "grad_norm": 2.2604360580444336, | |
| "learning_rate": 5.533911931471936e-05, | |
| "loss": 2.983, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.017042925278219396, | |
| "grad_norm": 2.3402767181396484, | |
| "learning_rate": 5.386588370213124e-05, | |
| "loss": 2.6614, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.017170111287758347, | |
| "grad_norm": 2.1931848526000977, | |
| "learning_rate": 5.240526069629265e-05, | |
| "loss": 3.2408, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.017297297297297298, | |
| "grad_norm": 2.2224552631378174, | |
| "learning_rate": 5.095764961694922e-05, | |
| "loss": 2.8428, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.01742448330683625, | |
| "grad_norm": 2.2872660160064697, | |
| "learning_rate": 4.952344622651566e-05, | |
| "loss": 2.928, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.0175516693163752, | |
| "grad_norm": 2.107144832611084, | |
| "learning_rate": 4.810304262187852e-05, | |
| "loss": 3.1571, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.01767885532591415, | |
| "grad_norm": 2.1836090087890625, | |
| "learning_rate": 4.669682712720065e-05, | |
| "loss": 2.9863, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.0178060413354531, | |
| "grad_norm": 2.202346086502075, | |
| "learning_rate": 4.530518418775733e-05, | |
| "loss": 3.0278, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.017933227344992052, | |
| "grad_norm": 2.1937851905822754, | |
| "learning_rate": 4.392849426483274e-05, | |
| "loss": 2.8878, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.018060413354531003, | |
| "grad_norm": 2.231384038925171, | |
| "learning_rate": 4.256713373170564e-05, | |
| "loss": 3.1869, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.018187599364069953, | |
| "grad_norm": 2.143895387649536, | |
| "learning_rate": 4.12214747707527e-05, | |
| "loss": 2.7143, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.018314785373608904, | |
| "grad_norm": 2.0842132568359375, | |
| "learning_rate": 3.9891885271697496e-05, | |
| "loss": 3.0679, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.018441971383147855, | |
| "grad_norm": 2.206463098526001, | |
| "learning_rate": 3.857872873103322e-05, | |
| "loss": 3.122, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.018569157392686806, | |
| "grad_norm": 2.248577833175659, | |
| "learning_rate": 3.7282364152646297e-05, | |
| "loss": 2.8049, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.018696343402225757, | |
| "grad_norm": 2.219324827194214, | |
| "learning_rate": 3.600314594966834e-05, | |
| "loss": 3.1196, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.018823529411764704, | |
| "grad_norm": 2.359001398086548, | |
| "learning_rate": 3.4741423847583134e-05, | |
| "loss": 3.1775, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.018950715421303655, | |
| "grad_norm": 2.1845834255218506, | |
| "learning_rate": 3.349754278861517e-05, | |
| "loss": 3.0865, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.019077901430842606, | |
| "grad_norm": 2.240842580795288, | |
| "learning_rate": 3.227184283742591e-05, | |
| "loss": 2.7162, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.019077901430842606, | |
| "eval_loss": 0.7287956476211548, | |
| "eval_runtime": 57.7804, | |
| "eval_samples_per_second": 57.303, | |
| "eval_steps_per_second": 28.66, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.019205087440381557, | |
| "grad_norm": 2.1659834384918213, | |
| "learning_rate": 3.106465908814342e-05, | |
| "loss": 3.0083, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.019332273449920508, | |
| "grad_norm": 2.5431900024414062, | |
| "learning_rate": 2.9876321572751144e-05, | |
| "loss": 2.8827, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.01945945945945946, | |
| "grad_norm": 2.1179680824279785, | |
| "learning_rate": 2.87071551708603e-05, | |
| "loss": 2.9347, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.01958664546899841, | |
| "grad_norm": 2.2814908027648926, | |
| "learning_rate": 2.7557479520891104e-05, | |
| "loss": 2.939, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.01971383147853736, | |
| "grad_norm": 2.2081267833709717, | |
| "learning_rate": 2.6427608932686843e-05, | |
| "loss": 3.1783, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.01984101748807631, | |
| "grad_norm": 2.1636998653411865, | |
| "learning_rate": 2.5317852301584643e-05, | |
| "loss": 2.8015, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.019968203497615262, | |
| "grad_norm": 2.2725632190704346, | |
| "learning_rate": 2.422851302396655e-05, | |
| "loss": 3.0367, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.020095389507154213, | |
| "grad_norm": 2.1158969402313232, | |
| "learning_rate": 2.315988891431412e-05, | |
| "loss": 2.9014, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.020222575516693164, | |
| "grad_norm": 2.225064277648926, | |
| "learning_rate": 2.2112272123788768e-05, | |
| "loss": 3.0376, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.020349761526232114, | |
| "grad_norm": 2.264939308166504, | |
| "learning_rate": 2.1085949060360654e-05, | |
| "loss": 2.7356, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.020476947535771065, | |
| "grad_norm": 2.242896556854248, | |
| "learning_rate": 2.008120031050753e-05, | |
| "loss": 2.7887, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.020604133545310016, | |
| "grad_norm": 2.4548873901367188, | |
| "learning_rate": 1.9098300562505266e-05, | |
| "loss": 2.9714, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.020731319554848967, | |
| "grad_norm": 2.1999237537384033, | |
| "learning_rate": 1.8137518531330767e-05, | |
| "loss": 2.9521, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.020858505564387918, | |
| "grad_norm": 2.446458101272583, | |
| "learning_rate": 1.7199116885197995e-05, | |
| "loss": 3.272, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.02098569157392687, | |
| "grad_norm": 2.216688394546509, | |
| "learning_rate": 1.6283352173747145e-05, | |
| "loss": 2.9213, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.02111287758346582, | |
| "grad_norm": 2.1515614986419678, | |
| "learning_rate": 1.5390474757906446e-05, | |
| "loss": 2.9302, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.02124006359300477, | |
| "grad_norm": 2.1468420028686523, | |
| "learning_rate": 1.4520728741446089e-05, | |
| "loss": 2.9918, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.02136724960254372, | |
| "grad_norm": 2.5838160514831543, | |
| "learning_rate": 1.3674351904242611e-05, | |
| "loss": 2.8896, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.021494435612082672, | |
| "grad_norm": 2.196659564971924, | |
| "learning_rate": 1.2851575637272262e-05, | |
| "loss": 2.9632, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.021621621621621623, | |
| "grad_norm": 2.193552017211914, | |
| "learning_rate": 1.2052624879351104e-05, | |
| "loss": 2.901, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.021748807631160574, | |
| "grad_norm": 2.2525877952575684, | |
| "learning_rate": 1.1277718055638819e-05, | |
| "loss": 3.1768, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.021875993640699525, | |
| "grad_norm": 2.2571728229522705, | |
| "learning_rate": 1.0527067017923654e-05, | |
| "loss": 3.2308, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.022003179650238475, | |
| "grad_norm": 2.3145599365234375, | |
| "learning_rate": 9.80087698670411e-06, | |
| "loss": 3.1365, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.022130365659777426, | |
| "grad_norm": 2.1300814151763916, | |
| "learning_rate": 9.09934649508375e-06, | |
| "loss": 2.7955, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.022257551669316374, | |
| "grad_norm": 2.23616361618042, | |
| "learning_rate": 8.422667334494249e-06, | |
| "loss": 2.9142, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.022384737678855324, | |
| "grad_norm": 2.1460986137390137, | |
| "learning_rate": 7.771024502261526e-06, | |
| "loss": 2.7426, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.022511923688394275, | |
| "grad_norm": 2.2808210849761963, | |
| "learning_rate": 7.144596151029303e-06, | |
| "loss": 2.9775, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.022639109697933226, | |
| "grad_norm": 2.477661609649658, | |
| "learning_rate": 6.543553540053926e-06, | |
| "loss": 3.4063, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.022766295707472177, | |
| "grad_norm": 2.2510106563568115, | |
| "learning_rate": 5.968060988383883e-06, | |
| "loss": 3.0301, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.022893481717011128, | |
| "grad_norm": 2.361362934112549, | |
| "learning_rate": 5.418275829936537e-06, | |
| "loss": 3.1077, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.02302066772655008, | |
| "grad_norm": 2.293454170227051, | |
| "learning_rate": 4.8943483704846475e-06, | |
| "loss": 3.3876, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.02314785373608903, | |
| "grad_norm": 2.142216682434082, | |
| "learning_rate": 4.3964218465642355e-06, | |
| "loss": 2.8311, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.02327503974562798, | |
| "grad_norm": 2.2283873558044434, | |
| "learning_rate": 3.924632386315186e-06, | |
| "loss": 3.2105, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.02340222575516693, | |
| "grad_norm": 2.265622138977051, | |
| "learning_rate": 3.4791089722651436e-06, | |
| "loss": 2.9194, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.023529411764705882, | |
| "grad_norm": 2.256969690322876, | |
| "learning_rate": 3.059973406066963e-06, | |
| "loss": 3.1245, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.023656597774244833, | |
| "grad_norm": 2.216651439666748, | |
| "learning_rate": 2.667340275199426e-06, | |
| "loss": 3.1906, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.023783783783783784, | |
| "grad_norm": 2.2530014514923096, | |
| "learning_rate": 2.3013169216400733e-06, | |
| "loss": 2.8496, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.023910969793322735, | |
| "grad_norm": 2.3641066551208496, | |
| "learning_rate": 1.9620034125190644e-06, | |
| "loss": 3.0763, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.024038155802861685, | |
| "grad_norm": 2.3235628604888916, | |
| "learning_rate": 1.6494925127617634e-06, | |
| "loss": 2.8297, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.024165341812400636, | |
| "grad_norm": 2.317734956741333, | |
| "learning_rate": 1.3638696597277679e-06, | |
| "loss": 2.9774, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.024292527821939587, | |
| "grad_norm": 2.316967725753784, | |
| "learning_rate": 1.1052129398531507e-06, | |
| "loss": 2.8554, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.024419713831478538, | |
| "grad_norm": 2.4230763912200928, | |
| "learning_rate": 8.735930673024806e-07, | |
| "loss": 3.3244, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.02454689984101749, | |
| "grad_norm": 2.1771557331085205, | |
| "learning_rate": 6.690733646361857e-07, | |
| "loss": 2.8446, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.02467408585055644, | |
| "grad_norm": 2.3441531658172607, | |
| "learning_rate": 4.917097454988584e-07, | |
| "loss": 2.7098, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.02480127186009539, | |
| "grad_norm": 2.175175905227661, | |
| "learning_rate": 3.415506993330153e-07, | |
| "loss": 2.7541, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.02492845786963434, | |
| "grad_norm": 2.375668525695801, | |
| "learning_rate": 2.1863727812254653e-07, | |
| "loss": 3.0021, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.025055643879173292, | |
| "grad_norm": 2.134728193283081, | |
| "learning_rate": 1.230030851695263e-07, | |
| "loss": 3.1438, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.025182829888712243, | |
| "grad_norm": 2.2377030849456787, | |
| "learning_rate": 5.467426590739511e-08, | |
| "loss": 2.9125, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.025310015898251194, | |
| "grad_norm": 2.040388584136963, | |
| "learning_rate": 1.3669500753099585e-08, | |
| "loss": 3.0197, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.025437201907790145, | |
| "grad_norm": 2.3986241817474365, | |
| "learning_rate": 0.0, | |
| "loss": 2.9025, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.025437201907790145, | |
| "eval_loss": 0.7257097959518433, | |
| "eval_runtime": 57.7187, | |
| "eval_samples_per_second": 57.364, | |
| "eval_steps_per_second": 28.691, | |
| "step": 200 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 200, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 50, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.1943156564099072e+16, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |