| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.898550724637681, | |
| "eval_steps": 500, | |
| "global_step": 136, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.057971014492753624, | |
| "grad_norm": 55.32845687866211, | |
| "learning_rate": 0.0001985294117647059, | |
| "loss": 76.4214, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.11594202898550725, | |
| "grad_norm": 59.647117614746094, | |
| "learning_rate": 0.00019558823529411764, | |
| "loss": 66.9992, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.17391304347826086, | |
| "grad_norm": 37.24838638305664, | |
| "learning_rate": 0.00019264705882352944, | |
| "loss": 53.3101, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.2318840579710145, | |
| "grad_norm": 31.579923629760742, | |
| "learning_rate": 0.00018970588235294117, | |
| "loss": 43.1133, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.2898550724637681, | |
| "grad_norm": 33.05708312988281, | |
| "learning_rate": 0.00018676470588235297, | |
| "loss": 41.4356, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.34782608695652173, | |
| "grad_norm": 19.261913299560547, | |
| "learning_rate": 0.0001838235294117647, | |
| "loss": 40.5694, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.4057971014492754, | |
| "grad_norm": 10.231746673583984, | |
| "learning_rate": 0.00018088235294117647, | |
| "loss": 35.4952, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.463768115942029, | |
| "grad_norm": 12.686185836791992, | |
| "learning_rate": 0.00017794117647058823, | |
| "loss": 35.9178, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.5217391304347826, | |
| "grad_norm": 5.934300899505615, | |
| "learning_rate": 0.000175, | |
| "loss": 36.0789, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.5797101449275363, | |
| "grad_norm": 6.5412092208862305, | |
| "learning_rate": 0.0001720588235294118, | |
| "loss": 36.4717, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.6376811594202898, | |
| "grad_norm": 7.47074031829834, | |
| "learning_rate": 0.00016911764705882353, | |
| "loss": 36.1205, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.6956521739130435, | |
| "grad_norm": 7.197345733642578, | |
| "learning_rate": 0.00016617647058823532, | |
| "loss": 35.8026, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.7536231884057971, | |
| "grad_norm": 8.729438781738281, | |
| "learning_rate": 0.00016323529411764706, | |
| "loss": 36.109, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.8115942028985508, | |
| "grad_norm": 10.730180740356445, | |
| "learning_rate": 0.00016029411764705885, | |
| "loss": 34.3386, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.8695652173913043, | |
| "grad_norm": 8.981538772583008, | |
| "learning_rate": 0.0001573529411764706, | |
| "loss": 33.3324, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.927536231884058, | |
| "grad_norm": 13.125741958618164, | |
| "learning_rate": 0.00015441176470588238, | |
| "loss": 34.8168, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.9855072463768116, | |
| "grad_norm": 7.202057361602783, | |
| "learning_rate": 0.00015147058823529412, | |
| "loss": 33.9379, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 1.0289855072463767, | |
| "grad_norm": 7.689969539642334, | |
| "learning_rate": 0.00014852941176470588, | |
| "loss": 26.4506, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 1.0869565217391304, | |
| "grad_norm": 12.119438171386719, | |
| "learning_rate": 0.00014558823529411765, | |
| "loss": 34.1455, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 1.144927536231884, | |
| "grad_norm": 10.431544303894043, | |
| "learning_rate": 0.0001426470588235294, | |
| "loss": 33.6817, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 1.2028985507246377, | |
| "grad_norm": 6.7350311279296875, | |
| "learning_rate": 0.00013970588235294118, | |
| "loss": 34.3042, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 1.2608695652173914, | |
| "grad_norm": 7.63714599609375, | |
| "learning_rate": 0.00013676470588235294, | |
| "loss": 33.6539, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 1.318840579710145, | |
| "grad_norm": 14.94995403289795, | |
| "learning_rate": 0.0001338235294117647, | |
| "loss": 34.8423, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 1.3768115942028984, | |
| "grad_norm": 5.653194904327393, | |
| "learning_rate": 0.00013088235294117647, | |
| "loss": 33.6543, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 1.434782608695652, | |
| "grad_norm": 7.0490946769714355, | |
| "learning_rate": 0.00012794117647058824, | |
| "loss": 34.4564, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.4927536231884058, | |
| "grad_norm": 9.5311279296875, | |
| "learning_rate": 0.000125, | |
| "loss": 34.0708, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 1.5507246376811594, | |
| "grad_norm": 9.313169479370117, | |
| "learning_rate": 0.00012205882352941178, | |
| "loss": 35.0762, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 1.608695652173913, | |
| "grad_norm": 6.120841026306152, | |
| "learning_rate": 0.00011911764705882353, | |
| "loss": 33.5708, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 1.6666666666666665, | |
| "grad_norm": 19.58426856994629, | |
| "learning_rate": 0.00011617647058823531, | |
| "loss": 32.583, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 1.7246376811594204, | |
| "grad_norm": 10.036286354064941, | |
| "learning_rate": 0.00011323529411764706, | |
| "loss": 35.4897, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.7826086956521738, | |
| "grad_norm": 9.211186408996582, | |
| "learning_rate": 0.00011029411764705884, | |
| "loss": 33.5828, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 1.8405797101449275, | |
| "grad_norm": 7.165535926818848, | |
| "learning_rate": 0.00010735294117647059, | |
| "loss": 34.0128, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 1.8985507246376812, | |
| "grad_norm": 8.84890079498291, | |
| "learning_rate": 0.00010441176470588237, | |
| "loss": 34.0061, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 1.9565217391304348, | |
| "grad_norm": 6.122589111328125, | |
| "learning_rate": 0.00010147058823529412, | |
| "loss": 34.1371, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 5.6373186111450195, | |
| "learning_rate": 9.852941176470589e-05, | |
| "loss": 25.4753, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 2.0579710144927534, | |
| "grad_norm": 10.314593315124512, | |
| "learning_rate": 9.558823529411765e-05, | |
| "loss": 33.9195, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 2.1159420289855073, | |
| "grad_norm": 8.193323135375977, | |
| "learning_rate": 9.264705882352942e-05, | |
| "loss": 33.1874, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 2.1739130434782608, | |
| "grad_norm": 9.070418357849121, | |
| "learning_rate": 8.970588235294118e-05, | |
| "loss": 33.3055, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 2.2318840579710146, | |
| "grad_norm": 8.063404083251953, | |
| "learning_rate": 8.676470588235295e-05, | |
| "loss": 34.1831, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 2.289855072463768, | |
| "grad_norm": 7.3460516929626465, | |
| "learning_rate": 8.382352941176471e-05, | |
| "loss": 32.5002, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 2.3478260869565215, | |
| "grad_norm": 7.9106974601745605, | |
| "learning_rate": 8.088235294117648e-05, | |
| "loss": 33.9427, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 2.4057971014492754, | |
| "grad_norm": 6.9293532371521, | |
| "learning_rate": 7.794117647058824e-05, | |
| "loss": 32.7437, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 2.463768115942029, | |
| "grad_norm": 5.583312511444092, | |
| "learning_rate": 7.500000000000001e-05, | |
| "loss": 32.4636, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 2.5217391304347827, | |
| "grad_norm": 9.602357864379883, | |
| "learning_rate": 7.205882352941177e-05, | |
| "loss": 34.4266, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 2.579710144927536, | |
| "grad_norm": 7.29642391204834, | |
| "learning_rate": 6.911764705882354e-05, | |
| "loss": 33.2233, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 2.63768115942029, | |
| "grad_norm": 7.916087627410889, | |
| "learning_rate": 6.61764705882353e-05, | |
| "loss": 32.328, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 2.6956521739130435, | |
| "grad_norm": 9.700789451599121, | |
| "learning_rate": 6.323529411764705e-05, | |
| "loss": 33.3864, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 2.753623188405797, | |
| "grad_norm": 7.925725936889648, | |
| "learning_rate": 6.0294117647058825e-05, | |
| "loss": 33.0956, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 2.8115942028985508, | |
| "grad_norm": 7.136451721191406, | |
| "learning_rate": 5.735294117647059e-05, | |
| "loss": 34.077, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 2.869565217391304, | |
| "grad_norm": 7.057907581329346, | |
| "learning_rate": 5.441176470588235e-05, | |
| "loss": 32.854, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 2.927536231884058, | |
| "grad_norm": 8.482731819152832, | |
| "learning_rate": 5.147058823529411e-05, | |
| "loss": 31.9702, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 2.9855072463768115, | |
| "grad_norm": 8.195663452148438, | |
| "learning_rate": 4.8529411764705885e-05, | |
| "loss": 32.0516, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 3.028985507246377, | |
| "grad_norm": 13.711084365844727, | |
| "learning_rate": 4.558823529411765e-05, | |
| "loss": 23.3778, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 3.0869565217391304, | |
| "grad_norm": 13.546501159667969, | |
| "learning_rate": 4.2647058823529415e-05, | |
| "loss": 32.14, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 3.1449275362318843, | |
| "grad_norm": 11.532342910766602, | |
| "learning_rate": 3.970588235294117e-05, | |
| "loss": 32.6737, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 3.2028985507246377, | |
| "grad_norm": 9.859041213989258, | |
| "learning_rate": 3.6764705882352945e-05, | |
| "loss": 32.2716, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 3.260869565217391, | |
| "grad_norm": 7.993894577026367, | |
| "learning_rate": 3.382352941176471e-05, | |
| "loss": 32.8686, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 3.318840579710145, | |
| "grad_norm": 5.935787677764893, | |
| "learning_rate": 3.0882352941176475e-05, | |
| "loss": 33.3099, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 3.3768115942028984, | |
| "grad_norm": 8.535140037536621, | |
| "learning_rate": 2.7941176470588236e-05, | |
| "loss": 33.1781, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 3.4347826086956523, | |
| "grad_norm": 8.691988945007324, | |
| "learning_rate": 2.5e-05, | |
| "loss": 31.1977, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 3.4927536231884058, | |
| "grad_norm": 7.092562675476074, | |
| "learning_rate": 2.2058823529411766e-05, | |
| "loss": 32.9427, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 3.550724637681159, | |
| "grad_norm": 10.397586822509766, | |
| "learning_rate": 1.9117647058823528e-05, | |
| "loss": 32.8582, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 3.608695652173913, | |
| "grad_norm": 7.606701374053955, | |
| "learning_rate": 1.6176470588235296e-05, | |
| "loss": 33.5666, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 3.6666666666666665, | |
| "grad_norm": 9.118799209594727, | |
| "learning_rate": 1.323529411764706e-05, | |
| "loss": 32.2489, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 3.7246376811594204, | |
| "grad_norm": 8.418676376342773, | |
| "learning_rate": 1.0294117647058824e-05, | |
| "loss": 32.8995, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 3.782608695652174, | |
| "grad_norm": 7.379878044128418, | |
| "learning_rate": 7.3529411764705884e-06, | |
| "loss": 33.5503, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 3.8405797101449277, | |
| "grad_norm": 7.183637619018555, | |
| "learning_rate": 4.411764705882353e-06, | |
| "loss": 31.911, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 3.898550724637681, | |
| "grad_norm": 12.431236267089844, | |
| "learning_rate": 1.4705882352941177e-06, | |
| "loss": 31.6491, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 3.898550724637681, | |
| "step": 136, | |
| "total_flos": 29261690983296.0, | |
| "train_loss": 35.055379671208996, | |
| "train_runtime": 354.4991, | |
| "train_samples_per_second": 6.229, | |
| "train_steps_per_second": 0.384 | |
| } | |
| ], | |
| "logging_steps": 2, | |
| "max_steps": 136, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 4, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 29261690983296.0, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |