Zineddinetranslate's picture
Upload folder using huggingface_hub
a2918ed verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 742,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.013490725126475547,
"grad_norm": 65.45332336425781,
"learning_rate": 0.00019964061096136568,
"loss": 22.1843,
"step": 10
},
{
"epoch": 0.026981450252951095,
"grad_norm": 6.102770805358887,
"learning_rate": 0.00019874213836477988,
"loss": 7.6527,
"step": 20
},
{
"epoch": 0.04047217537942664,
"grad_norm": 2.982379913330078,
"learning_rate": 0.00019784366576819408,
"loss": 3.4532,
"step": 30
},
{
"epoch": 0.05396290050590219,
"grad_norm": 3.4628562927246094,
"learning_rate": 0.00019694519317160826,
"loss": 2.4467,
"step": 40
},
{
"epoch": 0.06745362563237774,
"grad_norm": 3.0008349418640137,
"learning_rate": 0.00019604672057502246,
"loss": 1.6481,
"step": 50
},
{
"epoch": 0.08094435075885328,
"grad_norm": 1.767865777015686,
"learning_rate": 0.00019514824797843666,
"loss": 1.2119,
"step": 60
},
{
"epoch": 0.09443507588532883,
"grad_norm": 3.416968584060669,
"learning_rate": 0.00019424977538185087,
"loss": 0.967,
"step": 70
},
{
"epoch": 0.10792580101180438,
"grad_norm": 3.718262195587158,
"learning_rate": 0.00019335130278526507,
"loss": 0.8783,
"step": 80
},
{
"epoch": 0.12141652613827993,
"grad_norm": 3.689655065536499,
"learning_rate": 0.00019245283018867927,
"loss": 0.7619,
"step": 90
},
{
"epoch": 0.13490725126475547,
"grad_norm": 6.728384017944336,
"learning_rate": 0.00019155435759209347,
"loss": 0.7244,
"step": 100
},
{
"epoch": 0.14839797639123103,
"grad_norm": 4.880130290985107,
"learning_rate": 0.00019065588499550765,
"loss": 0.6931,
"step": 110
},
{
"epoch": 0.16188870151770657,
"grad_norm": 10.14767074584961,
"learning_rate": 0.00018975741239892185,
"loss": 0.6995,
"step": 120
},
{
"epoch": 0.17537942664418213,
"grad_norm": 10.897403717041016,
"learning_rate": 0.00018885893980233605,
"loss": 0.69,
"step": 130
},
{
"epoch": 0.18887015177065766,
"grad_norm": 7.759915351867676,
"learning_rate": 0.00018796046720575023,
"loss": 0.6899,
"step": 140
},
{
"epoch": 0.20236087689713322,
"grad_norm": 7.006256580352783,
"learning_rate": 0.00018706199460916443,
"loss": 0.6739,
"step": 150
},
{
"epoch": 0.21585160202360876,
"grad_norm": 5.777213096618652,
"learning_rate": 0.00018616352201257863,
"loss": 0.6731,
"step": 160
},
{
"epoch": 0.22934232715008432,
"grad_norm": 10.224111557006836,
"learning_rate": 0.00018526504941599283,
"loss": 0.6417,
"step": 170
},
{
"epoch": 0.24283305227655985,
"grad_norm": 6.332294464111328,
"learning_rate": 0.000184366576819407,
"loss": 0.6452,
"step": 180
},
{
"epoch": 0.2563237774030354,
"grad_norm": 3.9472005367279053,
"learning_rate": 0.0001834681042228212,
"loss": 0.6271,
"step": 190
},
{
"epoch": 0.26981450252951095,
"grad_norm": 6.640835762023926,
"learning_rate": 0.0001825696316262354,
"loss": 0.6161,
"step": 200
},
{
"epoch": 0.28330522765598654,
"grad_norm": 5.6504316329956055,
"learning_rate": 0.00018167115902964959,
"loss": 0.6452,
"step": 210
},
{
"epoch": 0.29679595278246207,
"grad_norm": 3.4219472408294678,
"learning_rate": 0.0001807726864330638,
"loss": 0.6228,
"step": 220
},
{
"epoch": 0.3102866779089376,
"grad_norm": 7.479050636291504,
"learning_rate": 0.000179874213836478,
"loss": 0.6035,
"step": 230
},
{
"epoch": 0.32377740303541314,
"grad_norm": 6.84128475189209,
"learning_rate": 0.0001789757412398922,
"loss": 0.5763,
"step": 240
},
{
"epoch": 0.3372681281618887,
"grad_norm": 3.407183885574341,
"learning_rate": 0.0001780772686433064,
"loss": 0.5981,
"step": 250
},
{
"epoch": 0.35075885328836426,
"grad_norm": 6.634527683258057,
"learning_rate": 0.0001771787960467206,
"loss": 0.6313,
"step": 260
},
{
"epoch": 0.3642495784148398,
"grad_norm": 6.274175643920898,
"learning_rate": 0.0001762803234501348,
"loss": 0.5971,
"step": 270
},
{
"epoch": 0.3777403035413153,
"grad_norm": 5.464880466461182,
"learning_rate": 0.00017538185085354897,
"loss": 0.629,
"step": 280
},
{
"epoch": 0.3912310286677909,
"grad_norm": 7.9421234130859375,
"learning_rate": 0.00017448337825696318,
"loss": 0.5604,
"step": 290
},
{
"epoch": 0.40472175379426645,
"grad_norm": 6.513714790344238,
"learning_rate": 0.00017358490566037738,
"loss": 0.5442,
"step": 300
},
{
"epoch": 0.418212478920742,
"grad_norm": 4.856393337249756,
"learning_rate": 0.00017268643306379155,
"loss": 0.5444,
"step": 310
},
{
"epoch": 0.4317032040472175,
"grad_norm": 7.252712249755859,
"learning_rate": 0.00017178796046720576,
"loss": 0.6071,
"step": 320
},
{
"epoch": 0.4451939291736931,
"grad_norm": 6.8953328132629395,
"learning_rate": 0.00017088948787061996,
"loss": 0.5622,
"step": 330
},
{
"epoch": 0.45868465430016864,
"grad_norm": 6.220505237579346,
"learning_rate": 0.00016999101527403416,
"loss": 0.6213,
"step": 340
},
{
"epoch": 0.47217537942664417,
"grad_norm": 1.9960452318191528,
"learning_rate": 0.00016909254267744833,
"loss": 0.5752,
"step": 350
},
{
"epoch": 0.4856661045531197,
"grad_norm": 4.8189778327941895,
"learning_rate": 0.00016819407008086254,
"loss": 0.5695,
"step": 360
},
{
"epoch": 0.4991568296795953,
"grad_norm": 2.157482624053955,
"learning_rate": 0.00016729559748427674,
"loss": 0.5818,
"step": 370
},
{
"epoch": 0.5126475548060708,
"grad_norm": 6.034278392791748,
"learning_rate": 0.00016639712488769091,
"loss": 0.5632,
"step": 380
},
{
"epoch": 0.5261382799325464,
"grad_norm": 3.6643030643463135,
"learning_rate": 0.00016549865229110512,
"loss": 0.5149,
"step": 390
},
{
"epoch": 0.5396290050590219,
"grad_norm": 3.4874682426452637,
"learning_rate": 0.00016460017969451932,
"loss": 0.5712,
"step": 400
},
{
"epoch": 0.5531197301854974,
"grad_norm": 5.94505500793457,
"learning_rate": 0.00016370170709793352,
"loss": 0.5579,
"step": 410
},
{
"epoch": 0.5666104553119731,
"grad_norm": 2.6858978271484375,
"learning_rate": 0.00016280323450134772,
"loss": 0.5601,
"step": 420
},
{
"epoch": 0.5801011804384486,
"grad_norm": 0.5919632315635681,
"learning_rate": 0.00016190476190476192,
"loss": 0.5122,
"step": 430
},
{
"epoch": 0.5935919055649241,
"grad_norm": 9.723355293273926,
"learning_rate": 0.00016100628930817613,
"loss": 0.5408,
"step": 440
},
{
"epoch": 0.6070826306913997,
"grad_norm": 3.421534776687622,
"learning_rate": 0.0001601078167115903,
"loss": 0.5125,
"step": 450
},
{
"epoch": 0.6205733558178752,
"grad_norm": 1.350819706916809,
"learning_rate": 0.0001592093441150045,
"loss": 0.5443,
"step": 460
},
{
"epoch": 0.6340640809443507,
"grad_norm": 2.9618473052978516,
"learning_rate": 0.0001583108715184187,
"loss": 0.5792,
"step": 470
},
{
"epoch": 0.6475548060708263,
"grad_norm": 11.16501522064209,
"learning_rate": 0.00015741239892183288,
"loss": 0.5499,
"step": 480
},
{
"epoch": 0.6610455311973018,
"grad_norm": 2.7122786045074463,
"learning_rate": 0.00015651392632524708,
"loss": 0.589,
"step": 490
},
{
"epoch": 0.6745362563237775,
"grad_norm": 6.2460408210754395,
"learning_rate": 0.00015561545372866128,
"loss": 0.5422,
"step": 500
},
{
"epoch": 0.688026981450253,
"grad_norm": 12.943184852600098,
"learning_rate": 0.0001547169811320755,
"loss": 0.5431,
"step": 510
},
{
"epoch": 0.7015177065767285,
"grad_norm": 2.4169883728027344,
"learning_rate": 0.00015381850853548966,
"loss": 0.4984,
"step": 520
},
{
"epoch": 0.715008431703204,
"grad_norm": 5.108154773712158,
"learning_rate": 0.00015292003593890386,
"loss": 0.4795,
"step": 530
},
{
"epoch": 0.7284991568296796,
"grad_norm": 3.268669605255127,
"learning_rate": 0.00015202156334231807,
"loss": 0.4961,
"step": 540
},
{
"epoch": 0.7419898819561551,
"grad_norm": 1.7926982641220093,
"learning_rate": 0.00015112309074573224,
"loss": 0.5339,
"step": 550
},
{
"epoch": 0.7554806070826307,
"grad_norm": 3.6382663249969482,
"learning_rate": 0.00015022461814914644,
"loss": 0.4855,
"step": 560
},
{
"epoch": 0.7689713322091062,
"grad_norm": 8.025957107543945,
"learning_rate": 0.00014932614555256067,
"loss": 0.4901,
"step": 570
},
{
"epoch": 0.7824620573355818,
"grad_norm": 4.220882892608643,
"learning_rate": 0.00014842767295597485,
"loss": 0.5448,
"step": 580
},
{
"epoch": 0.7959527824620574,
"grad_norm": 6.42817497253418,
"learning_rate": 0.00014752920035938905,
"loss": 0.4969,
"step": 590
},
{
"epoch": 0.8094435075885329,
"grad_norm": 4.182610988616943,
"learning_rate": 0.00014663072776280325,
"loss": 0.4863,
"step": 600
},
{
"epoch": 0.8229342327150084,
"grad_norm": 3.687934637069702,
"learning_rate": 0.00014573225516621745,
"loss": 0.5162,
"step": 610
},
{
"epoch": 0.836424957841484,
"grad_norm": 3.8240537643432617,
"learning_rate": 0.00014483378256963163,
"loss": 0.5089,
"step": 620
},
{
"epoch": 0.8499156829679595,
"grad_norm": 3.3823423385620117,
"learning_rate": 0.00014393530997304583,
"loss": 0.5447,
"step": 630
},
{
"epoch": 0.863406408094435,
"grad_norm": 4.591256618499756,
"learning_rate": 0.00014303683737646003,
"loss": 0.5067,
"step": 640
},
{
"epoch": 0.8768971332209107,
"grad_norm": 4.414494514465332,
"learning_rate": 0.0001421383647798742,
"loss": 0.4858,
"step": 650
},
{
"epoch": 0.8903878583473862,
"grad_norm": 3.4649970531463623,
"learning_rate": 0.0001412398921832884,
"loss": 0.5243,
"step": 660
},
{
"epoch": 0.9038785834738617,
"grad_norm": 4.813696384429932,
"learning_rate": 0.0001403414195867026,
"loss": 0.51,
"step": 670
},
{
"epoch": 0.9173693086003373,
"grad_norm": 4.597005844116211,
"learning_rate": 0.00013944294699011681,
"loss": 0.5079,
"step": 680
},
{
"epoch": 0.9308600337268128,
"grad_norm": 4.291003704071045,
"learning_rate": 0.000138544474393531,
"loss": 0.4966,
"step": 690
},
{
"epoch": 0.9443507588532883,
"grad_norm": 3.8949058055877686,
"learning_rate": 0.0001376460017969452,
"loss": 0.5086,
"step": 700
},
{
"epoch": 0.9578414839797639,
"grad_norm": 4.685634613037109,
"learning_rate": 0.0001367475292003594,
"loss": 0.5005,
"step": 710
},
{
"epoch": 0.9713322091062394,
"grad_norm": 3.9277801513671875,
"learning_rate": 0.0001358490566037736,
"loss": 0.4759,
"step": 720
},
{
"epoch": 0.984822934232715,
"grad_norm": 4.3011345863342285,
"learning_rate": 0.0001349505840071878,
"loss": 0.4371,
"step": 730
},
{
"epoch": 0.9983136593591906,
"grad_norm": 3.370706796646118,
"learning_rate": 0.000134052111410602,
"loss": 0.4811,
"step": 740
}
],
"logging_steps": 10,
"max_steps": 2226,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.690343765927526e+16,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}