trocr-large-printed-cmc7-ocr / trainer_state.json
rocassis's picture
All Dunn!!!
95b81a5 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 5.0,
"eval_steps": 500,
"global_step": 3120,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0016025641025641025,
"grad_norm": 213.62364196777344,
"learning_rate": 4.998397435897436e-05,
"loss": 9.3378,
"step": 1
},
{
"epoch": 0.049679487179487176,
"grad_norm": 43.086360931396484,
"learning_rate": 4.950320512820513e-05,
"loss": 2.5252,
"step": 31
},
{
"epoch": 0.09935897435897435,
"grad_norm": 48.219451904296875,
"learning_rate": 4.9006410256410256e-05,
"loss": 1.5213,
"step": 62
},
{
"epoch": 0.14903846153846154,
"grad_norm": 30.23198127746582,
"learning_rate": 4.850961538461539e-05,
"loss": 1.9636,
"step": 93
},
{
"epoch": 0.1987179487179487,
"grad_norm": 52.96772766113281,
"learning_rate": 4.8012820512820516e-05,
"loss": 1.3545,
"step": 124
},
{
"epoch": 0.2483974358974359,
"grad_norm": 48.79069900512695,
"learning_rate": 4.751602564102564e-05,
"loss": 1.4998,
"step": 155
},
{
"epoch": 0.2980769230769231,
"grad_norm": 38.8032341003418,
"learning_rate": 4.701923076923077e-05,
"loss": 1.4654,
"step": 186
},
{
"epoch": 0.34775641025641024,
"grad_norm": 106.62429809570312,
"learning_rate": 4.65224358974359e-05,
"loss": 1.8027,
"step": 217
},
{
"epoch": 0.3974358974358974,
"grad_norm": 25.370744705200195,
"learning_rate": 4.602564102564102e-05,
"loss": 1.4394,
"step": 248
},
{
"epoch": 0.44711538461538464,
"grad_norm": 82.52824401855469,
"learning_rate": 4.5528846153846157e-05,
"loss": 1.4573,
"step": 279
},
{
"epoch": 0.4967948717948718,
"grad_norm": 19.148210525512695,
"learning_rate": 4.503205128205128e-05,
"loss": 1.2547,
"step": 310
},
{
"epoch": 0.5464743589743589,
"grad_norm": 46.984779357910156,
"learning_rate": 4.453525641025642e-05,
"loss": 1.3297,
"step": 341
},
{
"epoch": 0.5961538461538461,
"grad_norm": 45.50709915161133,
"learning_rate": 4.403846153846154e-05,
"loss": 1.3359,
"step": 372
},
{
"epoch": 0.6458333333333334,
"grad_norm": 36.45267105102539,
"learning_rate": 4.354166666666667e-05,
"loss": 1.2466,
"step": 403
},
{
"epoch": 0.6955128205128205,
"grad_norm": 20.8362979888916,
"learning_rate": 4.30448717948718e-05,
"loss": 1.1206,
"step": 434
},
{
"epoch": 0.7451923076923077,
"grad_norm": 64.35557556152344,
"learning_rate": 4.2548076923076924e-05,
"loss": 1.3676,
"step": 465
},
{
"epoch": 0.7948717948717948,
"grad_norm": 12.930615425109863,
"learning_rate": 4.205128205128206e-05,
"loss": 1.3871,
"step": 496
},
{
"epoch": 0.844551282051282,
"grad_norm": 51.6870002746582,
"learning_rate": 4.1554487179487184e-05,
"loss": 1.0038,
"step": 527
},
{
"epoch": 0.8942307692307693,
"grad_norm": 14.677131652832031,
"learning_rate": 4.105769230769231e-05,
"loss": 1.2283,
"step": 558
},
{
"epoch": 0.9439102564102564,
"grad_norm": 43.24152755737305,
"learning_rate": 4.056089743589744e-05,
"loss": 1.0872,
"step": 589
},
{
"epoch": 0.9935897435897436,
"grad_norm": 25.641794204711914,
"learning_rate": 4.006410256410257e-05,
"loss": 0.9991,
"step": 620
},
{
"epoch": 1.0432692307692308,
"grad_norm": 36.622703552246094,
"learning_rate": 3.956730769230769e-05,
"loss": 0.9941,
"step": 651
},
{
"epoch": 1.092948717948718,
"grad_norm": 17.249757766723633,
"learning_rate": 3.9070512820512824e-05,
"loss": 0.9877,
"step": 682
},
{
"epoch": 1.142628205128205,
"grad_norm": 31.591552734375,
"learning_rate": 3.857371794871795e-05,
"loss": 0.8917,
"step": 713
},
{
"epoch": 1.1923076923076923,
"grad_norm": 28.23312759399414,
"learning_rate": 3.807692307692308e-05,
"loss": 0.9264,
"step": 744
},
{
"epoch": 1.2419871794871795,
"grad_norm": 37.707183837890625,
"learning_rate": 3.7580128205128204e-05,
"loss": 0.849,
"step": 775
},
{
"epoch": 1.2916666666666667,
"grad_norm": 12.67361831665039,
"learning_rate": 3.708333333333334e-05,
"loss": 0.8255,
"step": 806
},
{
"epoch": 1.3413461538461537,
"grad_norm": 39.1461181640625,
"learning_rate": 3.658653846153846e-05,
"loss": 0.7372,
"step": 837
},
{
"epoch": 1.391025641025641,
"grad_norm": 33.27256393432617,
"learning_rate": 3.608974358974359e-05,
"loss": 0.8227,
"step": 868
},
{
"epoch": 1.4407051282051282,
"grad_norm": 545.9265747070312,
"learning_rate": 3.559294871794872e-05,
"loss": 0.7941,
"step": 899
},
{
"epoch": 1.4903846153846154,
"grad_norm": 23.55372428894043,
"learning_rate": 3.5096153846153845e-05,
"loss": 0.8686,
"step": 930
},
{
"epoch": 1.5400641025641026,
"grad_norm": 22.64493179321289,
"learning_rate": 3.459935897435898e-05,
"loss": 0.7059,
"step": 961
},
{
"epoch": 1.5897435897435899,
"grad_norm": 79.4326171875,
"learning_rate": 3.4102564102564105e-05,
"loss": 0.7567,
"step": 992
},
{
"epoch": 1.6394230769230769,
"grad_norm": 13.156254768371582,
"learning_rate": 3.360576923076923e-05,
"loss": 0.7138,
"step": 1023
},
{
"epoch": 1.689102564102564,
"grad_norm": 39.97475051879883,
"learning_rate": 3.310897435897436e-05,
"loss": 0.676,
"step": 1054
},
{
"epoch": 1.7387820512820513,
"grad_norm": 24.62538719177246,
"learning_rate": 3.261217948717949e-05,
"loss": 0.683,
"step": 1085
},
{
"epoch": 1.7884615384615383,
"grad_norm": 18.064680099487305,
"learning_rate": 3.211538461538462e-05,
"loss": 0.5746,
"step": 1116
},
{
"epoch": 1.8381410256410255,
"grad_norm": 10.56497573852539,
"learning_rate": 3.1618589743589746e-05,
"loss": 0.5881,
"step": 1147
},
{
"epoch": 1.8878205128205128,
"grad_norm": 43.007911682128906,
"learning_rate": 3.112179487179487e-05,
"loss": 0.78,
"step": 1178
},
{
"epoch": 1.9375,
"grad_norm": 25.748680114746094,
"learning_rate": 3.0625000000000006e-05,
"loss": 0.592,
"step": 1209
},
{
"epoch": 1.9871794871794872,
"grad_norm": 20.511089324951172,
"learning_rate": 3.012820512820513e-05,
"loss": 0.6807,
"step": 1240
},
{
"epoch": 2.0368589743589745,
"grad_norm": 26.275846481323242,
"learning_rate": 2.963141025641026e-05,
"loss": 0.6935,
"step": 1271
},
{
"epoch": 2.0865384615384617,
"grad_norm": 69.18566131591797,
"learning_rate": 2.913461538461539e-05,
"loss": 0.511,
"step": 1302
},
{
"epoch": 2.136217948717949,
"grad_norm": 4.620207786560059,
"learning_rate": 2.8637820512820513e-05,
"loss": 0.5338,
"step": 1333
},
{
"epoch": 2.185897435897436,
"grad_norm": 44.329566955566406,
"learning_rate": 2.8141025641025643e-05,
"loss": 0.4471,
"step": 1364
},
{
"epoch": 2.235576923076923,
"grad_norm": 24.95821762084961,
"learning_rate": 2.7644230769230773e-05,
"loss": 0.5295,
"step": 1395
},
{
"epoch": 2.28525641025641,
"grad_norm": 32.0108757019043,
"learning_rate": 2.7147435897435896e-05,
"loss": 0.5075,
"step": 1426
},
{
"epoch": 2.3349358974358974,
"grad_norm": 15.610992431640625,
"learning_rate": 2.6650641025641026e-05,
"loss": 0.4898,
"step": 1457
},
{
"epoch": 2.3846153846153846,
"grad_norm": 6.399534225463867,
"learning_rate": 2.6153846153846157e-05,
"loss": 0.4273,
"step": 1488
},
{
"epoch": 2.434294871794872,
"grad_norm": 22.35305404663086,
"learning_rate": 2.565705128205128e-05,
"loss": 0.4826,
"step": 1519
},
{
"epoch": 2.483974358974359,
"grad_norm": 26.82170867919922,
"learning_rate": 2.516025641025641e-05,
"loss": 0.4472,
"step": 1550
},
{
"epoch": 2.5336538461538463,
"grad_norm": 31.885419845581055,
"learning_rate": 2.466346153846154e-05,
"loss": 0.3986,
"step": 1581
},
{
"epoch": 2.5833333333333335,
"grad_norm": 21.020639419555664,
"learning_rate": 2.4166666666666667e-05,
"loss": 0.4856,
"step": 1612
},
{
"epoch": 2.6330128205128203,
"grad_norm": 10.991323471069336,
"learning_rate": 2.3669871794871794e-05,
"loss": 0.5008,
"step": 1643
},
{
"epoch": 2.6826923076923075,
"grad_norm": 17.077383041381836,
"learning_rate": 2.3173076923076924e-05,
"loss": 0.3585,
"step": 1674
},
{
"epoch": 2.7323717948717947,
"grad_norm": 6.274226188659668,
"learning_rate": 2.2676282051282054e-05,
"loss": 0.4101,
"step": 1705
},
{
"epoch": 2.782051282051282,
"grad_norm": 7.574007511138916,
"learning_rate": 2.217948717948718e-05,
"loss": 0.3604,
"step": 1736
},
{
"epoch": 2.831730769230769,
"grad_norm": 15.08122730255127,
"learning_rate": 2.168269230769231e-05,
"loss": 0.407,
"step": 1767
},
{
"epoch": 2.8814102564102564,
"grad_norm": 32.56257247924805,
"learning_rate": 2.1185897435897437e-05,
"loss": 0.4002,
"step": 1798
},
{
"epoch": 2.9310897435897436,
"grad_norm": 34.24089431762695,
"learning_rate": 2.0689102564102564e-05,
"loss": 0.429,
"step": 1829
},
{
"epoch": 2.980769230769231,
"grad_norm": 5.287622928619385,
"learning_rate": 2.0192307692307694e-05,
"loss": 0.352,
"step": 1860
},
{
"epoch": 3.030448717948718,
"grad_norm": 33.49539566040039,
"learning_rate": 1.969551282051282e-05,
"loss": 0.3348,
"step": 1891
},
{
"epoch": 3.0801282051282053,
"grad_norm": 3.116619825363159,
"learning_rate": 1.919871794871795e-05,
"loss": 0.2613,
"step": 1922
},
{
"epoch": 3.1298076923076925,
"grad_norm": 5.825009346008301,
"learning_rate": 1.8701923076923078e-05,
"loss": 0.279,
"step": 1953
},
{
"epoch": 3.1794871794871793,
"grad_norm": 2.2447965145111084,
"learning_rate": 1.8205128205128204e-05,
"loss": 0.3405,
"step": 1984
},
{
"epoch": 3.2291666666666665,
"grad_norm": 6.297438144683838,
"learning_rate": 1.7708333333333335e-05,
"loss": 0.3294,
"step": 2015
},
{
"epoch": 3.2788461538461537,
"grad_norm": 17.42154884338379,
"learning_rate": 1.721153846153846e-05,
"loss": 0.296,
"step": 2046
},
{
"epoch": 3.328525641025641,
"grad_norm": 11.834112167358398,
"learning_rate": 1.6714743589743588e-05,
"loss": 0.3767,
"step": 2077
},
{
"epoch": 3.378205128205128,
"grad_norm": 2.2308290004730225,
"learning_rate": 1.6217948717948718e-05,
"loss": 0.2352,
"step": 2108
},
{
"epoch": 3.4278846153846154,
"grad_norm": 13.97610855102539,
"learning_rate": 1.5721153846153845e-05,
"loss": 0.2937,
"step": 2139
},
{
"epoch": 3.4775641025641026,
"grad_norm": 4.005247116088867,
"learning_rate": 1.5224358974358973e-05,
"loss": 0.283,
"step": 2170
},
{
"epoch": 3.52724358974359,
"grad_norm": 5.180654525756836,
"learning_rate": 1.4727564102564103e-05,
"loss": 0.2773,
"step": 2201
},
{
"epoch": 3.5769230769230766,
"grad_norm": 6.487146854400635,
"learning_rate": 1.423076923076923e-05,
"loss": 0.2665,
"step": 2232
},
{
"epoch": 3.626602564102564,
"grad_norm": 2.202457904815674,
"learning_rate": 1.373397435897436e-05,
"loss": 0.2541,
"step": 2263
},
{
"epoch": 3.676282051282051,
"grad_norm": 47.195499420166016,
"learning_rate": 1.3237179487179489e-05,
"loss": 0.2362,
"step": 2294
},
{
"epoch": 3.7259615384615383,
"grad_norm": 21.26088523864746,
"learning_rate": 1.2740384615384615e-05,
"loss": 0.2331,
"step": 2325
},
{
"epoch": 3.7756410256410255,
"grad_norm": 16.77073860168457,
"learning_rate": 1.2243589743589744e-05,
"loss": 0.2134,
"step": 2356
},
{
"epoch": 3.8253205128205128,
"grad_norm": 2.246936559677124,
"learning_rate": 1.1746794871794872e-05,
"loss": 0.2437,
"step": 2387
},
{
"epoch": 3.875,
"grad_norm": 1.4183766841888428,
"learning_rate": 1.125e-05,
"loss": 0.2012,
"step": 2418
},
{
"epoch": 3.9246794871794872,
"grad_norm": 3.9523444175720215,
"learning_rate": 1.0753205128205129e-05,
"loss": 0.2204,
"step": 2449
},
{
"epoch": 3.9743589743589745,
"grad_norm": 3.564764976501465,
"learning_rate": 1.0256410256410256e-05,
"loss": 0.2034,
"step": 2480
},
{
"epoch": 4.024038461538462,
"grad_norm": 1.7849295139312744,
"learning_rate": 9.759615384615384e-06,
"loss": 0.1819,
"step": 2511
},
{
"epoch": 4.073717948717949,
"grad_norm": 1.7893842458724976,
"learning_rate": 9.262820512820514e-06,
"loss": 0.1585,
"step": 2542
},
{
"epoch": 4.123397435897436,
"grad_norm": 3.4071810245513916,
"learning_rate": 8.766025641025641e-06,
"loss": 0.1596,
"step": 2573
},
{
"epoch": 4.173076923076923,
"grad_norm": 28.766258239746094,
"learning_rate": 8.26923076923077e-06,
"loss": 0.1563,
"step": 2604
},
{
"epoch": 4.222756410256411,
"grad_norm": 3.066229820251465,
"learning_rate": 7.772435897435898e-06,
"loss": 0.1612,
"step": 2635
},
{
"epoch": 4.272435897435898,
"grad_norm": 1.4539713859558105,
"learning_rate": 7.275641025641026e-06,
"loss": 0.151,
"step": 2666
},
{
"epoch": 4.322115384615385,
"grad_norm": 3.9048240184783936,
"learning_rate": 6.778846153846154e-06,
"loss": 0.1514,
"step": 2697
},
{
"epoch": 4.371794871794872,
"grad_norm": 1.2496715784072876,
"learning_rate": 6.282051282051282e-06,
"loss": 0.1632,
"step": 2728
},
{
"epoch": 4.421474358974359,
"grad_norm": 1.516411542892456,
"learning_rate": 5.785256410256411e-06,
"loss": 0.1497,
"step": 2759
},
{
"epoch": 4.471153846153846,
"grad_norm": 2.217137575149536,
"learning_rate": 5.288461538461538e-06,
"loss": 0.1514,
"step": 2790
},
{
"epoch": 4.520833333333333,
"grad_norm": 1.7278540134429932,
"learning_rate": 4.791666666666667e-06,
"loss": 0.1409,
"step": 2821
},
{
"epoch": 4.57051282051282,
"grad_norm": 2.3950905799865723,
"learning_rate": 4.294871794871795e-06,
"loss": 0.1524,
"step": 2852
},
{
"epoch": 4.6201923076923075,
"grad_norm": 2.2806148529052734,
"learning_rate": 3.798076923076923e-06,
"loss": 0.1498,
"step": 2883
},
{
"epoch": 4.669871794871795,
"grad_norm": 2.3029630184173584,
"learning_rate": 3.3012820512820517e-06,
"loss": 0.1471,
"step": 2914
},
{
"epoch": 4.719551282051282,
"grad_norm": 1.592606782913208,
"learning_rate": 2.8044871794871797e-06,
"loss": 0.14,
"step": 2945
},
{
"epoch": 4.769230769230769,
"grad_norm": 3.534292221069336,
"learning_rate": 2.307692307692308e-06,
"loss": 0.1426,
"step": 2976
},
{
"epoch": 4.818910256410256,
"grad_norm": 2.344146966934204,
"learning_rate": 1.810897435897436e-06,
"loss": 0.1399,
"step": 3007
},
{
"epoch": 4.868589743589744,
"grad_norm": 11.425804138183594,
"learning_rate": 1.3141025641025643e-06,
"loss": 0.1374,
"step": 3038
},
{
"epoch": 4.918269230769231,
"grad_norm": 2.704406499862671,
"learning_rate": 8.173076923076924e-07,
"loss": 0.1389,
"step": 3069
},
{
"epoch": 4.967948717948718,
"grad_norm": 6.8166728019714355,
"learning_rate": 3.205128205128205e-07,
"loss": 0.1721,
"step": 3100
},
{
"epoch": 5.0,
"step": 3120,
"total_flos": 1.845867535870722e+19,
"train_loss": 0.6135548925552613,
"train_runtime": 6226.8155,
"train_samples_per_second": 2.003,
"train_steps_per_second": 0.501
}
],
"logging_steps": 31,
"max_steps": 3120,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.845867535870722e+19,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}