{ "best_metric": 0.8470824949698189, "best_model_checkpoint": "PhoWhisper-small-vispeech-classifier-v4/checkpoint-490", "epoch": 1.0, "eval_steps": 500, "global_step": 490, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01020408163265306, "grad_norm": 56193.578125, "learning_rate": 2.5510204081632652e-08, "loss": 2.0798, "step": 5 }, { "epoch": 0.02040816326530612, "grad_norm": 50520.33203125, "learning_rate": 5.1020408163265303e-08, "loss": 2.0772, "step": 10 }, { "epoch": 0.030612244897959183, "grad_norm": 47583.30859375, "learning_rate": 7.653061224489796e-08, "loss": 2.0762, "step": 15 }, { "epoch": 0.04081632653061224, "grad_norm": 53174.3828125, "learning_rate": 1.0204081632653061e-07, "loss": 2.0754, "step": 20 }, { "epoch": 0.05102040816326531, "grad_norm": 65972.65625, "learning_rate": 1.2755102040816328e-07, "loss": 2.0755, "step": 25 }, { "epoch": 0.061224489795918366, "grad_norm": 53920.14453125, "learning_rate": 1.5306122448979592e-07, "loss": 2.0745, "step": 30 }, { "epoch": 0.07142857142857142, "grad_norm": 52966.375, "learning_rate": 1.7857142857142858e-07, "loss": 2.0734, "step": 35 }, { "epoch": 0.08163265306122448, "grad_norm": 60549.94921875, "learning_rate": 2.0408163265306121e-07, "loss": 2.0778, "step": 40 }, { "epoch": 0.09183673469387756, "grad_norm": 55827.60546875, "learning_rate": 2.295918367346939e-07, "loss": 2.0787, "step": 45 }, { "epoch": 0.10204081632653061, "grad_norm": 51334.3359375, "learning_rate": 2.5510204081632656e-07, "loss": 2.0759, "step": 50 }, { "epoch": 0.11224489795918367, "grad_norm": 70249.5078125, "learning_rate": 2.806122448979592e-07, "loss": 2.0698, "step": 55 }, { "epoch": 0.12244897959183673, "grad_norm": 54763.26171875, "learning_rate": 3.0612244897959183e-07, "loss": 2.0762, "step": 60 }, { "epoch": 0.1326530612244898, "grad_norm": 58244.69921875, "learning_rate": 3.3163265306122455e-07, "loss": 2.0696, "step": 65 }, { "epoch": 0.14285714285714285, "grad_norm": 51894.46484375, "learning_rate": 3.5714285714285716e-07, "loss": 2.0702, "step": 70 }, { "epoch": 0.15306122448979592, "grad_norm": 72845.734375, "learning_rate": 3.826530612244898e-07, "loss": 2.0676, "step": 75 }, { "epoch": 0.16326530612244897, "grad_norm": 56673.85546875, "learning_rate": 4.0816326530612243e-07, "loss": 2.0693, "step": 80 }, { "epoch": 0.17346938775510204, "grad_norm": 61739.0546875, "learning_rate": 4.3367346938775514e-07, "loss": 2.0668, "step": 85 }, { "epoch": 0.1836734693877551, "grad_norm": 68551.390625, "learning_rate": 4.591836734693878e-07, "loss": 2.0651, "step": 90 }, { "epoch": 0.19387755102040816, "grad_norm": 55835.78125, "learning_rate": 4.846938775510205e-07, "loss": 2.0701, "step": 95 }, { "epoch": 0.20408163265306123, "grad_norm": 60520.2734375, "learning_rate": 5.102040816326531e-07, "loss": 2.0654, "step": 100 }, { "epoch": 0.21428571428571427, "grad_norm": 45642.94921875, "learning_rate": 5.357142857142857e-07, "loss": 2.0672, "step": 105 }, { "epoch": 0.22448979591836735, "grad_norm": 67603.8515625, "learning_rate": 5.612244897959184e-07, "loss": 2.0673, "step": 110 }, { "epoch": 0.23469387755102042, "grad_norm": 54359.4375, "learning_rate": 5.867346938775511e-07, "loss": 2.0604, "step": 115 }, { "epoch": 0.24489795918367346, "grad_norm": 81289.5390625, "learning_rate": 6.122448979591837e-07, "loss": 2.061, "step": 120 }, { "epoch": 0.25510204081632654, "grad_norm": 70781.5703125, "learning_rate": 6.377551020408164e-07, "loss": 2.0573, "step": 125 }, { "epoch": 0.2653061224489796, "grad_norm": 73178.1875, "learning_rate": 6.632653061224491e-07, "loss": 2.0541, "step": 130 }, { "epoch": 0.2755102040816326, "grad_norm": 69632.25, "learning_rate": 6.887755102040817e-07, "loss": 2.0603, "step": 135 }, { "epoch": 0.2857142857142857, "grad_norm": 54642.47265625, "learning_rate": 7.142857142857143e-07, "loss": 2.0587, "step": 140 }, { "epoch": 0.29591836734693877, "grad_norm": 64884.2890625, "learning_rate": 7.39795918367347e-07, "loss": 2.0574, "step": 145 }, { "epoch": 0.30612244897959184, "grad_norm": 74709.9609375, "learning_rate": 7.653061224489796e-07, "loss": 2.0466, "step": 150 }, { "epoch": 0.3163265306122449, "grad_norm": 59361.31640625, "learning_rate": 7.908163265306124e-07, "loss": 2.0473, "step": 155 }, { "epoch": 0.32653061224489793, "grad_norm": 76060.171875, "learning_rate": 8.163265306122449e-07, "loss": 2.0413, "step": 160 }, { "epoch": 0.336734693877551, "grad_norm": 52821.26953125, "learning_rate": 8.418367346938776e-07, "loss": 2.0495, "step": 165 }, { "epoch": 0.3469387755102041, "grad_norm": 55647.62109375, "learning_rate": 8.673469387755103e-07, "loss": 2.0361, "step": 170 }, { "epoch": 0.35714285714285715, "grad_norm": 49648.828125, "learning_rate": 8.928571428571429e-07, "loss": 2.0395, "step": 175 }, { "epoch": 0.3673469387755102, "grad_norm": 63527.9140625, "learning_rate": 9.183673469387756e-07, "loss": 2.0334, "step": 180 }, { "epoch": 0.37755102040816324, "grad_norm": 80417.828125, "learning_rate": 9.438775510204082e-07, "loss": 2.0331, "step": 185 }, { "epoch": 0.3877551020408163, "grad_norm": 82782.7265625, "learning_rate": 9.69387755102041e-07, "loss": 2.0251, "step": 190 }, { "epoch": 0.3979591836734694, "grad_norm": 66964.8125, "learning_rate": 9.948979591836735e-07, "loss": 2.0289, "step": 195 }, { "epoch": 0.40816326530612246, "grad_norm": 73848.046875, "learning_rate": 1.0204081632653063e-06, "loss": 2.0259, "step": 200 }, { "epoch": 0.41836734693877553, "grad_norm": 70241.171875, "learning_rate": 1.0459183673469388e-06, "loss": 2.0082, "step": 205 }, { "epoch": 0.42857142857142855, "grad_norm": 63360.8125, "learning_rate": 1.0714285714285714e-06, "loss": 2.0091, "step": 210 }, { "epoch": 0.4387755102040816, "grad_norm": 82052.796875, "learning_rate": 1.0969387755102041e-06, "loss": 2.0008, "step": 215 }, { "epoch": 0.4489795918367347, "grad_norm": 63768.30078125, "learning_rate": 1.122448979591837e-06, "loss": 1.9958, "step": 220 }, { "epoch": 0.45918367346938777, "grad_norm": 90230.640625, "learning_rate": 1.1479591836734695e-06, "loss": 1.9816, "step": 225 }, { "epoch": 0.46938775510204084, "grad_norm": 48253.55078125, "learning_rate": 1.1734693877551022e-06, "loss": 2.0064, "step": 230 }, { "epoch": 0.47959183673469385, "grad_norm": 73846.0859375, "learning_rate": 1.1989795918367348e-06, "loss": 1.996, "step": 235 }, { "epoch": 0.4897959183673469, "grad_norm": 86008.8046875, "learning_rate": 1.2244897959183673e-06, "loss": 1.9858, "step": 240 }, { "epoch": 0.5, "grad_norm": 86437.734375, "learning_rate": 1.25e-06, "loss": 1.9696, "step": 245 }, { "epoch": 0.5102040816326531, "grad_norm": 82227.921875, "learning_rate": 1.2755102040816329e-06, "loss": 1.9803, "step": 250 }, { "epoch": 0.5204081632653061, "grad_norm": 96575.4765625, "learning_rate": 1.3010204081632654e-06, "loss": 1.9375, "step": 255 }, { "epoch": 0.5306122448979592, "grad_norm": 89167.7265625, "learning_rate": 1.3265306122448982e-06, "loss": 1.9585, "step": 260 }, { "epoch": 0.5408163265306123, "grad_norm": 101032.5, "learning_rate": 1.3520408163265307e-06, "loss": 1.947, "step": 265 }, { "epoch": 0.5510204081632653, "grad_norm": 105424.25, "learning_rate": 1.3775510204081633e-06, "loss": 1.9482, "step": 270 }, { "epoch": 0.5612244897959183, "grad_norm": 84136.484375, "learning_rate": 1.4030612244897959e-06, "loss": 1.9386, "step": 275 }, { "epoch": 0.5714285714285714, "grad_norm": 83034.078125, "learning_rate": 1.4285714285714286e-06, "loss": 1.9245, "step": 280 }, { "epoch": 0.5816326530612245, "grad_norm": 89200.4609375, "learning_rate": 1.4540816326530614e-06, "loss": 1.9138, "step": 285 }, { "epoch": 0.5918367346938775, "grad_norm": 103667.9609375, "learning_rate": 1.479591836734694e-06, "loss": 1.9283, "step": 290 }, { "epoch": 0.6020408163265306, "grad_norm": 98443.2734375, "learning_rate": 1.5051020408163267e-06, "loss": 1.9025, "step": 295 }, { "epoch": 0.6122448979591837, "grad_norm": 99525.0, "learning_rate": 1.5306122448979593e-06, "loss": 1.8933, "step": 300 }, { "epoch": 0.6224489795918368, "grad_norm": 145139.5625, "learning_rate": 1.556122448979592e-06, "loss": 1.8563, "step": 305 }, { "epoch": 0.6326530612244898, "grad_norm": 116158.859375, "learning_rate": 1.5816326530612248e-06, "loss": 1.8931, "step": 310 }, { "epoch": 0.6428571428571429, "grad_norm": 131590.09375, "learning_rate": 1.6071428571428574e-06, "loss": 1.8599, "step": 315 }, { "epoch": 0.6530612244897959, "grad_norm": 112483.40625, "learning_rate": 1.6326530612244897e-06, "loss": 1.8596, "step": 320 }, { "epoch": 0.6632653061224489, "grad_norm": 82979.0859375, "learning_rate": 1.6581632653061225e-06, "loss": 1.883, "step": 325 }, { "epoch": 0.673469387755102, "grad_norm": 96548.5703125, "learning_rate": 1.6836734693877552e-06, "loss": 1.8586, "step": 330 }, { "epoch": 0.6836734693877551, "grad_norm": 113625.171875, "learning_rate": 1.7091836734693878e-06, "loss": 1.8375, "step": 335 }, { "epoch": 0.6938775510204082, "grad_norm": 115079.0078125, "learning_rate": 1.7346938775510206e-06, "loss": 1.838, "step": 340 }, { "epoch": 0.7040816326530612, "grad_norm": 139868.21875, "learning_rate": 1.7602040816326531e-06, "loss": 1.8013, "step": 345 }, { "epoch": 0.7142857142857143, "grad_norm": 126350.546875, "learning_rate": 1.7857142857142859e-06, "loss": 1.847, "step": 350 }, { "epoch": 0.7244897959183674, "grad_norm": 149690.375, "learning_rate": 1.8112244897959187e-06, "loss": 1.8203, "step": 355 }, { "epoch": 0.7346938775510204, "grad_norm": 132708.578125, "learning_rate": 1.8367346938775512e-06, "loss": 1.7931, "step": 360 }, { "epoch": 0.7448979591836735, "grad_norm": 128203.3203125, "learning_rate": 1.862244897959184e-06, "loss": 1.7533, "step": 365 }, { "epoch": 0.7551020408163265, "grad_norm": 118367.4765625, "learning_rate": 1.8877551020408163e-06, "loss": 1.7662, "step": 370 }, { "epoch": 0.7653061224489796, "grad_norm": 162099.25, "learning_rate": 1.913265306122449e-06, "loss": 1.7775, "step": 375 }, { "epoch": 0.7755102040816326, "grad_norm": 132369.234375, "learning_rate": 1.938775510204082e-06, "loss": 1.7908, "step": 380 }, { "epoch": 0.7857142857142857, "grad_norm": 161412.09375, "learning_rate": 1.9642857142857144e-06, "loss": 1.7045, "step": 385 }, { "epoch": 0.7959183673469388, "grad_norm": 117275.65625, "learning_rate": 1.989795918367347e-06, "loss": 1.74, "step": 390 }, { "epoch": 0.8061224489795918, "grad_norm": 134175.890625, "learning_rate": 2.01530612244898e-06, "loss": 1.7041, "step": 395 }, { "epoch": 0.8163265306122449, "grad_norm": 165093.34375, "learning_rate": 2.0408163265306125e-06, "loss": 1.7217, "step": 400 }, { "epoch": 0.826530612244898, "grad_norm": 108558.96875, "learning_rate": 2.066326530612245e-06, "loss": 1.7211, "step": 405 }, { "epoch": 0.8367346938775511, "grad_norm": 159761.328125, "learning_rate": 2.0918367346938776e-06, "loss": 1.7056, "step": 410 }, { "epoch": 0.8469387755102041, "grad_norm": 131097.640625, "learning_rate": 2.1173469387755106e-06, "loss": 1.7289, "step": 415 }, { "epoch": 0.8571428571428571, "grad_norm": 162492.6875, "learning_rate": 2.1428571428571427e-06, "loss": 1.7312, "step": 420 }, { "epoch": 0.8673469387755102, "grad_norm": 122069.125, "learning_rate": 2.1683673469387757e-06, "loss": 1.6696, "step": 425 }, { "epoch": 0.8775510204081632, "grad_norm": 194332.5625, "learning_rate": 2.1938775510204083e-06, "loss": 1.6226, "step": 430 }, { "epoch": 0.8877551020408163, "grad_norm": 132711.28125, "learning_rate": 2.219387755102041e-06, "loss": 1.663, "step": 435 }, { "epoch": 0.8979591836734694, "grad_norm": 166239.375, "learning_rate": 2.244897959183674e-06, "loss": 1.6127, "step": 440 }, { "epoch": 0.9081632653061225, "grad_norm": 132459.359375, "learning_rate": 2.2704081632653064e-06, "loss": 1.6042, "step": 445 }, { "epoch": 0.9183673469387755, "grad_norm": 114693.3046875, "learning_rate": 2.295918367346939e-06, "loss": 1.6276, "step": 450 }, { "epoch": 0.9285714285714286, "grad_norm": 184974.53125, "learning_rate": 2.321428571428572e-06, "loss": 1.6448, "step": 455 }, { "epoch": 0.9387755102040817, "grad_norm": 134113.625, "learning_rate": 2.3469387755102044e-06, "loss": 1.6329, "step": 460 }, { "epoch": 0.9489795918367347, "grad_norm": 137895.78125, "learning_rate": 2.372448979591837e-06, "loss": 1.5441, "step": 465 }, { "epoch": 0.9591836734693877, "grad_norm": 158595.234375, "learning_rate": 2.3979591836734696e-06, "loss": 1.5524, "step": 470 }, { "epoch": 0.9693877551020408, "grad_norm": 162135.703125, "learning_rate": 2.423469387755102e-06, "loss": 1.5796, "step": 475 }, { "epoch": 0.9795918367346939, "grad_norm": 205519.296875, "learning_rate": 2.4489795918367347e-06, "loss": 1.5343, "step": 480 }, { "epoch": 0.9897959183673469, "grad_norm": 193819.75, "learning_rate": 2.4744897959183676e-06, "loss": 1.5776, "step": 485 }, { "epoch": 1.0, "grad_norm": 225175.609375, "learning_rate": 2.5e-06, "loss": 1.5384, "step": 490 }, { "epoch": 1.0, "eval_accuracy": 0.8470824949698189, "eval_loss": 1.4336239099502563, "eval_runtime": 214.7827, "eval_samples_per_second": 13.884, "eval_steps_per_second": 0.582, "step": 490 } ], "logging_steps": 5, "max_steps": 4900, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.49679069336576e+18, "train_batch_size": 24, "trial_name": null, "trial_params": null }