hr16's picture
Training in progress, epoch 2, checkpoint
2383570 verified
raw
history blame
33.9 kB
{
"best_metric": 0.8883299798792756,
"best_model_checkpoint": "PhoWhisper-small-vispeech-classifier-v4/checkpoint-490",
"epoch": 2.0,
"eval_steps": 500,
"global_step": 980,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01020408163265306,
"grad_norm": 51640.66015625,
"learning_rate": 5.1020408163265303e-08,
"loss": 2.0778,
"step": 5
},
{
"epoch": 0.02040816326530612,
"grad_norm": 46773.35546875,
"learning_rate": 1.0204081632653061e-07,
"loss": 2.0809,
"step": 10
},
{
"epoch": 0.030612244897959183,
"grad_norm": 45343.12109375,
"learning_rate": 1.5306122448979592e-07,
"loss": 2.0771,
"step": 15
},
{
"epoch": 0.04081632653061224,
"grad_norm": 49285.91796875,
"learning_rate": 2.0408163265306121e-07,
"loss": 2.077,
"step": 20
},
{
"epoch": 0.05102040816326531,
"grad_norm": 63271.82421875,
"learning_rate": 2.5510204081632656e-07,
"loss": 2.0755,
"step": 25
},
{
"epoch": 0.061224489795918366,
"grad_norm": 51649.79296875,
"learning_rate": 3.0612244897959183e-07,
"loss": 2.0764,
"step": 30
},
{
"epoch": 0.07142857142857142,
"grad_norm": 51633.98828125,
"learning_rate": 3.5714285714285716e-07,
"loss": 2.0763,
"step": 35
},
{
"epoch": 0.08163265306122448,
"grad_norm": 61777.23828125,
"learning_rate": 4.0816326530612243e-07,
"loss": 2.0772,
"step": 40
},
{
"epoch": 0.09183673469387756,
"grad_norm": 58467.15625,
"learning_rate": 4.591836734693878e-07,
"loss": 2.0743,
"step": 45
},
{
"epoch": 0.10204081632653061,
"grad_norm": 48974.37890625,
"learning_rate": 5.102040816326531e-07,
"loss": 2.072,
"step": 50
},
{
"epoch": 0.11224489795918367,
"grad_norm": 71007.515625,
"learning_rate": 5.612244897959184e-07,
"loss": 2.0751,
"step": 55
},
{
"epoch": 0.12244897959183673,
"grad_norm": 58237.77734375,
"learning_rate": 6.122448979591837e-07,
"loss": 2.0702,
"step": 60
},
{
"epoch": 0.1326530612244898,
"grad_norm": 58785.83203125,
"learning_rate": 6.632653061224491e-07,
"loss": 2.0713,
"step": 65
},
{
"epoch": 0.14285714285714285,
"grad_norm": 52771.265625,
"learning_rate": 7.142857142857143e-07,
"loss": 2.069,
"step": 70
},
{
"epoch": 0.15306122448979592,
"grad_norm": 69647.078125,
"learning_rate": 7.653061224489796e-07,
"loss": 2.0684,
"step": 75
},
{
"epoch": 0.16326530612244897,
"grad_norm": 59148.47265625,
"learning_rate": 8.163265306122449e-07,
"loss": 2.0634,
"step": 80
},
{
"epoch": 0.17346938775510204,
"grad_norm": 66723.921875,
"learning_rate": 8.673469387755103e-07,
"loss": 2.064,
"step": 85
},
{
"epoch": 0.1836734693877551,
"grad_norm": 72545.203125,
"learning_rate": 9.183673469387756e-07,
"loss": 2.0593,
"step": 90
},
{
"epoch": 0.19387755102040816,
"grad_norm": 59112.55859375,
"learning_rate": 9.69387755102041e-07,
"loss": 2.0553,
"step": 95
},
{
"epoch": 0.20408163265306123,
"grad_norm": 68369.265625,
"learning_rate": 1.0204081632653063e-06,
"loss": 2.0566,
"step": 100
},
{
"epoch": 0.21428571428571427,
"grad_norm": 51080.62890625,
"learning_rate": 1.0714285714285714e-06,
"loss": 2.0505,
"step": 105
},
{
"epoch": 0.22448979591836735,
"grad_norm": 78760.0859375,
"learning_rate": 1.122448979591837e-06,
"loss": 2.0496,
"step": 110
},
{
"epoch": 0.23469387755102042,
"grad_norm": 57198.7890625,
"learning_rate": 1.1734693877551022e-06,
"loss": 2.0458,
"step": 115
},
{
"epoch": 0.24489795918367346,
"grad_norm": 98369.4609375,
"learning_rate": 1.2244897959183673e-06,
"loss": 2.0336,
"step": 120
},
{
"epoch": 0.25510204081632654,
"grad_norm": 80801.5703125,
"learning_rate": 1.2755102040816329e-06,
"loss": 2.0354,
"step": 125
},
{
"epoch": 0.2653061224489796,
"grad_norm": 82179.9765625,
"learning_rate": 1.3265306122448982e-06,
"loss": 2.0263,
"step": 130
},
{
"epoch": 0.2755102040816326,
"grad_norm": 83225.078125,
"learning_rate": 1.3775510204081633e-06,
"loss": 2.0168,
"step": 135
},
{
"epoch": 0.2857142857142857,
"grad_norm": 62451.64453125,
"learning_rate": 1.4285714285714286e-06,
"loss": 2.023,
"step": 140
},
{
"epoch": 0.29591836734693877,
"grad_norm": 75567.9453125,
"learning_rate": 1.479591836734694e-06,
"loss": 2.0206,
"step": 145
},
{
"epoch": 0.30612244897959184,
"grad_norm": 98461.484375,
"learning_rate": 1.5306122448979593e-06,
"loss": 1.9984,
"step": 150
},
{
"epoch": 0.3163265306122449,
"grad_norm": 75078.1484375,
"learning_rate": 1.5816326530612248e-06,
"loss": 1.9899,
"step": 155
},
{
"epoch": 0.32653061224489793,
"grad_norm": 97895.671875,
"learning_rate": 1.6326530612244897e-06,
"loss": 1.9758,
"step": 160
},
{
"epoch": 0.336734693877551,
"grad_norm": 65880.2421875,
"learning_rate": 1.6836734693877552e-06,
"loss": 1.9918,
"step": 165
},
{
"epoch": 0.3469387755102041,
"grad_norm": 70675.578125,
"learning_rate": 1.7346938775510206e-06,
"loss": 1.9715,
"step": 170
},
{
"epoch": 0.35714285714285715,
"grad_norm": 64327.19921875,
"learning_rate": 1.7857142857142859e-06,
"loss": 1.9597,
"step": 175
},
{
"epoch": 0.3673469387755102,
"grad_norm": 80610.6171875,
"learning_rate": 1.8367346938775512e-06,
"loss": 1.9632,
"step": 180
},
{
"epoch": 0.37755102040816324,
"grad_norm": 110899.375,
"learning_rate": 1.8877551020408163e-06,
"loss": 1.9476,
"step": 185
},
{
"epoch": 0.3877551020408163,
"grad_norm": 111132.5703125,
"learning_rate": 1.938775510204082e-06,
"loss": 1.9369,
"step": 190
},
{
"epoch": 0.3979591836734694,
"grad_norm": 88722.625,
"learning_rate": 1.989795918367347e-06,
"loss": 1.9452,
"step": 195
},
{
"epoch": 0.40816326530612246,
"grad_norm": 105874.296875,
"learning_rate": 2.0408163265306125e-06,
"loss": 1.9459,
"step": 200
},
{
"epoch": 0.41836734693877553,
"grad_norm": 97286.0078125,
"learning_rate": 2.0918367346938776e-06,
"loss": 1.918,
"step": 205
},
{
"epoch": 0.42857142857142855,
"grad_norm": 82848.0625,
"learning_rate": 2.1428571428571427e-06,
"loss": 1.9135,
"step": 210
},
{
"epoch": 0.4387755102040816,
"grad_norm": 114854.4609375,
"learning_rate": 2.1938775510204083e-06,
"loss": 1.8869,
"step": 215
},
{
"epoch": 0.4489795918367347,
"grad_norm": 86661.25,
"learning_rate": 2.244897959183674e-06,
"loss": 1.8861,
"step": 220
},
{
"epoch": 0.45918367346938777,
"grad_norm": 127591.3984375,
"learning_rate": 2.295918367346939e-06,
"loss": 1.8396,
"step": 225
},
{
"epoch": 0.46938775510204084,
"grad_norm": 71038.8671875,
"learning_rate": 2.3469387755102044e-06,
"loss": 1.8995,
"step": 230
},
{
"epoch": 0.47959183673469385,
"grad_norm": 101402.421875,
"learning_rate": 2.3979591836734696e-06,
"loss": 1.8699,
"step": 235
},
{
"epoch": 0.4897959183673469,
"grad_norm": 113157.3828125,
"learning_rate": 2.4489795918367347e-06,
"loss": 1.8599,
"step": 240
},
{
"epoch": 0.5,
"grad_norm": 115485.0546875,
"learning_rate": 2.5e-06,
"loss": 1.8214,
"step": 245
},
{
"epoch": 0.5102040816326531,
"grad_norm": 139369.8125,
"learning_rate": 2.5510204081632657e-06,
"loss": 1.8539,
"step": 250
},
{
"epoch": 0.5204081632653061,
"grad_norm": 141401.46875,
"learning_rate": 2.602040816326531e-06,
"loss": 1.7507,
"step": 255
},
{
"epoch": 0.5306122448979592,
"grad_norm": 125767.3984375,
"learning_rate": 2.6530612244897964e-06,
"loss": 1.7948,
"step": 260
},
{
"epoch": 0.5408163265306123,
"grad_norm": 148848.5625,
"learning_rate": 2.7040816326530615e-06,
"loss": 1.7871,
"step": 265
},
{
"epoch": 0.5510204081632653,
"grad_norm": 152509.4375,
"learning_rate": 2.7551020408163266e-06,
"loss": 1.8054,
"step": 270
},
{
"epoch": 0.5612244897959183,
"grad_norm": 114681.078125,
"learning_rate": 2.8061224489795917e-06,
"loss": 1.7711,
"step": 275
},
{
"epoch": 0.5714285714285714,
"grad_norm": 119104.984375,
"learning_rate": 2.8571428571428573e-06,
"loss": 1.7139,
"step": 280
},
{
"epoch": 0.5816326530612245,
"grad_norm": 123160.5,
"learning_rate": 2.908163265306123e-06,
"loss": 1.703,
"step": 285
},
{
"epoch": 0.5918367346938775,
"grad_norm": 157218.203125,
"learning_rate": 2.959183673469388e-06,
"loss": 1.7685,
"step": 290
},
{
"epoch": 0.6020408163265306,
"grad_norm": 134804.828125,
"learning_rate": 3.0102040816326534e-06,
"loss": 1.7053,
"step": 295
},
{
"epoch": 0.6122448979591837,
"grad_norm": 153977.625,
"learning_rate": 3.0612244897959185e-06,
"loss": 1.6711,
"step": 300
},
{
"epoch": 0.6224489795918368,
"grad_norm": 204690.125,
"learning_rate": 3.112244897959184e-06,
"loss": 1.5873,
"step": 305
},
{
"epoch": 0.6326530612244898,
"grad_norm": 158604.1875,
"learning_rate": 3.1632653061224496e-06,
"loss": 1.6642,
"step": 310
},
{
"epoch": 0.6428571428571429,
"grad_norm": 181877.484375,
"learning_rate": 3.2142857142857147e-06,
"loss": 1.6204,
"step": 315
},
{
"epoch": 0.6530612244897959,
"grad_norm": 157735.734375,
"learning_rate": 3.2653061224489794e-06,
"loss": 1.6152,
"step": 320
},
{
"epoch": 0.6632653061224489,
"grad_norm": 121935.3828125,
"learning_rate": 3.316326530612245e-06,
"loss": 1.6357,
"step": 325
},
{
"epoch": 0.673469387755102,
"grad_norm": 133579.859375,
"learning_rate": 3.3673469387755105e-06,
"loss": 1.6085,
"step": 330
},
{
"epoch": 0.6836734693877551,
"grad_norm": 154886.953125,
"learning_rate": 3.4183673469387756e-06,
"loss": 1.5609,
"step": 335
},
{
"epoch": 0.6938775510204082,
"grad_norm": 151655.0625,
"learning_rate": 3.469387755102041e-06,
"loss": 1.5905,
"step": 340
},
{
"epoch": 0.7040816326530612,
"grad_norm": 198471.609375,
"learning_rate": 3.5204081632653062e-06,
"loss": 1.5052,
"step": 345
},
{
"epoch": 0.7142857142857143,
"grad_norm": 191516.453125,
"learning_rate": 3.5714285714285718e-06,
"loss": 1.5798,
"step": 350
},
{
"epoch": 0.7244897959183674,
"grad_norm": 207511.296875,
"learning_rate": 3.6224489795918373e-06,
"loss": 1.5382,
"step": 355
},
{
"epoch": 0.7346938775510204,
"grad_norm": 170820.609375,
"learning_rate": 3.6734693877551024e-06,
"loss": 1.485,
"step": 360
},
{
"epoch": 0.7448979591836735,
"grad_norm": 183925.796875,
"learning_rate": 3.724489795918368e-06,
"loss": 1.427,
"step": 365
},
{
"epoch": 0.7551020408163265,
"grad_norm": 302829.75,
"learning_rate": 3.7755102040816327e-06,
"loss": 1.4242,
"step": 370
},
{
"epoch": 0.7653061224489796,
"grad_norm": 292990.5625,
"learning_rate": 3.826530612244898e-06,
"loss": 1.484,
"step": 375
},
{
"epoch": 0.7755102040816326,
"grad_norm": 189517.421875,
"learning_rate": 3.877551020408164e-06,
"loss": 1.5039,
"step": 380
},
{
"epoch": 0.7857142857142857,
"grad_norm": 239271.890625,
"learning_rate": 3.928571428571429e-06,
"loss": 1.3332,
"step": 385
},
{
"epoch": 0.7959183673469388,
"grad_norm": 167386.28125,
"learning_rate": 3.979591836734694e-06,
"loss": 1.429,
"step": 390
},
{
"epoch": 0.8061224489795918,
"grad_norm": 165701.9375,
"learning_rate": 4.03061224489796e-06,
"loss": 1.3167,
"step": 395
},
{
"epoch": 0.8163265306122449,
"grad_norm": 284793.46875,
"learning_rate": 4.081632653061225e-06,
"loss": 1.3867,
"step": 400
},
{
"epoch": 0.826530612244898,
"grad_norm": 159807.671875,
"learning_rate": 4.13265306122449e-06,
"loss": 1.404,
"step": 405
},
{
"epoch": 0.8367346938775511,
"grad_norm": 225698.859375,
"learning_rate": 4.183673469387755e-06,
"loss": 1.3842,
"step": 410
},
{
"epoch": 0.8469387755102041,
"grad_norm": 148945.90625,
"learning_rate": 4.234693877551021e-06,
"loss": 1.4605,
"step": 415
},
{
"epoch": 0.8571428571428571,
"grad_norm": 301350.15625,
"learning_rate": 4.2857142857142855e-06,
"loss": 1.4279,
"step": 420
},
{
"epoch": 0.8673469387755102,
"grad_norm": 216519.640625,
"learning_rate": 4.336734693877551e-06,
"loss": 1.3122,
"step": 425
},
{
"epoch": 0.8775510204081632,
"grad_norm": 243937.1875,
"learning_rate": 4.3877551020408165e-06,
"loss": 1.2084,
"step": 430
},
{
"epoch": 0.8877551020408163,
"grad_norm": 195260.390625,
"learning_rate": 4.438775510204082e-06,
"loss": 1.3024,
"step": 435
},
{
"epoch": 0.8979591836734694,
"grad_norm": 218123.65625,
"learning_rate": 4.489795918367348e-06,
"loss": 1.2303,
"step": 440
},
{
"epoch": 0.9081632653061225,
"grad_norm": 343990.03125,
"learning_rate": 4.540816326530613e-06,
"loss": 1.2012,
"step": 445
},
{
"epoch": 0.9183673469387755,
"grad_norm": 157196.390625,
"learning_rate": 4.591836734693878e-06,
"loss": 1.2632,
"step": 450
},
{
"epoch": 0.9285714285714286,
"grad_norm": 238103.46875,
"learning_rate": 4.642857142857144e-06,
"loss": 1.3131,
"step": 455
},
{
"epoch": 0.9387755102040817,
"grad_norm": 158442.625,
"learning_rate": 4.693877551020409e-06,
"loss": 1.2849,
"step": 460
},
{
"epoch": 0.9489795918367347,
"grad_norm": 178316.640625,
"learning_rate": 4.744897959183674e-06,
"loss": 1.0987,
"step": 465
},
{
"epoch": 0.9591836734693877,
"grad_norm": 240848.03125,
"learning_rate": 4.795918367346939e-06,
"loss": 1.154,
"step": 470
},
{
"epoch": 0.9693877551020408,
"grad_norm": 677730.8125,
"learning_rate": 4.846938775510204e-06,
"loss": 1.1682,
"step": 475
},
{
"epoch": 0.9795918367346939,
"grad_norm": 300622.875,
"learning_rate": 4.897959183673469e-06,
"loss": 1.1212,
"step": 480
},
{
"epoch": 0.9897959183673469,
"grad_norm": 234745.03125,
"learning_rate": 4.948979591836735e-06,
"loss": 1.1859,
"step": 485
},
{
"epoch": 1.0,
"grad_norm": 254966.984375,
"learning_rate": 5e-06,
"loss": 1.1763,
"step": 490
},
{
"epoch": 1.0,
"eval_accuracy": 0.8883299798792756,
"eval_loss": 0.968666136264801,
"eval_runtime": 208.5533,
"eval_samples_per_second": 14.299,
"eval_steps_per_second": 0.599,
"step": 490
},
{
"epoch": 1.010204081632653,
"grad_norm": 227501.953125,
"learning_rate": 4.994331065759638e-06,
"loss": 1.1466,
"step": 495
},
{
"epoch": 1.0204081632653061,
"grad_norm": 184310.828125,
"learning_rate": 4.9886621315192745e-06,
"loss": 1.0584,
"step": 500
},
{
"epoch": 1.030612244897959,
"grad_norm": 568293.25,
"learning_rate": 4.982993197278912e-06,
"loss": 1.0574,
"step": 505
},
{
"epoch": 1.0408163265306123,
"grad_norm": 208197.4375,
"learning_rate": 4.9773242630385495e-06,
"loss": 1.0964,
"step": 510
},
{
"epoch": 1.0510204081632653,
"grad_norm": 302428.34375,
"learning_rate": 4.971655328798186e-06,
"loss": 1.1017,
"step": 515
},
{
"epoch": 1.0612244897959184,
"grad_norm": 331227.1875,
"learning_rate": 4.965986394557824e-06,
"loss": 1.0954,
"step": 520
},
{
"epoch": 1.0714285714285714,
"grad_norm": 525813.5625,
"learning_rate": 4.960317460317461e-06,
"loss": 1.061,
"step": 525
},
{
"epoch": 1.0816326530612246,
"grad_norm": 578634.125,
"learning_rate": 4.954648526077098e-06,
"loss": 1.135,
"step": 530
},
{
"epoch": 1.0918367346938775,
"grad_norm": 208531.90625,
"learning_rate": 4.948979591836735e-06,
"loss": 0.9104,
"step": 535
},
{
"epoch": 1.1020408163265305,
"grad_norm": 621899.75,
"learning_rate": 4.943310657596373e-06,
"loss": 1.0909,
"step": 540
},
{
"epoch": 1.1122448979591837,
"grad_norm": 362684.3125,
"learning_rate": 4.9376417233560094e-06,
"loss": 0.8759,
"step": 545
},
{
"epoch": 1.1224489795918366,
"grad_norm": 274587.8125,
"learning_rate": 4.931972789115647e-06,
"loss": 0.9977,
"step": 550
},
{
"epoch": 1.1326530612244898,
"grad_norm": 379287.3125,
"learning_rate": 4.9263038548752836e-06,
"loss": 1.013,
"step": 555
},
{
"epoch": 1.1428571428571428,
"grad_norm": 277623.0,
"learning_rate": 4.920634920634921e-06,
"loss": 0.9425,
"step": 560
},
{
"epoch": 1.153061224489796,
"grad_norm": 341604.53125,
"learning_rate": 4.914965986394558e-06,
"loss": 0.9697,
"step": 565
},
{
"epoch": 1.163265306122449,
"grad_norm": 742070.625,
"learning_rate": 4.909297052154195e-06,
"loss": 1.1181,
"step": 570
},
{
"epoch": 1.1734693877551021,
"grad_norm": 225296.84375,
"learning_rate": 4.903628117913833e-06,
"loss": 0.9465,
"step": 575
},
{
"epoch": 1.183673469387755,
"grad_norm": 952885.3125,
"learning_rate": 4.897959183673469e-06,
"loss": 1.1135,
"step": 580
},
{
"epoch": 1.193877551020408,
"grad_norm": 527028.875,
"learning_rate": 4.892290249433107e-06,
"loss": 1.109,
"step": 585
},
{
"epoch": 1.2040816326530612,
"grad_norm": 243262.0625,
"learning_rate": 4.886621315192744e-06,
"loss": 0.9313,
"step": 590
},
{
"epoch": 1.2142857142857142,
"grad_norm": 287492.28125,
"learning_rate": 4.880952380952381e-06,
"loss": 0.9603,
"step": 595
},
{
"epoch": 1.2244897959183674,
"grad_norm": 400522.90625,
"learning_rate": 4.8752834467120185e-06,
"loss": 0.9105,
"step": 600
},
{
"epoch": 1.2346938775510203,
"grad_norm": 190966.40625,
"learning_rate": 4.869614512471656e-06,
"loss": 0.9654,
"step": 605
},
{
"epoch": 1.2448979591836735,
"grad_norm": 223976.515625,
"learning_rate": 4.863945578231293e-06,
"loss": 0.9511,
"step": 610
},
{
"epoch": 1.2551020408163265,
"grad_norm": 345172.09375,
"learning_rate": 4.85827664399093e-06,
"loss": 0.8092,
"step": 615
},
{
"epoch": 1.2653061224489797,
"grad_norm": 378395.75,
"learning_rate": 4.852607709750568e-06,
"loss": 0.9777,
"step": 620
},
{
"epoch": 1.2755102040816326,
"grad_norm": 408057.8125,
"learning_rate": 4.846938775510204e-06,
"loss": 1.006,
"step": 625
},
{
"epoch": 1.2857142857142856,
"grad_norm": 254345.203125,
"learning_rate": 4.841269841269842e-06,
"loss": 0.9107,
"step": 630
},
{
"epoch": 1.2959183673469388,
"grad_norm": 465929.21875,
"learning_rate": 4.835600907029479e-06,
"loss": 1.0237,
"step": 635
},
{
"epoch": 1.306122448979592,
"grad_norm": 665911.5625,
"learning_rate": 4.829931972789116e-06,
"loss": 0.886,
"step": 640
},
{
"epoch": 1.316326530612245,
"grad_norm": 541784.25,
"learning_rate": 4.824263038548753e-06,
"loss": 0.9266,
"step": 645
},
{
"epoch": 1.3265306122448979,
"grad_norm": 181121.25,
"learning_rate": 4.818594104308391e-06,
"loss": 0.9289,
"step": 650
},
{
"epoch": 1.336734693877551,
"grad_norm": 465088.90625,
"learning_rate": 4.8129251700680275e-06,
"loss": 0.8976,
"step": 655
},
{
"epoch": 1.346938775510204,
"grad_norm": 309529.28125,
"learning_rate": 4.807256235827665e-06,
"loss": 0.9182,
"step": 660
},
{
"epoch": 1.3571428571428572,
"grad_norm": 672721.125,
"learning_rate": 4.8015873015873025e-06,
"loss": 0.8919,
"step": 665
},
{
"epoch": 1.3673469387755102,
"grad_norm": 300983.96875,
"learning_rate": 4.795918367346939e-06,
"loss": 0.7491,
"step": 670
},
{
"epoch": 1.3775510204081631,
"grad_norm": 168024.203125,
"learning_rate": 4.790249433106577e-06,
"loss": 0.803,
"step": 675
},
{
"epoch": 1.3877551020408163,
"grad_norm": 252603.5,
"learning_rate": 4.784580498866213e-06,
"loss": 0.8481,
"step": 680
},
{
"epoch": 1.3979591836734695,
"grad_norm": 495847.71875,
"learning_rate": 4.778911564625851e-06,
"loss": 0.9796,
"step": 685
},
{
"epoch": 1.4081632653061225,
"grad_norm": 213929.046875,
"learning_rate": 4.773242630385488e-06,
"loss": 0.8509,
"step": 690
},
{
"epoch": 1.4183673469387754,
"grad_norm": 203737.90625,
"learning_rate": 4.767573696145125e-06,
"loss": 0.8133,
"step": 695
},
{
"epoch": 1.4285714285714286,
"grad_norm": 340089.6875,
"learning_rate": 4.761904761904762e-06,
"loss": 0.8628,
"step": 700
},
{
"epoch": 1.4387755102040816,
"grad_norm": 292163.90625,
"learning_rate": 4.756235827664399e-06,
"loss": 0.9341,
"step": 705
},
{
"epoch": 1.4489795918367347,
"grad_norm": 266707.96875,
"learning_rate": 4.7505668934240365e-06,
"loss": 1.0158,
"step": 710
},
{
"epoch": 1.4591836734693877,
"grad_norm": 389876.40625,
"learning_rate": 4.744897959183674e-06,
"loss": 0.9575,
"step": 715
},
{
"epoch": 1.469387755102041,
"grad_norm": 699534.0625,
"learning_rate": 4.739229024943311e-06,
"loss": 0.9293,
"step": 720
},
{
"epoch": 1.4795918367346939,
"grad_norm": 495432.5,
"learning_rate": 4.733560090702948e-06,
"loss": 0.7233,
"step": 725
},
{
"epoch": 1.489795918367347,
"grad_norm": 297580.71875,
"learning_rate": 4.727891156462586e-06,
"loss": 0.7624,
"step": 730
},
{
"epoch": 1.5,
"grad_norm": 214458.125,
"learning_rate": 4.722222222222222e-06,
"loss": 0.8704,
"step": 735
},
{
"epoch": 1.510204081632653,
"grad_norm": 483185.90625,
"learning_rate": 4.71655328798186e-06,
"loss": 0.7841,
"step": 740
},
{
"epoch": 1.5204081632653061,
"grad_norm": 279528.75,
"learning_rate": 4.710884353741497e-06,
"loss": 0.8634,
"step": 745
},
{
"epoch": 1.5306122448979593,
"grad_norm": 398520.15625,
"learning_rate": 4.705215419501134e-06,
"loss": 0.902,
"step": 750
},
{
"epoch": 1.5408163265306123,
"grad_norm": 343986.875,
"learning_rate": 4.699546485260771e-06,
"loss": 0.8896,
"step": 755
},
{
"epoch": 1.5510204081632653,
"grad_norm": 941463.0625,
"learning_rate": 4.693877551020409e-06,
"loss": 0.8483,
"step": 760
},
{
"epoch": 1.5612244897959182,
"grad_norm": 382110.125,
"learning_rate": 4.6882086167800455e-06,
"loss": 0.944,
"step": 765
},
{
"epoch": 1.5714285714285714,
"grad_norm": 321413.15625,
"learning_rate": 4.682539682539683e-06,
"loss": 0.8019,
"step": 770
},
{
"epoch": 1.5816326530612246,
"grad_norm": 404689.875,
"learning_rate": 4.6768707482993205e-06,
"loss": 0.7069,
"step": 775
},
{
"epoch": 1.5918367346938775,
"grad_norm": 653981.25,
"learning_rate": 4.671201814058957e-06,
"loss": 0.9424,
"step": 780
},
{
"epoch": 1.6020408163265305,
"grad_norm": 387232.34375,
"learning_rate": 4.665532879818595e-06,
"loss": 0.9593,
"step": 785
},
{
"epoch": 1.6122448979591837,
"grad_norm": 527808.625,
"learning_rate": 4.659863945578232e-06,
"loss": 0.8147,
"step": 790
},
{
"epoch": 1.6224489795918369,
"grad_norm": 350020.8125,
"learning_rate": 4.654195011337869e-06,
"loss": 1.0195,
"step": 795
},
{
"epoch": 1.6326530612244898,
"grad_norm": 510409.9375,
"learning_rate": 4.648526077097506e-06,
"loss": 0.865,
"step": 800
},
{
"epoch": 1.6428571428571428,
"grad_norm": 238838.65625,
"learning_rate": 4.642857142857144e-06,
"loss": 0.8115,
"step": 805
},
{
"epoch": 1.6530612244897958,
"grad_norm": 266616.84375,
"learning_rate": 4.6371882086167804e-06,
"loss": 0.8091,
"step": 810
},
{
"epoch": 1.663265306122449,
"grad_norm": 398578.84375,
"learning_rate": 4.631519274376418e-06,
"loss": 0.6485,
"step": 815
},
{
"epoch": 1.6734693877551021,
"grad_norm": 242011.765625,
"learning_rate": 4.6258503401360546e-06,
"loss": 0.8242,
"step": 820
},
{
"epoch": 1.683673469387755,
"grad_norm": 646366.6875,
"learning_rate": 4.620181405895692e-06,
"loss": 0.7616,
"step": 825
},
{
"epoch": 1.693877551020408,
"grad_norm": 447861.9375,
"learning_rate": 4.614512471655329e-06,
"loss": 0.8694,
"step": 830
},
{
"epoch": 1.7040816326530612,
"grad_norm": 422529.3125,
"learning_rate": 4.608843537414966e-06,
"loss": 0.9344,
"step": 835
},
{
"epoch": 1.7142857142857144,
"grad_norm": 448338.375,
"learning_rate": 4.603174603174604e-06,
"loss": 0.8459,
"step": 840
},
{
"epoch": 1.7244897959183674,
"grad_norm": 528423.75,
"learning_rate": 4.59750566893424e-06,
"loss": 0.8312,
"step": 845
},
{
"epoch": 1.7346938775510203,
"grad_norm": 422879.84375,
"learning_rate": 4.591836734693878e-06,
"loss": 0.8151,
"step": 850
},
{
"epoch": 1.7448979591836735,
"grad_norm": 543066.5,
"learning_rate": 4.5861678004535145e-06,
"loss": 0.8013,
"step": 855
},
{
"epoch": 1.7551020408163265,
"grad_norm": 362556.6875,
"learning_rate": 4.580498866213152e-06,
"loss": 0.9636,
"step": 860
},
{
"epoch": 1.7653061224489797,
"grad_norm": 305676.09375,
"learning_rate": 4.5748299319727895e-06,
"loss": 0.8208,
"step": 865
},
{
"epoch": 1.7755102040816326,
"grad_norm": 477851.96875,
"learning_rate": 4.569160997732427e-06,
"loss": 0.8109,
"step": 870
},
{
"epoch": 1.7857142857142856,
"grad_norm": 380819.1875,
"learning_rate": 4.563492063492064e-06,
"loss": 0.8019,
"step": 875
},
{
"epoch": 1.7959183673469388,
"grad_norm": 276121.21875,
"learning_rate": 4.557823129251701e-06,
"loss": 0.7365,
"step": 880
},
{
"epoch": 1.806122448979592,
"grad_norm": 807004.5625,
"learning_rate": 4.5521541950113386e-06,
"loss": 0.8284,
"step": 885
},
{
"epoch": 1.816326530612245,
"grad_norm": 612872.5625,
"learning_rate": 4.546485260770975e-06,
"loss": 0.9153,
"step": 890
},
{
"epoch": 1.8265306122448979,
"grad_norm": 384264.40625,
"learning_rate": 4.540816326530613e-06,
"loss": 0.8697,
"step": 895
},
{
"epoch": 1.836734693877551,
"grad_norm": 323550.15625,
"learning_rate": 4.53514739229025e-06,
"loss": 0.816,
"step": 900
},
{
"epoch": 1.8469387755102042,
"grad_norm": 588055.625,
"learning_rate": 4.529478458049887e-06,
"loss": 0.9765,
"step": 905
},
{
"epoch": 1.8571428571428572,
"grad_norm": 1386326.375,
"learning_rate": 4.523809523809524e-06,
"loss": 0.8051,
"step": 910
},
{
"epoch": 1.8673469387755102,
"grad_norm": 351753.28125,
"learning_rate": 4.518140589569162e-06,
"loss": 0.8377,
"step": 915
},
{
"epoch": 1.8775510204081631,
"grad_norm": 547828.25,
"learning_rate": 4.5124716553287985e-06,
"loss": 0.6776,
"step": 920
},
{
"epoch": 1.8877551020408163,
"grad_norm": 332551.40625,
"learning_rate": 4.506802721088436e-06,
"loss": 0.9434,
"step": 925
},
{
"epoch": 1.8979591836734695,
"grad_norm": 597800.125,
"learning_rate": 4.5011337868480735e-06,
"loss": 0.6165,
"step": 930
},
{
"epoch": 1.9081632653061225,
"grad_norm": 598426.1875,
"learning_rate": 4.49546485260771e-06,
"loss": 0.9668,
"step": 935
},
{
"epoch": 1.9183673469387754,
"grad_norm": 508502.5,
"learning_rate": 4.489795918367348e-06,
"loss": 0.8707,
"step": 940
},
{
"epoch": 1.9285714285714286,
"grad_norm": 825221.0625,
"learning_rate": 4.484126984126984e-06,
"loss": 0.8437,
"step": 945
},
{
"epoch": 1.9387755102040818,
"grad_norm": 674555.8125,
"learning_rate": 4.478458049886622e-06,
"loss": 0.7054,
"step": 950
},
{
"epoch": 1.9489795918367347,
"grad_norm": 547481.0,
"learning_rate": 4.472789115646259e-06,
"loss": 0.7896,
"step": 955
},
{
"epoch": 1.9591836734693877,
"grad_norm": 869727.6875,
"learning_rate": 4.467120181405896e-06,
"loss": 0.8161,
"step": 960
},
{
"epoch": 1.9693877551020407,
"grad_norm": 517315.71875,
"learning_rate": 4.461451247165533e-06,
"loss": 0.8131,
"step": 965
},
{
"epoch": 1.9795918367346939,
"grad_norm": 651501.0625,
"learning_rate": 4.45578231292517e-06,
"loss": 0.6217,
"step": 970
},
{
"epoch": 1.989795918367347,
"grad_norm": 559789.5625,
"learning_rate": 4.4501133786848075e-06,
"loss": 0.7849,
"step": 975
},
{
"epoch": 2.0,
"grad_norm": 586872.375,
"learning_rate": 4.444444444444444e-06,
"loss": 0.7778,
"step": 980
},
{
"epoch": 2.0,
"eval_accuracy": 0.8779342723004695,
"eval_loss": 0.5649504661560059,
"eval_runtime": 182.7428,
"eval_samples_per_second": 16.318,
"eval_steps_per_second": 0.684,
"step": 980
}
],
"logging_steps": 5,
"max_steps": 4900,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2.99358138673152e+18,
"train_batch_size": 24,
"trial_name": null,
"trial_params": null
}