|
{ |
|
"best_metric": 1.0, |
|
"best_model_checkpoint": "./models/checkpoint-1003", |
|
"epoch": 97.0, |
|
"global_step": 97291, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.98e-05, |
|
"loss": 6.8202, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 3.5011327266693115, |
|
"eval_runtime": 260.4478, |
|
"eval_samples_per_second": 17.735, |
|
"eval_steps_per_second": 0.372, |
|
"eval_wer": 1.0, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 9.899194360523666e-05, |
|
"loss": 1.0758, |
|
"step": 2006 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.5050923228263855, |
|
"eval_runtime": 330.4397, |
|
"eval_samples_per_second": 13.978, |
|
"eval_steps_per_second": 0.294, |
|
"eval_wer": 0.4667033908679153, |
|
"step": 2006 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 9.798187311178248e-05, |
|
"loss": 0.5297, |
|
"step": 3009 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.3798016309738159, |
|
"eval_runtime": 325.2274, |
|
"eval_samples_per_second": 14.202, |
|
"eval_steps_per_second": 0.298, |
|
"eval_wer": 0.39802913453299055, |
|
"step": 3009 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 9.69718026183283e-05, |
|
"loss": 0.4266, |
|
"step": 4012 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.32800063490867615, |
|
"eval_runtime": 361.1443, |
|
"eval_samples_per_second": 12.79, |
|
"eval_steps_per_second": 0.269, |
|
"eval_wer": 0.36733382298934997, |
|
"step": 4012 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 9.596173212487412e-05, |
|
"loss": 0.3668, |
|
"step": 5015 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.31621453166007996, |
|
"eval_runtime": 367.1388, |
|
"eval_samples_per_second": 12.581, |
|
"eval_steps_per_second": 0.264, |
|
"eval_wer": 0.3481760313379851, |
|
"step": 5015 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 9.495166163141995e-05, |
|
"loss": 0.326, |
|
"step": 6018 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 0.29389986395835876, |
|
"eval_runtime": 416.389, |
|
"eval_samples_per_second": 11.093, |
|
"eval_steps_per_second": 0.233, |
|
"eval_wer": 0.33691394295507404, |
|
"step": 6018 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 9.394159113796577e-05, |
|
"loss": 0.2961, |
|
"step": 7021 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 0.28012460470199585, |
|
"eval_runtime": 270.2817, |
|
"eval_samples_per_second": 17.09, |
|
"eval_steps_per_second": 0.359, |
|
"eval_wer": 0.3236320235034888, |
|
"step": 7021 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 9.293152064451159e-05, |
|
"loss": 0.2751, |
|
"step": 8024 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 0.2852382957935333, |
|
"eval_runtime": 304.5141, |
|
"eval_samples_per_second": 15.168, |
|
"eval_steps_per_second": 0.319, |
|
"eval_wer": 0.31775615130370916, |
|
"step": 8024 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 9.192145015105741e-05, |
|
"loss": 0.2561, |
|
"step": 9027 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 0.26253288984298706, |
|
"eval_runtime": 284.9667, |
|
"eval_samples_per_second": 16.209, |
|
"eval_steps_per_second": 0.34, |
|
"eval_wer": 0.30986044803525525, |
|
"step": 9027 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 9.091137965760322e-05, |
|
"loss": 0.2432, |
|
"step": 10030 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 0.2666085958480835, |
|
"eval_runtime": 406.9804, |
|
"eval_samples_per_second": 11.349, |
|
"eval_steps_per_second": 0.238, |
|
"eval_wer": 0.30536173338229894, |
|
"step": 10030 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 8.990130916414905e-05, |
|
"loss": 0.2242, |
|
"step": 11033 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_loss": 0.269761860370636, |
|
"eval_runtime": 400.3745, |
|
"eval_samples_per_second": 11.537, |
|
"eval_steps_per_second": 0.242, |
|
"eval_wer": 0.30089362223038313, |
|
"step": 11033 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 8.889123867069486e-05, |
|
"loss": 0.2153, |
|
"step": 12036 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_loss": 0.2758134603500366, |
|
"eval_runtime": 389.1981, |
|
"eval_samples_per_second": 11.868, |
|
"eval_steps_per_second": 0.249, |
|
"eval_wer": 0.30067939772309954, |
|
"step": 12036 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 8.78821752265861e-05, |
|
"loss": 0.2036, |
|
"step": 13039 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_loss": 0.2647706866264343, |
|
"eval_runtime": 436.7457, |
|
"eval_samples_per_second": 10.576, |
|
"eval_steps_per_second": 0.222, |
|
"eval_wer": 0.29743542661280453, |
|
"step": 13039 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 8.687210473313193e-05, |
|
"loss": 0.1972, |
|
"step": 14042 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_loss": 0.2667146325111389, |
|
"eval_runtime": 402.5926, |
|
"eval_samples_per_second": 11.473, |
|
"eval_steps_per_second": 0.241, |
|
"eval_wer": 0.29238584894111885, |
|
"step": 14042 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 8.586203423967774e-05, |
|
"loss": 0.186, |
|
"step": 15045 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_loss": 0.2569558024406433, |
|
"eval_runtime": 397.1089, |
|
"eval_samples_per_second": 11.632, |
|
"eval_steps_per_second": 0.244, |
|
"eval_wer": 0.29097808789325497, |
|
"step": 15045 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 8.485196374622358e-05, |
|
"loss": 0.179, |
|
"step": 16048 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_loss": 0.270986944437027, |
|
"eval_runtime": 443.1466, |
|
"eval_samples_per_second": 10.423, |
|
"eval_steps_per_second": 0.219, |
|
"eval_wer": 0.289386705839148, |
|
"step": 16048 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 8.384390735146023e-05, |
|
"loss": 0.1721, |
|
"step": 17051 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_loss": 0.2602141797542572, |
|
"eval_runtime": 438.2891, |
|
"eval_samples_per_second": 10.539, |
|
"eval_steps_per_second": 0.221, |
|
"eval_wer": 0.2873056677683927, |
|
"step": 17051 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 8.283383685800604e-05, |
|
"loss": 0.1656, |
|
"step": 18054 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_loss": 0.26877614855766296, |
|
"eval_runtime": 441.5132, |
|
"eval_samples_per_second": 10.462, |
|
"eval_steps_per_second": 0.22, |
|
"eval_wer": 0.28403109315705716, |
|
"step": 18054 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"learning_rate": 8.182376636455186e-05, |
|
"loss": 0.1613, |
|
"step": 19057 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_loss": 0.28074949979782104, |
|
"eval_runtime": 481.2314, |
|
"eval_samples_per_second": 9.598, |
|
"eval_steps_per_second": 0.202, |
|
"eval_wer": 0.28369445464561144, |
|
"step": 19057 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"learning_rate": 8.081369587109768e-05, |
|
"loss": 0.1546, |
|
"step": 20060 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_loss": 0.26929938793182373, |
|
"eval_runtime": 412.4546, |
|
"eval_samples_per_second": 11.199, |
|
"eval_steps_per_second": 0.235, |
|
"eval_wer": 0.2782776349614396, |
|
"step": 20060 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"learning_rate": 7.980463242698892e-05, |
|
"loss": 0.1506, |
|
"step": 21063 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_loss": 0.2805509865283966, |
|
"eval_runtime": 402.9001, |
|
"eval_samples_per_second": 11.464, |
|
"eval_steps_per_second": 0.241, |
|
"eval_wer": 0.28161341657485617, |
|
"step": 21063 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"learning_rate": 7.879456193353474e-05, |
|
"loss": 0.1459, |
|
"step": 22066 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_loss": 0.28382739424705505, |
|
"eval_runtime": 393.1186, |
|
"eval_samples_per_second": 11.75, |
|
"eval_steps_per_second": 0.247, |
|
"eval_wer": 0.27885910148120946, |
|
"step": 22066 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"learning_rate": 7.778449144008056e-05, |
|
"loss": 0.14, |
|
"step": 23069 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_loss": 0.2746136486530304, |
|
"eval_runtime": 425.1655, |
|
"eval_samples_per_second": 10.864, |
|
"eval_steps_per_second": 0.228, |
|
"eval_wer": 0.2754315093646713, |
|
"step": 23069 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"learning_rate": 7.677442094662639e-05, |
|
"loss": 0.1367, |
|
"step": 24072 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_loss": 0.2868611514568329, |
|
"eval_runtime": 427.1205, |
|
"eval_samples_per_second": 10.814, |
|
"eval_steps_per_second": 0.227, |
|
"eval_wer": 0.27466642183865836, |
|
"step": 24072 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"learning_rate": 7.576435045317221e-05, |
|
"loss": 0.1303, |
|
"step": 25075 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_loss": 0.26962676644325256, |
|
"eval_runtime": 414.5701, |
|
"eval_samples_per_second": 11.142, |
|
"eval_steps_per_second": 0.234, |
|
"eval_wer": 0.27371771330640227, |
|
"step": 25075 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"learning_rate": 7.475528700906345e-05, |
|
"loss": 0.1299, |
|
"step": 26078 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_loss": 0.2933480143547058, |
|
"eval_runtime": 423.3941, |
|
"eval_samples_per_second": 10.909, |
|
"eval_steps_per_second": 0.229, |
|
"eval_wer": 0.27429917982617213, |
|
"step": 26078 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"learning_rate": 7.374521651560927e-05, |
|
"loss": 0.1241, |
|
"step": 27081 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_loss": 0.2790047824382782, |
|
"eval_runtime": 413.5975, |
|
"eval_samples_per_second": 11.168, |
|
"eval_steps_per_second": 0.235, |
|
"eval_wer": 0.27019831068674255, |
|
"step": 27081 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"learning_rate": 7.273615307150052e-05, |
|
"loss": 0.122, |
|
"step": 28084 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_loss": 0.29522669315338135, |
|
"eval_runtime": 397.9723, |
|
"eval_samples_per_second": 11.606, |
|
"eval_steps_per_second": 0.244, |
|
"eval_wer": 0.27243236626270045, |
|
"step": 28084 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"learning_rate": 7.172708962739174e-05, |
|
"loss": 0.1195, |
|
"step": 29087 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_loss": 0.2860444486141205, |
|
"eval_runtime": 401.6884, |
|
"eval_samples_per_second": 11.499, |
|
"eval_steps_per_second": 0.241, |
|
"eval_wer": 0.26967805116905375, |
|
"step": 29087 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"learning_rate": 7.071701913393757e-05, |
|
"loss": 0.1171, |
|
"step": 30090 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_loss": 0.30656594038009644, |
|
"eval_runtime": 400.5872, |
|
"eval_samples_per_second": 11.531, |
|
"eval_steps_per_second": 0.242, |
|
"eval_wer": 0.27148365773044436, |
|
"step": 30090 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"learning_rate": 6.970694864048339e-05, |
|
"loss": 0.1131, |
|
"step": 31093 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_loss": 0.3087522089481354, |
|
"eval_runtime": 410.4211, |
|
"eval_samples_per_second": 11.254, |
|
"eval_steps_per_second": 0.236, |
|
"eval_wer": 0.26646468355979924, |
|
"step": 31093 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"learning_rate": 6.869687814702921e-05, |
|
"loss": 0.1111, |
|
"step": 32096 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_loss": 0.305497407913208, |
|
"eval_runtime": 397.2313, |
|
"eval_samples_per_second": 11.628, |
|
"eval_steps_per_second": 0.244, |
|
"eval_wer": 0.26823968662014935, |
|
"step": 32096 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"learning_rate": 6.768680765357503e-05, |
|
"loss": 0.1081, |
|
"step": 33099 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_loss": 0.3094805181026459, |
|
"eval_runtime": 398.4287, |
|
"eval_samples_per_second": 11.593, |
|
"eval_steps_per_second": 0.243, |
|
"eval_wer": 0.2667401150691639, |
|
"step": 33099 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"learning_rate": 6.667673716012085e-05, |
|
"loss": 0.106, |
|
"step": 34102 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_loss": 0.31470388174057007, |
|
"eval_runtime": 394.7251, |
|
"eval_samples_per_second": 11.702, |
|
"eval_steps_per_second": 0.246, |
|
"eval_wer": 0.2654241645244216, |
|
"step": 34102 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"learning_rate": 6.566666666666666e-05, |
|
"loss": 0.1023, |
|
"step": 35105 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_loss": 0.32143208384513855, |
|
"eval_runtime": 406.2767, |
|
"eval_samples_per_second": 11.369, |
|
"eval_steps_per_second": 0.239, |
|
"eval_wer": 0.2643530419880034, |
|
"step": 35105 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"learning_rate": 6.46565961732125e-05, |
|
"loss": 0.1031, |
|
"step": 36108 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_loss": 0.3046073317527771, |
|
"eval_runtime": 414.3275, |
|
"eval_samples_per_second": 11.148, |
|
"eval_steps_per_second": 0.234, |
|
"eval_wer": 0.26306769494430166, |
|
"step": 36108 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"learning_rate": 6.364853977844915e-05, |
|
"loss": 0.1008, |
|
"step": 37111 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_loss": 0.3285815715789795, |
|
"eval_runtime": 474.4359, |
|
"eval_samples_per_second": 9.736, |
|
"eval_steps_per_second": 0.204, |
|
"eval_wer": 0.2653323540213, |
|
"step": 37111 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"learning_rate": 6.263846928499497e-05, |
|
"loss": 0.0977, |
|
"step": 38114 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_loss": 0.2976115047931671, |
|
"eval_runtime": 433.6527, |
|
"eval_samples_per_second": 10.651, |
|
"eval_steps_per_second": 0.224, |
|
"eval_wer": 0.2623638144203697, |
|
"step": 38114 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"learning_rate": 6.162839879154079e-05, |
|
"loss": 0.0951, |
|
"step": 39117 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_loss": 0.3340004086494446, |
|
"eval_runtime": 479.8072, |
|
"eval_samples_per_second": 9.627, |
|
"eval_steps_per_second": 0.202, |
|
"eval_wer": 0.26282286693597745, |
|
"step": 39117 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"learning_rate": 6.0618328298086614e-05, |
|
"loss": 0.0923, |
|
"step": 40120 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_loss": 0.3238993287086487, |
|
"eval_runtime": 375.954, |
|
"eval_samples_per_second": 12.286, |
|
"eval_steps_per_second": 0.258, |
|
"eval_wer": 0.26407761047863876, |
|
"step": 40120 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"learning_rate": 5.960926485397785e-05, |
|
"loss": 0.0931, |
|
"step": 41123 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"eval_loss": 0.3135124444961548, |
|
"eval_runtime": 457.0228, |
|
"eval_samples_per_second": 10.107, |
|
"eval_steps_per_second": 0.212, |
|
"eval_wer": 0.2601909658464928, |
|
"step": 41123 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"learning_rate": 5.860020140986908e-05, |
|
"loss": 0.0902, |
|
"step": 42126 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"eval_loss": 0.3057383894920349, |
|
"eval_runtime": 459.7321, |
|
"eval_samples_per_second": 10.047, |
|
"eval_steps_per_second": 0.211, |
|
"eval_wer": 0.25957889582568244, |
|
"step": 42126 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"learning_rate": 5.759013091641491e-05, |
|
"loss": 0.0899, |
|
"step": 43129 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"eval_loss": 0.3292476534843445, |
|
"eval_runtime": 439.0502, |
|
"eval_samples_per_second": 10.52, |
|
"eval_steps_per_second": 0.221, |
|
"eval_wer": 0.26068062186314117, |
|
"step": 43129 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"learning_rate": 5.658106747230615e-05, |
|
"loss": 0.0873, |
|
"step": 44132 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"eval_loss": 0.3199455142021179, |
|
"eval_runtime": 360.0807, |
|
"eval_samples_per_second": 12.828, |
|
"eval_steps_per_second": 0.269, |
|
"eval_wer": 0.2638633859713551, |
|
"step": 44132 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"learning_rate": 5.5570996978851965e-05, |
|
"loss": 0.0856, |
|
"step": 45135 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"eval_loss": 0.33823832869529724, |
|
"eval_runtime": 343.4666, |
|
"eval_samples_per_second": 13.448, |
|
"eval_steps_per_second": 0.282, |
|
"eval_wer": 0.2601909658464928, |
|
"step": 45135 |
|
}, |
|
{ |
|
"epoch": 46.0, |
|
"learning_rate": 5.45619335347432e-05, |
|
"loss": 0.084, |
|
"step": 46138 |
|
}, |
|
{ |
|
"epoch": 46.0, |
|
"eval_loss": 0.33022576570510864, |
|
"eval_runtime": 342.0224, |
|
"eval_samples_per_second": 13.505, |
|
"eval_steps_per_second": 0.284, |
|
"eval_wer": 0.25795691027053497, |
|
"step": 46138 |
|
}, |
|
{ |
|
"epoch": 47.0, |
|
"learning_rate": 5.355186304128903e-05, |
|
"loss": 0.0815, |
|
"step": 47141 |
|
}, |
|
{ |
|
"epoch": 47.0, |
|
"eval_loss": 0.32415875792503357, |
|
"eval_runtime": 344.3556, |
|
"eval_samples_per_second": 13.413, |
|
"eval_steps_per_second": 0.282, |
|
"eval_wer": 0.26098665687354633, |
|
"step": 47141 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"learning_rate": 5.2541792547834846e-05, |
|
"loss": 0.0815, |
|
"step": 48144 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"eval_loss": 0.3343857526779175, |
|
"eval_runtime": 343.7802, |
|
"eval_samples_per_second": 13.436, |
|
"eval_steps_per_second": 0.282, |
|
"eval_wer": 0.2582629452809401, |
|
"step": 48144 |
|
}, |
|
{ |
|
"epoch": 49.0, |
|
"learning_rate": 5.1532729103726084e-05, |
|
"loss": 0.0785, |
|
"step": 49147 |
|
}, |
|
{ |
|
"epoch": 49.0, |
|
"eval_loss": 0.34900155663490295, |
|
"eval_runtime": 343.1634, |
|
"eval_samples_per_second": 13.46, |
|
"eval_steps_per_second": 0.283, |
|
"eval_wer": 0.25798751377157547, |
|
"step": 49147 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"learning_rate": 5.052265861027191e-05, |
|
"loss": 0.0777, |
|
"step": 50150 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"eval_loss": 0.34195998311042786, |
|
"eval_runtime": 299.1953, |
|
"eval_samples_per_second": 15.438, |
|
"eval_steps_per_second": 0.324, |
|
"eval_wer": 0.2567633737299547, |
|
"step": 50150 |
|
}, |
|
{ |
|
"epoch": 51.0, |
|
"learning_rate": 4.951258811681773e-05, |
|
"loss": 0.0769, |
|
"step": 51153 |
|
}, |
|
{ |
|
"epoch": 51.0, |
|
"eval_loss": 0.33568263053894043, |
|
"eval_runtime": 342.9594, |
|
"eval_samples_per_second": 13.468, |
|
"eval_steps_per_second": 0.283, |
|
"eval_wer": 0.25716121924348145, |
|
"step": 51153 |
|
}, |
|
{ |
|
"epoch": 52.0, |
|
"learning_rate": 4.8503524672708965e-05, |
|
"loss": 0.0775, |
|
"step": 52156 |
|
}, |
|
{ |
|
"epoch": 52.0, |
|
"eval_loss": 0.32406118512153625, |
|
"eval_runtime": 315.6469, |
|
"eval_samples_per_second": 14.633, |
|
"eval_steps_per_second": 0.307, |
|
"eval_wer": 0.2555698371893745, |
|
"step": 52156 |
|
}, |
|
{ |
|
"epoch": 53.0, |
|
"learning_rate": 4.749345417925479e-05, |
|
"loss": 0.0751, |
|
"step": 53159 |
|
}, |
|
{ |
|
"epoch": 53.0, |
|
"eval_loss": 0.3319285213947296, |
|
"eval_runtime": 350.258, |
|
"eval_samples_per_second": 13.187, |
|
"eval_steps_per_second": 0.277, |
|
"eval_wer": 0.25670216672787366, |
|
"step": 53159 |
|
}, |
|
{ |
|
"epoch": 54.0, |
|
"learning_rate": 4.6484390735146025e-05, |
|
"loss": 0.0745, |
|
"step": 54162 |
|
}, |
|
{ |
|
"epoch": 54.0, |
|
"eval_loss": 0.34319552779197693, |
|
"eval_runtime": 461.2436, |
|
"eval_samples_per_second": 10.014, |
|
"eval_steps_per_second": 0.21, |
|
"eval_wer": 0.25440690414983474, |
|
"step": 54162 |
|
}, |
|
{ |
|
"epoch": 55.0, |
|
"learning_rate": 4.547432024169184e-05, |
|
"loss": 0.0731, |
|
"step": 55165 |
|
}, |
|
{ |
|
"epoch": 55.0, |
|
"eval_loss": 0.340921550989151, |
|
"eval_runtime": 375.608, |
|
"eval_samples_per_second": 12.297, |
|
"eval_steps_per_second": 0.258, |
|
"eval_wer": 0.25575345819561757, |
|
"step": 55165 |
|
}, |
|
{ |
|
"epoch": 56.0, |
|
"learning_rate": 4.446424974823767e-05, |
|
"loss": 0.0725, |
|
"step": 56168 |
|
}, |
|
{ |
|
"epoch": 56.0, |
|
"eval_loss": 0.35223615169525146, |
|
"eval_runtime": 356.2173, |
|
"eval_samples_per_second": 12.967, |
|
"eval_steps_per_second": 0.272, |
|
"eval_wer": 0.255661647692496, |
|
"step": 56168 |
|
}, |
|
{ |
|
"epoch": 57.0, |
|
"learning_rate": 4.345417925478349e-05, |
|
"loss": 0.0712, |
|
"step": 57171 |
|
}, |
|
{ |
|
"epoch": 57.0, |
|
"eval_loss": 0.35607481002807617, |
|
"eval_runtime": 336.6291, |
|
"eval_samples_per_second": 13.721, |
|
"eval_steps_per_second": 0.288, |
|
"eval_wer": 0.2550495776716856, |
|
"step": 57171 |
|
}, |
|
{ |
|
"epoch": 58.0, |
|
"learning_rate": 4.244410876132931e-05, |
|
"loss": 0.0703, |
|
"step": 58174 |
|
}, |
|
{ |
|
"epoch": 58.0, |
|
"eval_loss": 0.34993934631347656, |
|
"eval_runtime": 338.604, |
|
"eval_samples_per_second": 13.641, |
|
"eval_steps_per_second": 0.286, |
|
"eval_wer": 0.2562737177133064, |
|
"step": 58174 |
|
}, |
|
{ |
|
"epoch": 59.0, |
|
"learning_rate": 4.143403826787513e-05, |
|
"loss": 0.0689, |
|
"step": 59177 |
|
}, |
|
{ |
|
"epoch": 59.0, |
|
"eval_loss": 0.35656341910362244, |
|
"eval_runtime": 282.9543, |
|
"eval_samples_per_second": 16.324, |
|
"eval_steps_per_second": 0.343, |
|
"eval_wer": 0.25489656016648304, |
|
"step": 59177 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"learning_rate": 4.0424974823766364e-05, |
|
"loss": 0.0667, |
|
"step": 60180 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"eval_loss": 0.35375431180000305, |
|
"eval_runtime": 291.8817, |
|
"eval_samples_per_second": 15.825, |
|
"eval_steps_per_second": 0.332, |
|
"eval_wer": 0.2550495776716856, |
|
"step": 60180 |
|
}, |
|
{ |
|
"epoch": 61.0, |
|
"learning_rate": 3.9414904330312185e-05, |
|
"loss": 0.0665, |
|
"step": 61183 |
|
}, |
|
{ |
|
"epoch": 61.0, |
|
"eval_loss": 0.3356081545352936, |
|
"eval_runtime": 356.5563, |
|
"eval_samples_per_second": 12.954, |
|
"eval_steps_per_second": 0.272, |
|
"eval_wer": 0.2545293181539968, |
|
"step": 61183 |
|
}, |
|
{ |
|
"epoch": 62.0, |
|
"learning_rate": 3.8405840886203424e-05, |
|
"loss": 0.0662, |
|
"step": 62186 |
|
}, |
|
{ |
|
"epoch": 62.0, |
|
"eval_loss": 0.347979873418808, |
|
"eval_runtime": 337.8069, |
|
"eval_samples_per_second": 13.673, |
|
"eval_steps_per_second": 0.287, |
|
"eval_wer": 0.2541314726404701, |
|
"step": 62186 |
|
}, |
|
{ |
|
"epoch": 63.0, |
|
"learning_rate": 3.7395770392749245e-05, |
|
"loss": 0.0653, |
|
"step": 63189 |
|
}, |
|
{ |
|
"epoch": 63.0, |
|
"eval_loss": 0.34537407755851746, |
|
"eval_runtime": 361.1194, |
|
"eval_samples_per_second": 12.791, |
|
"eval_steps_per_second": 0.269, |
|
"eval_wer": 0.2513771575468234, |
|
"step": 63189 |
|
}, |
|
{ |
|
"epoch": 64.0, |
|
"learning_rate": 3.6386706948640483e-05, |
|
"loss": 0.0633, |
|
"step": 64192 |
|
}, |
|
{ |
|
"epoch": 64.0, |
|
"eval_loss": 0.36930787563323975, |
|
"eval_runtime": 342.2873, |
|
"eval_samples_per_second": 13.495, |
|
"eval_steps_per_second": 0.283, |
|
"eval_wer": 0.25220345207491734, |
|
"step": 64192 |
|
}, |
|
{ |
|
"epoch": 65.0, |
|
"learning_rate": 3.5376636455186305e-05, |
|
"loss": 0.0632, |
|
"step": 65195 |
|
}, |
|
{ |
|
"epoch": 65.0, |
|
"eval_loss": 0.36128777265548706, |
|
"eval_runtime": 361.0289, |
|
"eval_samples_per_second": 12.794, |
|
"eval_steps_per_second": 0.269, |
|
"eval_wer": 0.254009058636308, |
|
"step": 65195 |
|
}, |
|
{ |
|
"epoch": 66.0, |
|
"learning_rate": 3.4366565961732126e-05, |
|
"loss": 0.0615, |
|
"step": 66198 |
|
}, |
|
{ |
|
"epoch": 66.0, |
|
"eval_loss": 0.3568029999732971, |
|
"eval_runtime": 343.0543, |
|
"eval_samples_per_second": 13.464, |
|
"eval_steps_per_second": 0.283, |
|
"eval_wer": 0.25211164157179583, |
|
"step": 66198 |
|
}, |
|
{ |
|
"epoch": 67.0, |
|
"learning_rate": 3.3357502517623365e-05, |
|
"loss": 0.0618, |
|
"step": 67201 |
|
}, |
|
{ |
|
"epoch": 67.0, |
|
"eval_loss": 0.3640798032283783, |
|
"eval_runtime": 350.9436, |
|
"eval_samples_per_second": 13.162, |
|
"eval_steps_per_second": 0.276, |
|
"eval_wer": 0.2523258660790794, |
|
"step": 67201 |
|
}, |
|
{ |
|
"epoch": 68.0, |
|
"learning_rate": 3.2347432024169186e-05, |
|
"loss": 0.0609, |
|
"step": 68204 |
|
}, |
|
{ |
|
"epoch": 68.0, |
|
"eval_loss": 0.36202770471572876, |
|
"eval_runtime": 325.6728, |
|
"eval_samples_per_second": 14.183, |
|
"eval_steps_per_second": 0.298, |
|
"eval_wer": 0.25293793609988985, |
|
"step": 68204 |
|
}, |
|
{ |
|
"epoch": 69.0, |
|
"learning_rate": 3.133736153071501e-05, |
|
"loss": 0.0594, |
|
"step": 69207 |
|
}, |
|
{ |
|
"epoch": 69.0, |
|
"eval_loss": 0.3554952144622803, |
|
"eval_runtime": 322.2715, |
|
"eval_samples_per_second": 14.333, |
|
"eval_steps_per_second": 0.301, |
|
"eval_wer": 0.25205043456971477, |
|
"step": 69207 |
|
}, |
|
{ |
|
"epoch": 70.0, |
|
"learning_rate": 3.0328298086606243e-05, |
|
"loss": 0.0594, |
|
"step": 70210 |
|
}, |
|
{ |
|
"epoch": 70.0, |
|
"eval_loss": 0.35941246151924133, |
|
"eval_runtime": 323.1251, |
|
"eval_samples_per_second": 14.295, |
|
"eval_steps_per_second": 0.3, |
|
"eval_wer": 0.25208103807075527, |
|
"step": 70210 |
|
}, |
|
{ |
|
"epoch": 71.0, |
|
"learning_rate": 2.9318227593152064e-05, |
|
"loss": 0.0598, |
|
"step": 71213 |
|
}, |
|
{ |
|
"epoch": 71.0, |
|
"eval_loss": 0.35894548892974854, |
|
"eval_runtime": 348.1512, |
|
"eval_samples_per_second": 13.267, |
|
"eval_steps_per_second": 0.279, |
|
"eval_wer": 0.25067327702289144, |
|
"step": 71213 |
|
}, |
|
{ |
|
"epoch": 72.0, |
|
"learning_rate": 2.8308157099697885e-05, |
|
"loss": 0.0579, |
|
"step": 72216 |
|
}, |
|
{ |
|
"epoch": 72.0, |
|
"eval_loss": 0.34253111481666565, |
|
"eval_runtime": 354.641, |
|
"eval_samples_per_second": 13.024, |
|
"eval_steps_per_second": 0.274, |
|
"eval_wer": 0.25119353654058024, |
|
"step": 72216 |
|
}, |
|
{ |
|
"epoch": 73.0, |
|
"learning_rate": 2.7299093655589124e-05, |
|
"loss": 0.0582, |
|
"step": 73219 |
|
}, |
|
{ |
|
"epoch": 73.0, |
|
"eval_loss": 0.35133859515190125, |
|
"eval_runtime": 330.0079, |
|
"eval_samples_per_second": 13.997, |
|
"eval_steps_per_second": 0.294, |
|
"eval_wer": 0.24917370547190598, |
|
"step": 73219 |
|
}, |
|
{ |
|
"epoch": 74.0, |
|
"learning_rate": 2.6289023162134945e-05, |
|
"loss": 0.0575, |
|
"step": 74222 |
|
}, |
|
{ |
|
"epoch": 74.0, |
|
"eval_loss": 0.36137738823890686, |
|
"eval_runtime": 330.3739, |
|
"eval_samples_per_second": 13.981, |
|
"eval_steps_per_second": 0.294, |
|
"eval_wer": 0.25156077855306647, |
|
"step": 74222 |
|
}, |
|
{ |
|
"epoch": 75.0, |
|
"learning_rate": 2.5278952668680767e-05, |
|
"loss": 0.0569, |
|
"step": 75225 |
|
}, |
|
{ |
|
"epoch": 75.0, |
|
"eval_loss": 0.36236804723739624, |
|
"eval_runtime": 316.6665, |
|
"eval_samples_per_second": 14.586, |
|
"eval_steps_per_second": 0.306, |
|
"eval_wer": 0.24947974048231117, |
|
"step": 75225 |
|
}, |
|
{ |
|
"epoch": 76.0, |
|
"learning_rate": 2.4269889224572005e-05, |
|
"loss": 0.0558, |
|
"step": 76228 |
|
}, |
|
{ |
|
"epoch": 76.0, |
|
"eval_loss": 0.3634468615055084, |
|
"eval_runtime": 319.3864, |
|
"eval_samples_per_second": 14.462, |
|
"eval_steps_per_second": 0.304, |
|
"eval_wer": 0.24908189496878444, |
|
"step": 76228 |
|
}, |
|
{ |
|
"epoch": 77.0, |
|
"learning_rate": 2.3259818731117823e-05, |
|
"loss": 0.0556, |
|
"step": 77231 |
|
}, |
|
{ |
|
"epoch": 77.0, |
|
"eval_loss": 0.3610389530658722, |
|
"eval_runtime": 350.196, |
|
"eval_samples_per_second": 13.19, |
|
"eval_steps_per_second": 0.277, |
|
"eval_wer": 0.248408617945893, |
|
"step": 77231 |
|
}, |
|
{ |
|
"epoch": 78.0, |
|
"learning_rate": 2.2249748237663648e-05, |
|
"loss": 0.0548, |
|
"step": 78234 |
|
}, |
|
{ |
|
"epoch": 78.0, |
|
"eval_loss": 0.3590007722377777, |
|
"eval_runtime": 341.784, |
|
"eval_samples_per_second": 13.514, |
|
"eval_steps_per_second": 0.284, |
|
"eval_wer": 0.24657240788346188, |
|
"step": 78234 |
|
}, |
|
{ |
|
"epoch": 79.0, |
|
"learning_rate": 2.123967774420947e-05, |
|
"loss": 0.0528, |
|
"step": 79237 |
|
}, |
|
{ |
|
"epoch": 79.0, |
|
"eval_loss": 0.3674350678920746, |
|
"eval_runtime": 327.8795, |
|
"eval_samples_per_second": 14.087, |
|
"eval_steps_per_second": 0.296, |
|
"eval_wer": 0.248408617945893, |
|
"step": 79237 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"learning_rate": 2.0230614300100704e-05, |
|
"loss": 0.0521, |
|
"step": 80240 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"eval_loss": 0.3592623770236969, |
|
"eval_runtime": 329.3838, |
|
"eval_samples_per_second": 14.023, |
|
"eval_steps_per_second": 0.294, |
|
"eval_wer": 0.2469702533969886, |
|
"step": 80240 |
|
}, |
|
{ |
|
"epoch": 81.0, |
|
"learning_rate": 1.9221550855991943e-05, |
|
"loss": 0.0532, |
|
"step": 81243 |
|
}, |
|
{ |
|
"epoch": 81.0, |
|
"eval_loss": 0.370604932308197, |
|
"eval_runtime": 303.5979, |
|
"eval_samples_per_second": 15.214, |
|
"eval_steps_per_second": 0.32, |
|
"eval_wer": 0.24947974048231117, |
|
"step": 81243 |
|
}, |
|
{ |
|
"epoch": 82.0, |
|
"learning_rate": 1.8211480362537764e-05, |
|
"loss": 0.0517, |
|
"step": 82246 |
|
}, |
|
{ |
|
"epoch": 82.0, |
|
"eval_loss": 0.37441593408584595, |
|
"eval_runtime": 316.1258, |
|
"eval_samples_per_second": 14.611, |
|
"eval_steps_per_second": 0.307, |
|
"eval_wer": 0.24975517199167585, |
|
"step": 82246 |
|
}, |
|
{ |
|
"epoch": 83.0, |
|
"learning_rate": 1.7201409869083586e-05, |
|
"loss": 0.0516, |
|
"step": 83249 |
|
}, |
|
{ |
|
"epoch": 83.0, |
|
"eval_loss": 0.3725976347923279, |
|
"eval_runtime": 353.3018, |
|
"eval_samples_per_second": 13.074, |
|
"eval_steps_per_second": 0.275, |
|
"eval_wer": 0.2485922389521361, |
|
"step": 83249 |
|
}, |
|
{ |
|
"epoch": 84.0, |
|
"learning_rate": 1.6192346424974824e-05, |
|
"loss": 0.0496, |
|
"step": 84252 |
|
}, |
|
{ |
|
"epoch": 84.0, |
|
"eval_loss": 0.36621466279029846, |
|
"eval_runtime": 358.7351, |
|
"eval_samples_per_second": 12.876, |
|
"eval_steps_per_second": 0.27, |
|
"eval_wer": 0.24730689190843433, |
|
"step": 84252 |
|
}, |
|
{ |
|
"epoch": 85.0, |
|
"learning_rate": 1.5182275931520645e-05, |
|
"loss": 0.0503, |
|
"step": 85255 |
|
}, |
|
{ |
|
"epoch": 85.0, |
|
"eval_loss": 0.3839229941368103, |
|
"eval_runtime": 322.7657, |
|
"eval_samples_per_second": 14.311, |
|
"eval_steps_per_second": 0.301, |
|
"eval_wer": 0.24819439343860938, |
|
"step": 85255 |
|
}, |
|
{ |
|
"epoch": 86.0, |
|
"learning_rate": 1.4173212487411884e-05, |
|
"loss": 0.0505, |
|
"step": 86258 |
|
}, |
|
{ |
|
"epoch": 86.0, |
|
"eval_loss": 0.37594547867774963, |
|
"eval_runtime": 344.4521, |
|
"eval_samples_per_second": 13.41, |
|
"eval_steps_per_second": 0.282, |
|
"eval_wer": 0.24843922144693353, |
|
"step": 86258 |
|
}, |
|
{ |
|
"epoch": 87.0, |
|
"learning_rate": 1.3163141993957703e-05, |
|
"loss": 0.0489, |
|
"step": 87261 |
|
}, |
|
{ |
|
"epoch": 87.0, |
|
"eval_loss": 0.3759251832962036, |
|
"eval_runtime": 348.9851, |
|
"eval_samples_per_second": 13.236, |
|
"eval_steps_per_second": 0.278, |
|
"eval_wer": 0.24718447790427225, |
|
"step": 87261 |
|
}, |
|
{ |
|
"epoch": 88.0, |
|
"learning_rate": 1.2153071500503525e-05, |
|
"loss": 0.0498, |
|
"step": 88264 |
|
}, |
|
{ |
|
"epoch": 88.0, |
|
"eval_loss": 0.3712981045246124, |
|
"eval_runtime": 358.3358, |
|
"eval_samples_per_second": 12.89, |
|
"eval_steps_per_second": 0.271, |
|
"eval_wer": 0.2474599094136369, |
|
"step": 88264 |
|
}, |
|
{ |
|
"epoch": 89.0, |
|
"learning_rate": 1.1143001007049346e-05, |
|
"loss": 0.0481, |
|
"step": 89267 |
|
}, |
|
{ |
|
"epoch": 89.0, |
|
"eval_loss": 0.37631461024284363, |
|
"eval_runtime": 376.3412, |
|
"eval_samples_per_second": 12.273, |
|
"eval_steps_per_second": 0.258, |
|
"eval_wer": 0.2474293059125964, |
|
"step": 89267 |
|
}, |
|
{ |
|
"epoch": 90.0, |
|
"learning_rate": 1.0134944612286003e-05, |
|
"loss": 0.0483, |
|
"step": 90270 |
|
}, |
|
{ |
|
"epoch": 90.0, |
|
"eval_loss": 0.380001425743103, |
|
"eval_runtime": 340.9184, |
|
"eval_samples_per_second": 13.549, |
|
"eval_steps_per_second": 0.285, |
|
"eval_wer": 0.24651120088138084, |
|
"step": 90270 |
|
}, |
|
{ |
|
"epoch": 91.0, |
|
"learning_rate": 9.124874118831823e-06, |
|
"loss": 0.0471, |
|
"step": 91273 |
|
}, |
|
{ |
|
"epoch": 91.0, |
|
"eval_loss": 0.37985819578170776, |
|
"eval_runtime": 326.6601, |
|
"eval_samples_per_second": 14.14, |
|
"eval_steps_per_second": 0.297, |
|
"eval_wer": 0.2472762884073938, |
|
"step": 91273 |
|
}, |
|
{ |
|
"epoch": 92.0, |
|
"learning_rate": 8.114803625377644e-06, |
|
"loss": 0.0477, |
|
"step": 92276 |
|
}, |
|
{ |
|
"epoch": 92.0, |
|
"eval_loss": 0.3750379979610443, |
|
"eval_runtime": 337.4116, |
|
"eval_samples_per_second": 13.69, |
|
"eval_steps_per_second": 0.287, |
|
"eval_wer": 0.2474293059125964, |
|
"step": 92276 |
|
}, |
|
{ |
|
"epoch": 93.0, |
|
"learning_rate": 7.104733131923465e-06, |
|
"loss": 0.0468, |
|
"step": 93279 |
|
}, |
|
{ |
|
"epoch": 93.0, |
|
"eval_loss": 0.3763069212436676, |
|
"eval_runtime": 317.2801, |
|
"eval_samples_per_second": 14.558, |
|
"eval_steps_per_second": 0.306, |
|
"eval_wer": 0.24767413392092055, |
|
"step": 93279 |
|
}, |
|
{ |
|
"epoch": 94.0, |
|
"learning_rate": 6.0946626384692855e-06, |
|
"loss": 0.047, |
|
"step": 94282 |
|
}, |
|
{ |
|
"epoch": 94.0, |
|
"eval_loss": 0.3740558326244354, |
|
"eval_runtime": 286.4453, |
|
"eval_samples_per_second": 16.125, |
|
"eval_steps_per_second": 0.339, |
|
"eval_wer": 0.24687844289386707, |
|
"step": 94282 |
|
}, |
|
{ |
|
"epoch": 95.0, |
|
"learning_rate": 5.084592145015106e-06, |
|
"loss": 0.0458, |
|
"step": 95285 |
|
}, |
|
{ |
|
"epoch": 95.0, |
|
"eval_loss": 0.37711286544799805, |
|
"eval_runtime": 358.0674, |
|
"eval_samples_per_second": 12.9, |
|
"eval_steps_per_second": 0.271, |
|
"eval_wer": 0.24706206390011018, |
|
"step": 95285 |
|
}, |
|
{ |
|
"epoch": 96.0, |
|
"learning_rate": 4.0755287009063444e-06, |
|
"loss": 0.0462, |
|
"step": 96288 |
|
}, |
|
{ |
|
"epoch": 96.0, |
|
"eval_loss": 0.37909871339797974, |
|
"eval_runtime": 326.4087, |
|
"eval_samples_per_second": 14.151, |
|
"eval_steps_per_second": 0.297, |
|
"eval_wer": 0.24672542538866446, |
|
"step": 96288 |
|
}, |
|
{ |
|
"epoch": 97.0, |
|
"learning_rate": 3.066465256797583e-06, |
|
"loss": 0.0458, |
|
"step": 97291 |
|
}, |
|
{ |
|
"epoch": 97.0, |
|
"eval_loss": 0.37906551361083984, |
|
"eval_runtime": 319.3611, |
|
"eval_samples_per_second": 14.463, |
|
"eval_steps_per_second": 0.304, |
|
"eval_wer": 0.2458073203574489, |
|
"step": 97291 |
|
} |
|
], |
|
"max_steps": 100300, |
|
"num_train_epochs": 100, |
|
"total_flos": 1.283405734102556e+21, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|