|
{ |
|
"best_metric": 21.258466244264802, |
|
"best_model_checkpoint": "whisper-medium-mn-10/checkpoint-39000", |
|
"epoch": 3.465303647232089, |
|
"global_step": 40000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.2000000000000006e-07, |
|
"loss": 4.4103, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.200000000000001e-07, |
|
"loss": 3.7386, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.42e-06, |
|
"loss": 2.9759, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9200000000000003e-06, |
|
"loss": 2.3506, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.42e-06, |
|
"loss": 1.9105, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.92e-06, |
|
"loss": 1.6286, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.4200000000000007e-06, |
|
"loss": 1.4143, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.920000000000001e-06, |
|
"loss": 1.1879, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.42e-06, |
|
"loss": 1.0801, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.92e-06, |
|
"loss": 0.9285, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5.420000000000001e-06, |
|
"loss": 0.8351, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.92e-06, |
|
"loss": 0.7879, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.42e-06, |
|
"loss": 0.7726, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.92e-06, |
|
"loss": 0.715, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.420000000000001e-06, |
|
"loss": 0.6641, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.92e-06, |
|
"loss": 0.6486, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.42e-06, |
|
"loss": 0.6421, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.920000000000001e-06, |
|
"loss": 0.6152, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.42e-06, |
|
"loss": 0.6019, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.920000000000002e-06, |
|
"loss": 0.6037, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.994683544303798e-06, |
|
"loss": 0.5603, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.988354430379748e-06, |
|
"loss": 0.5327, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.982025316455697e-06, |
|
"loss": 0.5164, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.975696202531647e-06, |
|
"loss": 0.5326, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.969367088607596e-06, |
|
"loss": 0.516, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.963037974683545e-06, |
|
"loss": 0.498, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.956708860759495e-06, |
|
"loss": 0.4944, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.950379746835444e-06, |
|
"loss": 0.4356, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.944050632911392e-06, |
|
"loss": 0.4679, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.937721518987341e-06, |
|
"loss": 0.4567, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.931392405063291e-06, |
|
"loss": 0.4482, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.92506329113924e-06, |
|
"loss": 0.4783, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.91873417721519e-06, |
|
"loss": 0.42, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.912405063291141e-06, |
|
"loss": 0.4189, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.90607594936709e-06, |
|
"loss": 0.4223, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.89974683544304e-06, |
|
"loss": 0.4355, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.89341772151899e-06, |
|
"loss": 0.4153, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.887088607594939e-06, |
|
"loss": 0.4361, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.880759493670888e-06, |
|
"loss": 0.3909, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.874430379746836e-06, |
|
"loss": 0.4197, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_cer": 19.094707051649205, |
|
"eval_loss": 0.44623005390167236, |
|
"eval_runtime": 2202.0442, |
|
"eval_samples_per_second": 0.856, |
|
"eval_steps_per_second": 0.107, |
|
"eval_wer": 53.96001747869784, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.868101265822785e-06, |
|
"loss": 0.4227, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.861772151898735e-06, |
|
"loss": 0.4155, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.855443037974684e-06, |
|
"loss": 0.3636, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.849113924050633e-06, |
|
"loss": 0.3911, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.842784810126583e-06, |
|
"loss": 0.3817, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.836455696202532e-06, |
|
"loss": 0.3686, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.830126582278482e-06, |
|
"loss": 0.3813, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.823797468354431e-06, |
|
"loss": 0.3835, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.81746835443038e-06, |
|
"loss": 0.3831, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.81113924050633e-06, |
|
"loss": 0.3746, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.80481012658228e-06, |
|
"loss": 0.3418, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.798481012658229e-06, |
|
"loss": 0.3441, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.792151898734178e-06, |
|
"loss": 0.3762, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.785822784810128e-06, |
|
"loss": 0.3589, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.779493670886077e-06, |
|
"loss": 0.3635, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.773164556962027e-06, |
|
"loss": 0.3618, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.766835443037976e-06, |
|
"loss": 0.3712, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.760506329113924e-06, |
|
"loss": 0.3402, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.754177215189873e-06, |
|
"loss": 0.3421, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.747848101265823e-06, |
|
"loss": 0.3498, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.741518987341772e-06, |
|
"loss": 0.3499, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.735189873417721e-06, |
|
"loss": 0.3289, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.728860759493671e-06, |
|
"loss": 0.3452, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.72253164556962e-06, |
|
"loss": 0.3384, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.71620253164557e-06, |
|
"loss": 0.33, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.70987341772152e-06, |
|
"loss": 0.3348, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.70354430379747e-06, |
|
"loss": 0.3276, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.69721518987342e-06, |
|
"loss": 0.3217, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.69088607594937e-06, |
|
"loss": 0.3284, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.684556962025317e-06, |
|
"loss": 0.3175, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.678227848101266e-06, |
|
"loss": 0.3225, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.671898734177216e-06, |
|
"loss": 0.3169, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.665569620253165e-06, |
|
"loss": 0.3023, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.659240506329115e-06, |
|
"loss": 0.3277, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.652911392405064e-06, |
|
"loss": 0.3254, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.646582278481013e-06, |
|
"loss": 0.3059, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.640253164556963e-06, |
|
"loss": 0.3378, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.633924050632912e-06, |
|
"loss": 0.3163, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.627594936708862e-06, |
|
"loss": 0.3269, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.621265822784811e-06, |
|
"loss": 0.3288, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"eval_cer": 14.801556551329135, |
|
"eval_loss": 0.346794068813324, |
|
"eval_runtime": 2133.941, |
|
"eval_samples_per_second": 0.884, |
|
"eval_steps_per_second": 0.111, |
|
"eval_wer": 44.210181341490056, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.61493670886076e-06, |
|
"loss": 0.3164, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.60860759493671e-06, |
|
"loss": 0.3281, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.60227848101266e-06, |
|
"loss": 0.2904, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.595949367088609e-06, |
|
"loss": 0.2981, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.589620253164558e-06, |
|
"loss": 0.3325, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.583291139240508e-06, |
|
"loss": 0.3252, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.576962025316457e-06, |
|
"loss": 0.3206, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.570632911392405e-06, |
|
"loss": 0.3195, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.564303797468354e-06, |
|
"loss": 0.2796, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.557974683544304e-06, |
|
"loss": 0.3053, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.551645569620253e-06, |
|
"loss": 0.3021, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.545316455696203e-06, |
|
"loss": 0.2993, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.538987341772152e-06, |
|
"loss": 0.3056, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.532658227848101e-06, |
|
"loss": 0.2964, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.526329113924051e-06, |
|
"loss": 0.3082, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.52e-06, |
|
"loss": 0.3009, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.51367088607595e-06, |
|
"loss": 0.2803, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.5073417721519e-06, |
|
"loss": 0.278, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.501012658227849e-06, |
|
"loss": 0.2902, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.494683544303798e-06, |
|
"loss": 0.2936, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.488354430379747e-06, |
|
"loss": 0.2925, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.482025316455697e-06, |
|
"loss": 0.282, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.475696202531646e-06, |
|
"loss": 0.2949, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.469367088607596e-06, |
|
"loss": 0.3076, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.463037974683545e-06, |
|
"loss": 0.2993, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.456708860759495e-06, |
|
"loss": 0.2825, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.450379746835444e-06, |
|
"loss": 0.328, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.444050632911393e-06, |
|
"loss": 0.278, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.437721518987343e-06, |
|
"loss": 0.2919, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.431392405063292e-06, |
|
"loss": 0.3111, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.425063291139242e-06, |
|
"loss": 0.2916, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.418734177215191e-06, |
|
"loss": 0.2909, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.41240506329114e-06, |
|
"loss": 0.2753, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.40607594936709e-06, |
|
"loss": 0.2827, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.39974683544304e-06, |
|
"loss": 0.3033, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.393417721518989e-06, |
|
"loss": 0.2756, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.387088607594937e-06, |
|
"loss": 0.3157, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.380759493670886e-06, |
|
"loss": 0.2881, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.374430379746835e-06, |
|
"loss": 0.2873, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.368101265822785e-06, |
|
"loss": 0.2737, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_cer": 12.34712442303157, |
|
"eval_loss": 0.3020499050617218, |
|
"eval_runtime": 2105.7024, |
|
"eval_samples_per_second": 0.896, |
|
"eval_steps_per_second": 0.112, |
|
"eval_wer": 36.16998033646494, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.361772151898734e-06, |
|
"loss": 0.2999, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.355443037974684e-06, |
|
"loss": 0.2745, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.349113924050633e-06, |
|
"loss": 0.2675, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.342784810126583e-06, |
|
"loss": 0.2656, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.336455696202532e-06, |
|
"loss": 0.2781, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.330126582278481e-06, |
|
"loss": 0.2615, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.323797468354431e-06, |
|
"loss": 0.2572, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.31746835443038e-06, |
|
"loss": 0.2633, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.31113924050633e-06, |
|
"loss": 0.2788, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.30481012658228e-06, |
|
"loss": 0.2528, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.298481012658229e-06, |
|
"loss": 0.271, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.292151898734178e-06, |
|
"loss": 0.3011, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.285822784810127e-06, |
|
"loss": 0.2924, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.279493670886077e-06, |
|
"loss": 0.2556, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.273164556962026e-06, |
|
"loss": 0.2671, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.266835443037976e-06, |
|
"loss": 0.2622, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.260506329113925e-06, |
|
"loss": 0.276, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.254177215189875e-06, |
|
"loss": 0.2652, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.247848101265824e-06, |
|
"loss": 0.2663, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.241518987341773e-06, |
|
"loss": 0.275, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.235189873417723e-06, |
|
"loss": 0.2741, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.228860759493672e-06, |
|
"loss": 0.2864, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.222531645569622e-06, |
|
"loss": 0.256, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.216202531645571e-06, |
|
"loss": 0.283, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.20987341772152e-06, |
|
"loss": 0.2825, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.20354430379747e-06, |
|
"loss": 0.2567, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.197215189873418e-06, |
|
"loss": 0.2591, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.190886075949367e-06, |
|
"loss": 0.2542, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.184556962025317e-06, |
|
"loss": 0.2737, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.178227848101266e-06, |
|
"loss": 0.293, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.171898734177215e-06, |
|
"loss": 0.2686, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.165569620253165e-06, |
|
"loss": 0.2508, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.159240506329114e-06, |
|
"loss": 0.2581, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.152911392405064e-06, |
|
"loss": 0.2788, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.146582278481013e-06, |
|
"loss": 0.2472, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.140253164556963e-06, |
|
"loss": 0.2504, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.133924050632912e-06, |
|
"loss": 0.2515, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.127594936708861e-06, |
|
"loss": 0.2526, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.121265822784811e-06, |
|
"loss": 0.2611, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.11493670886076e-06, |
|
"loss": 0.2558, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_cer": 11.717091742191975, |
|
"eval_loss": 0.2823735475540161, |
|
"eval_runtime": 2062.27, |
|
"eval_samples_per_second": 0.915, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 34.17085427135678, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.10860759493671e-06, |
|
"loss": 0.2471, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.10227848101266e-06, |
|
"loss": 0.2682, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.095949367088609e-06, |
|
"loss": 0.2501, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.089620253164558e-06, |
|
"loss": 0.2591, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.083291139240506e-06, |
|
"loss": 0.236, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.076962025316457e-06, |
|
"loss": 0.2607, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.070632911392406e-06, |
|
"loss": 0.2639, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.064303797468356e-06, |
|
"loss": 0.2521, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.057974683544305e-06, |
|
"loss": 0.2528, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.051645569620255e-06, |
|
"loss": 0.2735, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.045316455696204e-06, |
|
"loss": 0.2626, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.038987341772153e-06, |
|
"loss": 0.2688, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.032658227848103e-06, |
|
"loss": 0.2764, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.026329113924052e-06, |
|
"loss": 0.2522, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.020000000000002e-06, |
|
"loss": 0.2347, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.01367088607595e-06, |
|
"loss": 0.2504, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.007341772151899e-06, |
|
"loss": 0.2759, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.001012658227848e-06, |
|
"loss": 0.2342, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 8.994683544303798e-06, |
|
"loss": 0.2776, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 8.988354430379747e-06, |
|
"loss": 0.2484, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 8.982025316455697e-06, |
|
"loss": 0.2619, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 8.975696202531646e-06, |
|
"loss": 0.2625, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 8.969367088607595e-06, |
|
"loss": 0.2474, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 8.963037974683545e-06, |
|
"loss": 0.2444, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 8.956708860759494e-06, |
|
"loss": 0.2463, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 8.950379746835444e-06, |
|
"loss": 0.265, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 8.944050632911393e-06, |
|
"loss": 0.2261, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 8.937721518987343e-06, |
|
"loss": 0.2087, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 8.931392405063292e-06, |
|
"loss": 0.2399, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 8.925063291139241e-06, |
|
"loss": 0.2358, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 8.918734177215191e-06, |
|
"loss": 0.2617, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 8.91240506329114e-06, |
|
"loss": 0.2234, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 8.90607594936709e-06, |
|
"loss": 0.2403, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 8.899746835443037e-06, |
|
"loss": 0.2553, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 8.893417721518987e-06, |
|
"loss": 0.2396, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 8.887088607594936e-06, |
|
"loss": 0.2561, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 8.880759493670886e-06, |
|
"loss": 0.2618, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 8.874430379746835e-06, |
|
"loss": 0.2445, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 8.868101265822786e-06, |
|
"loss": 0.2402, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 8.861772151898736e-06, |
|
"loss": 0.2406, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_cer": 10.355109329200499, |
|
"eval_loss": 0.25939086079597473, |
|
"eval_runtime": 2083.4671, |
|
"eval_samples_per_second": 0.905, |
|
"eval_steps_per_second": 0.113, |
|
"eval_wer": 31.123006336027963, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 8.855443037974685e-06, |
|
"loss": 0.253, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 8.849113924050635e-06, |
|
"loss": 0.221, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 8.842784810126584e-06, |
|
"loss": 0.2436, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 8.836455696202533e-06, |
|
"loss": 0.235, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 8.830126582278483e-06, |
|
"loss": 0.2416, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 8.82379746835443e-06, |
|
"loss": 0.2606, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 8.81746835443038e-06, |
|
"loss": 0.2341, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 8.81113924050633e-06, |
|
"loss": 0.2441, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 8.804810126582279e-06, |
|
"loss": 0.2464, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 8.798481012658228e-06, |
|
"loss": 0.2438, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 8.792151898734178e-06, |
|
"loss": 0.2336, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 8.785822784810127e-06, |
|
"loss": 0.2296, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 8.779493670886077e-06, |
|
"loss": 0.231, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 8.773164556962026e-06, |
|
"loss": 0.2319, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 8.766835443037975e-06, |
|
"loss": 0.2464, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 8.760506329113925e-06, |
|
"loss": 0.2242, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 8.754177215189874e-06, |
|
"loss": 0.2346, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 8.747848101265824e-06, |
|
"loss": 0.2184, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 8.741518987341773e-06, |
|
"loss": 0.255, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 8.735189873417723e-06, |
|
"loss": 0.2423, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 8.728860759493672e-06, |
|
"loss": 0.2521, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 8.722531645569621e-06, |
|
"loss": 0.251, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 8.716202531645571e-06, |
|
"loss": 0.2227, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 8.709873417721519e-06, |
|
"loss": 0.2165, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 8.703544303797468e-06, |
|
"loss": 0.2472, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 8.697215189873417e-06, |
|
"loss": 0.2309, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 8.690886075949367e-06, |
|
"loss": 0.2428, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 8.684556962025316e-06, |
|
"loss": 0.2115, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 8.678227848101266e-06, |
|
"loss": 0.2445, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 8.671898734177215e-06, |
|
"loss": 0.2371, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 8.665569620253165e-06, |
|
"loss": 0.2125, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 8.659240506329116e-06, |
|
"loss": 0.2126, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 8.652911392405065e-06, |
|
"loss": 0.2521, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 8.646582278481015e-06, |
|
"loss": 0.2287, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 8.640253164556964e-06, |
|
"loss": 0.2427, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 8.633924050632912e-06, |
|
"loss": 0.232, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 8.627594936708861e-06, |
|
"loss": 0.2578, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 8.62126582278481e-06, |
|
"loss": 0.2249, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 8.61493670886076e-06, |
|
"loss": 0.2324, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 8.60860759493671e-06, |
|
"loss": 0.218, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"eval_cer": 9.781510056938783, |
|
"eval_loss": 0.24516499042510986, |
|
"eval_runtime": 2045.6229, |
|
"eval_samples_per_second": 0.922, |
|
"eval_steps_per_second": 0.115, |
|
"eval_wer": 29.68647585754861, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 8.602278481012659e-06, |
|
"loss": 0.251, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 8.595949367088608e-06, |
|
"loss": 0.2421, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 8.589620253164558e-06, |
|
"loss": 0.223, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 8.583291139240507e-06, |
|
"loss": 0.2185, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 8.576962025316457e-06, |
|
"loss": 0.2334, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 8.570632911392406e-06, |
|
"loss": 0.2384, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 8.564303797468355e-06, |
|
"loss": 0.2263, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 8.557974683544305e-06, |
|
"loss": 0.2403, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 8.551645569620254e-06, |
|
"loss": 0.2131, |
|
"step": 6225 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 8.545316455696204e-06, |
|
"loss": 0.2334, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 8.538987341772153e-06, |
|
"loss": 0.2314, |
|
"step": 6275 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.532658227848103e-06, |
|
"loss": 0.2331, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.52632911392405e-06, |
|
"loss": 0.218, |
|
"step": 6325 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.52e-06, |
|
"loss": 0.2325, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.51367088607595e-06, |
|
"loss": 0.2316, |
|
"step": 6375 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.507341772151899e-06, |
|
"loss": 0.2424, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.501012658227848e-06, |
|
"loss": 0.218, |
|
"step": 6425 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.494683544303797e-06, |
|
"loss": 0.2196, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.488354430379747e-06, |
|
"loss": 0.2286, |
|
"step": 6475 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.482025316455696e-06, |
|
"loss": 0.2278, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.475696202531646e-06, |
|
"loss": 0.2209, |
|
"step": 6525 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.469367088607595e-06, |
|
"loss": 0.2357, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.463037974683545e-06, |
|
"loss": 0.2322, |
|
"step": 6575 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.456708860759494e-06, |
|
"loss": 0.2201, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.450379746835443e-06, |
|
"loss": 0.2047, |
|
"step": 6625 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.444050632911393e-06, |
|
"loss": 0.2099, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.437721518987342e-06, |
|
"loss": 0.2214, |
|
"step": 6675 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.431392405063292e-06, |
|
"loss": 0.2438, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.425063291139241e-06, |
|
"loss": 0.2348, |
|
"step": 6725 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.41873417721519e-06, |
|
"loss": 0.2267, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.41240506329114e-06, |
|
"loss": 0.2388, |
|
"step": 6775 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.40607594936709e-06, |
|
"loss": 0.2177, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.399746835443039e-06, |
|
"loss": 0.2132, |
|
"step": 6825 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.393417721518988e-06, |
|
"loss": 0.2148, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.387088607594938e-06, |
|
"loss": 0.2281, |
|
"step": 6875 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.380759493670887e-06, |
|
"loss": 0.2309, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.374430379746837e-06, |
|
"loss": 0.2286, |
|
"step": 6925 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.368101265822786e-06, |
|
"loss": 0.235, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.361772151898735e-06, |
|
"loss": 0.2163, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 8.355443037974685e-06, |
|
"loss": 0.2253, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"eval_cer": 9.671170108823826, |
|
"eval_loss": 0.2343638390302658, |
|
"eval_runtime": 2080.7719, |
|
"eval_samples_per_second": 0.906, |
|
"eval_steps_per_second": 0.113, |
|
"eval_wer": 29.293205156215862, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 8.349113924050634e-06, |
|
"loss": 0.2332, |
|
"step": 7025 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 8.342784810126584e-06, |
|
"loss": 0.1907, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 8.336455696202531e-06, |
|
"loss": 0.2151, |
|
"step": 7075 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 8.330126582278481e-06, |
|
"loss": 0.2338, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 8.32379746835443e-06, |
|
"loss": 0.2123, |
|
"step": 7125 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 8.31746835443038e-06, |
|
"loss": 0.227, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 8.31113924050633e-06, |
|
"loss": 0.2196, |
|
"step": 7175 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 8.304810126582279e-06, |
|
"loss": 0.2308, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 8.298481012658228e-06, |
|
"loss": 0.2076, |
|
"step": 7225 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 8.292151898734177e-06, |
|
"loss": 0.2151, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 8.285822784810127e-06, |
|
"loss": 0.2178, |
|
"step": 7275 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 8.279493670886076e-06, |
|
"loss": 0.2203, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 8.273164556962026e-06, |
|
"loss": 0.2082, |
|
"step": 7325 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 8.266835443037975e-06, |
|
"loss": 0.2151, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 8.260506329113925e-06, |
|
"loss": 0.229, |
|
"step": 7375 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 8.254177215189874e-06, |
|
"loss": 0.208, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 8.247848101265823e-06, |
|
"loss": 0.21, |
|
"step": 7425 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 8.241518987341773e-06, |
|
"loss": 0.2263, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 8.235189873417722e-06, |
|
"loss": 0.2324, |
|
"step": 7475 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 8.228860759493672e-06, |
|
"loss": 0.2293, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 8.222531645569621e-06, |
|
"loss": 0.2201, |
|
"step": 7525 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 8.21620253164557e-06, |
|
"loss": 0.2229, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 8.20987341772152e-06, |
|
"loss": 0.2249, |
|
"step": 7575 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 8.20354430379747e-06, |
|
"loss": 0.2506, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 8.197215189873419e-06, |
|
"loss": 0.2068, |
|
"step": 7625 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 8.190886075949368e-06, |
|
"loss": 0.2197, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 8.184556962025318e-06, |
|
"loss": 0.2143, |
|
"step": 7675 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 8.178227848101267e-06, |
|
"loss": 0.1915, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 8.171898734177217e-06, |
|
"loss": 0.1992, |
|
"step": 7725 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 8.165569620253166e-06, |
|
"loss": 0.2133, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 8.159240506329115e-06, |
|
"loss": 0.2002, |
|
"step": 7775 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.152911392405065e-06, |
|
"loss": 0.2093, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.146582278481013e-06, |
|
"loss": 0.2117, |
|
"step": 7825 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.140253164556962e-06, |
|
"loss": 0.2178, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.133924050632911e-06, |
|
"loss": 0.2118, |
|
"step": 7875 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.127594936708861e-06, |
|
"loss": 0.2126, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 8.12126582278481e-06, |
|
"loss": 0.227, |
|
"step": 7925 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 8.11493670886076e-06, |
|
"loss": 0.2156, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 8.108607594936709e-06, |
|
"loss": 0.1955, |
|
"step": 7975 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 8.102278481012659e-06, |
|
"loss": 0.2071, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"eval_cer": 9.426063811866177, |
|
"eval_loss": 0.22829954326152802, |
|
"eval_runtime": 2075.4318, |
|
"eval_samples_per_second": 0.909, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 28.50666375355036, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.095949367088608e-06, |
|
"loss": 0.2234, |
|
"step": 8025 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.089620253164557e-06, |
|
"loss": 0.2353, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.083291139240507e-06, |
|
"loss": 0.2276, |
|
"step": 8075 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.076962025316456e-06, |
|
"loss": 0.2285, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.070632911392406e-06, |
|
"loss": 0.2199, |
|
"step": 8125 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.064303797468355e-06, |
|
"loss": 0.2118, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.057974683544305e-06, |
|
"loss": 0.2063, |
|
"step": 8175 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.051645569620254e-06, |
|
"loss": 0.2299, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.045316455696203e-06, |
|
"loss": 0.1978, |
|
"step": 8225 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.038987341772151e-06, |
|
"loss": 0.1999, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 8.032658227848102e-06, |
|
"loss": 0.205, |
|
"step": 8275 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 8.026329113924052e-06, |
|
"loss": 0.2107, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 8.020000000000001e-06, |
|
"loss": 0.2243, |
|
"step": 8325 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 8.01367088607595e-06, |
|
"loss": 0.2223, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 8.0073417721519e-06, |
|
"loss": 0.2009, |
|
"step": 8375 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 8.00101265822785e-06, |
|
"loss": 0.2009, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 7.994683544303799e-06, |
|
"loss": 0.1942, |
|
"step": 8425 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 7.988354430379748e-06, |
|
"loss": 0.2197, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 7.982025316455698e-06, |
|
"loss": 0.2101, |
|
"step": 8475 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 7.975696202531647e-06, |
|
"loss": 0.205, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 7.969367088607597e-06, |
|
"loss": 0.2103, |
|
"step": 8525 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 7.963037974683544e-06, |
|
"loss": 0.237, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 7.956708860759494e-06, |
|
"loss": 0.2074, |
|
"step": 8575 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 7.950379746835443e-06, |
|
"loss": 0.1982, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 7.944050632911393e-06, |
|
"loss": 0.2201, |
|
"step": 8625 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 7.937721518987342e-06, |
|
"loss": 0.2177, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 7.931392405063291e-06, |
|
"loss": 0.2266, |
|
"step": 8675 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 7.925063291139241e-06, |
|
"loss": 0.2224, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 7.91873417721519e-06, |
|
"loss": 0.2025, |
|
"step": 8725 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 7.91240506329114e-06, |
|
"loss": 0.2137, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 7.906075949367089e-06, |
|
"loss": 0.2019, |
|
"step": 8775 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 7.899746835443039e-06, |
|
"loss": 0.2054, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 7.893417721518988e-06, |
|
"loss": 0.2123, |
|
"step": 8825 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 7.887088607594937e-06, |
|
"loss": 0.2001, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 7.880759493670887e-06, |
|
"loss": 0.2229, |
|
"step": 8875 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 7.874430379746836e-06, |
|
"loss": 0.1963, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 7.868101265822786e-06, |
|
"loss": 0.2022, |
|
"step": 8925 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 7.861772151898735e-06, |
|
"loss": 0.1863, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 7.855443037974685e-06, |
|
"loss": 0.2168, |
|
"step": 8975 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 7.849113924050632e-06, |
|
"loss": 0.2051, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"eval_cer": 9.06556382871197, |
|
"eval_loss": 0.22236889600753784, |
|
"eval_runtime": 2053.8157, |
|
"eval_samples_per_second": 0.918, |
|
"eval_steps_per_second": 0.115, |
|
"eval_wer": 27.403320952589034, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 7.842784810126582e-06, |
|
"loss": 0.1973, |
|
"step": 9025 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 7.836455696202531e-06, |
|
"loss": 0.2398, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 7.83012658227848e-06, |
|
"loss": 0.2131, |
|
"step": 9075 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 7.823797468354432e-06, |
|
"loss": 0.2031, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 7.817468354430381e-06, |
|
"loss": 0.2045, |
|
"step": 9125 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 7.81113924050633e-06, |
|
"loss": 0.203, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 7.80481012658228e-06, |
|
"loss": 0.2042, |
|
"step": 9175 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 7.79848101265823e-06, |
|
"loss": 0.2211, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 7.792151898734179e-06, |
|
"loss": 0.2226, |
|
"step": 9225 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 7.785822784810128e-06, |
|
"loss": 0.2159, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 7.779493670886078e-06, |
|
"loss": 0.204, |
|
"step": 9275 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 7.773164556962025e-06, |
|
"loss": 0.2032, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 7.766835443037975e-06, |
|
"loss": 0.2126, |
|
"step": 9325 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 7.760506329113924e-06, |
|
"loss": 0.2256, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 7.754177215189874e-06, |
|
"loss": 0.1887, |
|
"step": 9375 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 7.747848101265823e-06, |
|
"loss": 0.2088, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.741518987341773e-06, |
|
"loss": 0.2112, |
|
"step": 9425 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.735189873417722e-06, |
|
"loss": 0.1933, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.728860759493671e-06, |
|
"loss": 0.2005, |
|
"step": 9475 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.722531645569621e-06, |
|
"loss": 0.2181, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.71620253164557e-06, |
|
"loss": 0.1877, |
|
"step": 9525 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.70987341772152e-06, |
|
"loss": 0.2094, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.703797468354431e-06, |
|
"loss": 0.2233, |
|
"step": 9575 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.69746835443038e-06, |
|
"loss": 0.2073, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.69113924050633e-06, |
|
"loss": 0.2211, |
|
"step": 9625 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 7.68481012658228e-06, |
|
"loss": 0.1978, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 7.678481012658229e-06, |
|
"loss": 0.2195, |
|
"step": 9675 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 7.672151898734178e-06, |
|
"loss": 0.2102, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 7.665822784810128e-06, |
|
"loss": 0.2057, |
|
"step": 9725 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 7.659493670886077e-06, |
|
"loss": 0.2119, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 7.653164556962027e-06, |
|
"loss": 0.2113, |
|
"step": 9775 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 7.646835443037976e-06, |
|
"loss": 0.1976, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 7.640506329113925e-06, |
|
"loss": 0.2192, |
|
"step": 9825 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 7.634177215189875e-06, |
|
"loss": 0.2044, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 7.6278481012658234e-06, |
|
"loss": 0.1941, |
|
"step": 9875 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 7.621518987341773e-06, |
|
"loss": 0.2056, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 7.615189873417722e-06, |
|
"loss": 0.1973, |
|
"step": 9925 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 7.608860759493672e-06, |
|
"loss": 0.2291, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 7.602531645569621e-06, |
|
"loss": 0.2398, |
|
"step": 9975 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 7.59620253164557e-06, |
|
"loss": 0.2064, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"eval_cer": 8.785081365183114, |
|
"eval_loss": 0.21381241083145142, |
|
"eval_runtime": 2070.2248, |
|
"eval_samples_per_second": 0.911, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 26.720559318330782, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 7.589873417721519e-06, |
|
"loss": 0.1906, |
|
"step": 10025 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 7.583544303797469e-06, |
|
"loss": 0.1896, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 7.577215189873418e-06, |
|
"loss": 0.198, |
|
"step": 10075 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 7.5708860759493674e-06, |
|
"loss": 0.2106, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.564556962025317e-06, |
|
"loss": 0.2085, |
|
"step": 10125 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.558227848101266e-06, |
|
"loss": 0.2013, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.551898734177216e-06, |
|
"loss": 0.2103, |
|
"step": 10175 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.545569620253165e-06, |
|
"loss": 0.2084, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.539240506329114e-06, |
|
"loss": 0.1906, |
|
"step": 10225 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.532911392405063e-06, |
|
"loss": 0.2018, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.526582278481013e-06, |
|
"loss": 0.2012, |
|
"step": 10275 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.520253164556963e-06, |
|
"loss": 0.197, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 7.513924050632912e-06, |
|
"loss": 0.1894, |
|
"step": 10325 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 7.507594936708862e-06, |
|
"loss": 0.2194, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 7.501265822784811e-06, |
|
"loss": 0.1977, |
|
"step": 10375 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 7.494936708860761e-06, |
|
"loss": 0.1955, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 7.48860759493671e-06, |
|
"loss": 0.2095, |
|
"step": 10425 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 7.4822784810126594e-06, |
|
"loss": 0.2067, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 7.475949367088609e-06, |
|
"loss": 0.211, |
|
"step": 10475 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 7.4696202531645574e-06, |
|
"loss": 0.1978, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 7.463291139240507e-06, |
|
"loss": 0.2346, |
|
"step": 10525 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 7.456962025316456e-06, |
|
"loss": 0.2047, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 7.450632911392406e-06, |
|
"loss": 0.227, |
|
"step": 10575 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 7.444303797468355e-06, |
|
"loss": 0.1832, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 7.437974683544305e-06, |
|
"loss": 0.1913, |
|
"step": 10625 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 7.431645569620254e-06, |
|
"loss": 0.1885, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 7.4253164556962034e-06, |
|
"loss": 0.2216, |
|
"step": 10675 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 7.418987341772153e-06, |
|
"loss": 0.2003, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 7.4126582278481014e-06, |
|
"loss": 0.2024, |
|
"step": 10725 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 7.406329113924051e-06, |
|
"loss": 0.2049, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 7.4e-06, |
|
"loss": 0.1988, |
|
"step": 10775 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 7.39367088607595e-06, |
|
"loss": 0.2114, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 7.387341772151899e-06, |
|
"loss": 0.1769, |
|
"step": 10825 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 7.381012658227849e-06, |
|
"loss": 0.2016, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 7.374683544303798e-06, |
|
"loss": 0.1893, |
|
"step": 10875 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 7.3683544303797474e-06, |
|
"loss": 0.2159, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 7.362025316455697e-06, |
|
"loss": 0.1952, |
|
"step": 10925 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 7.3556962025316454e-06, |
|
"loss": 0.2119, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 7.349367088607595e-06, |
|
"loss": 0.2128, |
|
"step": 10975 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 7.343037974683544e-06, |
|
"loss": 0.193, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"eval_cer": 8.502072032613455, |
|
"eval_loss": 0.20893645286560059, |
|
"eval_runtime": 2059.0007, |
|
"eval_samples_per_second": 0.916, |
|
"eval_steps_per_second": 0.115, |
|
"eval_wer": 25.57898186585099, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.336708860759494e-06, |
|
"loss": 0.1933, |
|
"step": 11025 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.330379746835443e-06, |
|
"loss": 0.1912, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.324050632911393e-06, |
|
"loss": 0.2052, |
|
"step": 11075 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.317721518987342e-06, |
|
"loss": 0.2175, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.311392405063292e-06, |
|
"loss": 0.2193, |
|
"step": 11125 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 7.305063291139242e-06, |
|
"loss": 0.2016, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 7.298734177215191e-06, |
|
"loss": 0.1815, |
|
"step": 11175 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 7.292658227848102e-06, |
|
"loss": 0.1804, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 7.286329113924051e-06, |
|
"loss": 0.1717, |
|
"step": 11225 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 7.280000000000001e-06, |
|
"loss": 0.21, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 7.27367088607595e-06, |
|
"loss": 0.2051, |
|
"step": 11275 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 7.2673417721518995e-06, |
|
"loss": 0.2048, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 7.261012658227848e-06, |
|
"loss": 0.1996, |
|
"step": 11325 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 7.2546835443037975e-06, |
|
"loss": 0.183, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.248354430379747e-06, |
|
"loss": 0.1968, |
|
"step": 11375 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.242025316455696e-06, |
|
"loss": 0.2058, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.235696202531646e-06, |
|
"loss": 0.191, |
|
"step": 11425 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.229367088607595e-06, |
|
"loss": 0.1908, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.223037974683545e-06, |
|
"loss": 0.2063, |
|
"step": 11475 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.216708860759494e-06, |
|
"loss": 0.1933, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.210379746835444e-06, |
|
"loss": 0.1887, |
|
"step": 11525 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.204050632911394e-06, |
|
"loss": 0.1878, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.197721518987343e-06, |
|
"loss": 0.1483, |
|
"step": 11575 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.191392405063293e-06, |
|
"loss": 0.1558, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 7.185063291139241e-06, |
|
"loss": 0.1367, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 7.178734177215191e-06, |
|
"loss": 0.1333, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 7.17240506329114e-06, |
|
"loss": 0.1374, |
|
"step": 11675 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 7.1660759493670895e-06, |
|
"loss": 0.1369, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 7.159746835443039e-06, |
|
"loss": 0.1344, |
|
"step": 11725 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 7.153417721518988e-06, |
|
"loss": 0.1592, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 7.147088607594938e-06, |
|
"loss": 0.1457, |
|
"step": 11775 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 7.140759493670887e-06, |
|
"loss": 0.1481, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 7.134430379746836e-06, |
|
"loss": 0.1624, |
|
"step": 11825 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.128101265822785e-06, |
|
"loss": 0.1467, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.121772151898735e-06, |
|
"loss": 0.1471, |
|
"step": 11875 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.115443037974684e-06, |
|
"loss": 0.1477, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.1091139240506335e-06, |
|
"loss": 0.1358, |
|
"step": 11925 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.102784810126583e-06, |
|
"loss": 0.158, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.096455696202532e-06, |
|
"loss": 0.1441, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.090126582278482e-06, |
|
"loss": 0.1577, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"eval_cer": 8.287288164145414, |
|
"eval_loss": 0.20717626810073853, |
|
"eval_runtime": 2051.8904, |
|
"eval_samples_per_second": 0.919, |
|
"eval_steps_per_second": 0.115, |
|
"eval_wer": 25.61175442429539, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.083797468354431e-06, |
|
"loss": 0.1462, |
|
"step": 12025 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.07746835443038e-06, |
|
"loss": 0.1584, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 7.071139240506329e-06, |
|
"loss": 0.1432, |
|
"step": 12075 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 7.064810126582279e-06, |
|
"loss": 0.1537, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 7.058481012658228e-06, |
|
"loss": 0.1405, |
|
"step": 12125 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 7.0521518987341775e-06, |
|
"loss": 0.1437, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 7.045822784810127e-06, |
|
"loss": 0.133, |
|
"step": 12175 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.039493670886076e-06, |
|
"loss": 0.1292, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.033164556962026e-06, |
|
"loss": 0.138, |
|
"step": 12225 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.026835443037975e-06, |
|
"loss": 0.1512, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.020506329113924e-06, |
|
"loss": 0.1648, |
|
"step": 12275 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 7.014177215189873e-06, |
|
"loss": 0.1338, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 7.007848101265823e-06, |
|
"loss": 0.1438, |
|
"step": 12325 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 7.001518987341773e-06, |
|
"loss": 0.1646, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 6.995189873417722e-06, |
|
"loss": 0.1359, |
|
"step": 12375 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 6.988860759493672e-06, |
|
"loss": 0.1322, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 6.982531645569621e-06, |
|
"loss": 0.136, |
|
"step": 12425 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 6.976202531645571e-06, |
|
"loss": 0.1429, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 6.96987341772152e-06, |
|
"loss": 0.134, |
|
"step": 12475 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 6.9635443037974694e-06, |
|
"loss": 0.1382, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 6.957215189873419e-06, |
|
"loss": 0.1354, |
|
"step": 12525 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 6.950886075949368e-06, |
|
"loss": 0.1374, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 6.944556962025317e-06, |
|
"loss": 0.1572, |
|
"step": 12575 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 6.938227848101266e-06, |
|
"loss": 0.1424, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 6.931898734177216e-06, |
|
"loss": 0.157, |
|
"step": 12625 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 6.925569620253165e-06, |
|
"loss": 0.1371, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 6.919240506329115e-06, |
|
"loss": 0.1518, |
|
"step": 12675 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 6.912911392405064e-06, |
|
"loss": 0.1394, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 6.9065822784810134e-06, |
|
"loss": 0.1479, |
|
"step": 12725 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 6.900253164556963e-06, |
|
"loss": 0.1554, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 6.893924050632912e-06, |
|
"loss": 0.1405, |
|
"step": 12775 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 6.887594936708861e-06, |
|
"loss": 0.1192, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 6.88126582278481e-06, |
|
"loss": 0.1475, |
|
"step": 12825 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 6.87493670886076e-06, |
|
"loss": 0.1359, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 6.868607594936709e-06, |
|
"loss": 0.1453, |
|
"step": 12875 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 6.862278481012659e-06, |
|
"loss": 0.1368, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 6.855949367088608e-06, |
|
"loss": 0.1522, |
|
"step": 12925 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 6.8496202531645574e-06, |
|
"loss": 0.1377, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 6.843291139240507e-06, |
|
"loss": 0.1433, |
|
"step": 12975 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 6.836962025316456e-06, |
|
"loss": 0.1397, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"eval_cer": 8.236750783329402, |
|
"eval_loss": 0.2046031802892685, |
|
"eval_runtime": 2062.3221, |
|
"eval_samples_per_second": 0.915, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 25.114703954555385, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 6.830632911392405e-06, |
|
"loss": 0.1588, |
|
"step": 13025 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 6.824303797468354e-06, |
|
"loss": 0.1379, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 6.817974683544304e-06, |
|
"loss": 0.1337, |
|
"step": 13075 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 6.811645569620253e-06, |
|
"loss": 0.1418, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 6.805316455696203e-06, |
|
"loss": 0.1267, |
|
"step": 13125 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 6.798987341772152e-06, |
|
"loss": 0.1493, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 6.792658227848102e-06, |
|
"loss": 0.149, |
|
"step": 13175 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 6.786329113924052e-06, |
|
"loss": 0.1369, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 6.780000000000001e-06, |
|
"loss": 0.1425, |
|
"step": 13225 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 6.773670886075951e-06, |
|
"loss": 0.1495, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 6.7673417721519e-06, |
|
"loss": 0.1534, |
|
"step": 13275 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 6.7610126582278494e-06, |
|
"loss": 0.1432, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 6.754683544303798e-06, |
|
"loss": 0.1454, |
|
"step": 13325 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 6.7483544303797474e-06, |
|
"loss": 0.1409, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 6.742025316455697e-06, |
|
"loss": 0.1553, |
|
"step": 13375 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 6.735696202531646e-06, |
|
"loss": 0.1577, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 6.729367088607596e-06, |
|
"loss": 0.1543, |
|
"step": 13425 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 6.723037974683545e-06, |
|
"loss": 0.1415, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 6.716708860759495e-06, |
|
"loss": 0.1488, |
|
"step": 13475 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 6.710379746835444e-06, |
|
"loss": 0.1416, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 6.7040506329113934e-06, |
|
"loss": 0.1553, |
|
"step": 13525 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 6.697721518987342e-06, |
|
"loss": 0.1331, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 6.6913924050632914e-06, |
|
"loss": 0.1526, |
|
"step": 13575 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 6.685063291139241e-06, |
|
"loss": 0.137, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 6.67873417721519e-06, |
|
"loss": 0.141, |
|
"step": 13625 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 6.67240506329114e-06, |
|
"loss": 0.1446, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 6.666075949367089e-06, |
|
"loss": 0.1545, |
|
"step": 13675 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.659746835443039e-06, |
|
"loss": 0.1503, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.653417721518988e-06, |
|
"loss": 0.1584, |
|
"step": 13725 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.647088607594937e-06, |
|
"loss": 0.1499, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.640759493670886e-06, |
|
"loss": 0.1386, |
|
"step": 13775 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.6344303797468355e-06, |
|
"loss": 0.1357, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.628101265822785e-06, |
|
"loss": 0.1297, |
|
"step": 13825 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.621772151898734e-06, |
|
"loss": 0.139, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.615443037974684e-06, |
|
"loss": 0.129, |
|
"step": 13875 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.609113924050633e-06, |
|
"loss": 0.1353, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 6.602784810126583e-06, |
|
"loss": 0.1526, |
|
"step": 13925 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 6.596455696202532e-06, |
|
"loss": 0.146, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 6.590126582278481e-06, |
|
"loss": 0.1413, |
|
"step": 13975 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 6.583797468354432e-06, |
|
"loss": 0.1526, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"eval_cer": 8.76149725413564, |
|
"eval_loss": 0.2065056562423706, |
|
"eval_runtime": 2128.998, |
|
"eval_samples_per_second": 0.886, |
|
"eval_steps_per_second": 0.111, |
|
"eval_wer": 26.463840943849682, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 6.577468354430381e-06, |
|
"loss": 0.1356, |
|
"step": 14025 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 6.57113924050633e-06, |
|
"loss": 0.1356, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 6.564810126582279e-06, |
|
"loss": 0.1495, |
|
"step": 14075 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 6.558481012658229e-06, |
|
"loss": 0.1395, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 6.552151898734178e-06, |
|
"loss": 0.1504, |
|
"step": 14125 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 6.5458227848101274e-06, |
|
"loss": 0.1401, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 6.539493670886077e-06, |
|
"loss": 0.1373, |
|
"step": 14175 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 6.533164556962026e-06, |
|
"loss": 0.1375, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 6.526835443037976e-06, |
|
"loss": 0.1381, |
|
"step": 14225 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 6.520506329113925e-06, |
|
"loss": 0.1547, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 6.514177215189874e-06, |
|
"loss": 0.156, |
|
"step": 14275 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 6.507848101265823e-06, |
|
"loss": 0.1397, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 6.501518987341773e-06, |
|
"loss": 0.1502, |
|
"step": 14325 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 6.495189873417722e-06, |
|
"loss": 0.1458, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 6.4888607594936714e-06, |
|
"loss": 0.1579, |
|
"step": 14375 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 6.482531645569621e-06, |
|
"loss": 0.1413, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 6.47620253164557e-06, |
|
"loss": 0.1282, |
|
"step": 14425 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 6.46987341772152e-06, |
|
"loss": 0.1393, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 6.463544303797469e-06, |
|
"loss": 0.1558, |
|
"step": 14475 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 6.457215189873418e-06, |
|
"loss": 0.1413, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 6.450886075949367e-06, |
|
"loss": 0.1474, |
|
"step": 14525 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 6.444556962025317e-06, |
|
"loss": 0.1644, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 6.438227848101266e-06, |
|
"loss": 0.1319, |
|
"step": 14575 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 6.4318987341772154e-06, |
|
"loss": 0.155, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 6.425569620253165e-06, |
|
"loss": 0.1437, |
|
"step": 14625 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 6.419240506329114e-06, |
|
"loss": 0.1397, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 6.412911392405064e-06, |
|
"loss": 0.1493, |
|
"step": 14675 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 6.406582278481013e-06, |
|
"loss": 0.1442, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 6.400253164556962e-06, |
|
"loss": 0.1541, |
|
"step": 14725 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 6.393924050632911e-06, |
|
"loss": 0.1413, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 6.387594936708861e-06, |
|
"loss": 0.1343, |
|
"step": 14775 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 6.38126582278481e-06, |
|
"loss": 0.1554, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 6.37493670886076e-06, |
|
"loss": 0.161, |
|
"step": 14825 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 6.36860759493671e-06, |
|
"loss": 0.14, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 6.362278481012659e-06, |
|
"loss": 0.1393, |
|
"step": 14875 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 6.3559493670886086e-06, |
|
"loss": 0.1316, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 6.349620253164558e-06, |
|
"loss": 0.1433, |
|
"step": 14925 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.3432911392405074e-06, |
|
"loss": 0.1414, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.336962025316457e-06, |
|
"loss": 0.1691, |
|
"step": 14975 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.330632911392406e-06, |
|
"loss": 0.1497, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"eval_cer": 7.9587951888413455, |
|
"eval_loss": 0.20039086043834686, |
|
"eval_runtime": 2071.9347, |
|
"eval_samples_per_second": 0.91, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 24.4865632510378, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.324303797468355e-06, |
|
"loss": 0.1395, |
|
"step": 15025 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.317974683544304e-06, |
|
"loss": 0.1484, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.311645569620254e-06, |
|
"loss": 0.1391, |
|
"step": 15075 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.305316455696203e-06, |
|
"loss": 0.1316, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.2989873417721526e-06, |
|
"loss": 0.1555, |
|
"step": 15125 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.292658227848102e-06, |
|
"loss": 0.1371, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.2863291139240514e-06, |
|
"loss": 0.1477, |
|
"step": 15175 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.280000000000001e-06, |
|
"loss": 0.1382, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.27367088607595e-06, |
|
"loss": 0.146, |
|
"step": 15225 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.267341772151899e-06, |
|
"loss": 0.1363, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.261012658227848e-06, |
|
"loss": 0.1415, |
|
"step": 15275 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.254683544303798e-06, |
|
"loss": 0.1535, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.248354430379747e-06, |
|
"loss": 0.1415, |
|
"step": 15325 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.242025316455697e-06, |
|
"loss": 0.1527, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.235696202531646e-06, |
|
"loss": 0.1305, |
|
"step": 15375 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.2293670886075954e-06, |
|
"loss": 0.1487, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.223037974683545e-06, |
|
"loss": 0.1513, |
|
"step": 15425 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.216708860759494e-06, |
|
"loss": 0.1359, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.210379746835443e-06, |
|
"loss": 0.1344, |
|
"step": 15475 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.204050632911392e-06, |
|
"loss": 0.1517, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.197721518987342e-06, |
|
"loss": 0.1446, |
|
"step": 15525 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.191392405063291e-06, |
|
"loss": 0.1314, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.185063291139241e-06, |
|
"loss": 0.1452, |
|
"step": 15575 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.17873417721519e-06, |
|
"loss": 0.1507, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.1724050632911394e-06, |
|
"loss": 0.1332, |
|
"step": 15625 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 6.16607594936709e-06, |
|
"loss": 0.1318, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 6.159746835443039e-06, |
|
"loss": 0.1457, |
|
"step": 15675 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 6.1534177215189886e-06, |
|
"loss": 0.1468, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 6.147088607594938e-06, |
|
"loss": 0.1339, |
|
"step": 15725 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 6.1407594936708866e-06, |
|
"loss": 0.139, |
|
"step": 15750 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 6.134430379746836e-06, |
|
"loss": 0.1347, |
|
"step": 15775 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 6.1281012658227854e-06, |
|
"loss": 0.1395, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 6.121772151898735e-06, |
|
"loss": 0.1403, |
|
"step": 15825 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 6.115443037974684e-06, |
|
"loss": 0.1465, |
|
"step": 15850 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 6.109113924050634e-06, |
|
"loss": 0.1483, |
|
"step": 15875 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 6.102784810126583e-06, |
|
"loss": 0.1368, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 6.0964556962025326e-06, |
|
"loss": 0.138, |
|
"step": 15925 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 6.090126582278482e-06, |
|
"loss": 0.1592, |
|
"step": 15950 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 6.0837974683544306e-06, |
|
"loss": 0.1437, |
|
"step": 15975 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.07746835443038e-06, |
|
"loss": 0.1569, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"eval_cer": 7.955426030120279, |
|
"eval_loss": 0.19895771145820618, |
|
"eval_runtime": 2076.4885, |
|
"eval_samples_per_second": 0.908, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 24.224382783482632, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.0711392405063294e-06, |
|
"loss": 0.1337, |
|
"step": 16025 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.064810126582279e-06, |
|
"loss": 0.1441, |
|
"step": 16050 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.058481012658228e-06, |
|
"loss": 0.1584, |
|
"step": 16075 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.052151898734178e-06, |
|
"loss": 0.1373, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 6.045822784810127e-06, |
|
"loss": 0.1582, |
|
"step": 16125 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 6.0394936708860766e-06, |
|
"loss": 0.1573, |
|
"step": 16150 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 6.033164556962026e-06, |
|
"loss": 0.1473, |
|
"step": 16175 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 6.026835443037975e-06, |
|
"loss": 0.1447, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 6.020506329113924e-06, |
|
"loss": 0.1351, |
|
"step": 16225 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 6.0141772151898734e-06, |
|
"loss": 0.1511, |
|
"step": 16250 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 6.007848101265823e-06, |
|
"loss": 0.1321, |
|
"step": 16275 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 6.001518987341772e-06, |
|
"loss": 0.1316, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.995189873417722e-06, |
|
"loss": 0.1323, |
|
"step": 16325 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.988860759493671e-06, |
|
"loss": 0.1338, |
|
"step": 16350 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.9825316455696206e-06, |
|
"loss": 0.1463, |
|
"step": 16375 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.97620253164557e-06, |
|
"loss": 0.1366, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.969873417721519e-06, |
|
"loss": 0.1553, |
|
"step": 16425 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.963544303797468e-06, |
|
"loss": 0.1234, |
|
"step": 16450 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.957215189873419e-06, |
|
"loss": 0.1343, |
|
"step": 16475 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.950886075949368e-06, |
|
"loss": 0.1409, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.944556962025317e-06, |
|
"loss": 0.1309, |
|
"step": 16525 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.9382278481012666e-06, |
|
"loss": 0.1477, |
|
"step": 16550 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.931898734177216e-06, |
|
"loss": 0.1319, |
|
"step": 16575 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.925569620253165e-06, |
|
"loss": 0.1388, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.919240506329115e-06, |
|
"loss": 0.1441, |
|
"step": 16625 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.912911392405064e-06, |
|
"loss": 0.1496, |
|
"step": 16650 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.906582278481014e-06, |
|
"loss": 0.1344, |
|
"step": 16675 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.900253164556963e-06, |
|
"loss": 0.1241, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.893924050632912e-06, |
|
"loss": 0.1369, |
|
"step": 16725 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.887594936708861e-06, |
|
"loss": 0.1415, |
|
"step": 16750 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.8812658227848106e-06, |
|
"loss": 0.1236, |
|
"step": 16775 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.87493670886076e-06, |
|
"loss": 0.1434, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.8686075949367094e-06, |
|
"loss": 0.1358, |
|
"step": 16825 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.862278481012659e-06, |
|
"loss": 0.1449, |
|
"step": 16850 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.855949367088608e-06, |
|
"loss": 0.1589, |
|
"step": 16875 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.849620253164558e-06, |
|
"loss": 0.157, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.843291139240507e-06, |
|
"loss": 0.1453, |
|
"step": 16925 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.836962025316456e-06, |
|
"loss": 0.1423, |
|
"step": 16950 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.830632911392405e-06, |
|
"loss": 0.1472, |
|
"step": 16975 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.8243037974683546e-06, |
|
"loss": 0.1416, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"eval_cer": 7.87540851049493, |
|
"eval_loss": 0.20009736716747284, |
|
"eval_runtime": 2089.8591, |
|
"eval_samples_per_second": 0.902, |
|
"eval_steps_per_second": 0.113, |
|
"eval_wer": 24.229844876556697, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.817974683544304e-06, |
|
"loss": 0.143, |
|
"step": 17025 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.8116455696202534e-06, |
|
"loss": 0.1402, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.805316455696203e-06, |
|
"loss": 0.1229, |
|
"step": 17075 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.798987341772152e-06, |
|
"loss": 0.1343, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.792658227848102e-06, |
|
"loss": 0.1397, |
|
"step": 17125 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.786329113924051e-06, |
|
"loss": 0.153, |
|
"step": 17150 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.78e-06, |
|
"loss": 0.1356, |
|
"step": 17175 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.773670886075949e-06, |
|
"loss": 0.1403, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.7673417721518986e-06, |
|
"loss": 0.1481, |
|
"step": 17225 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.761012658227848e-06, |
|
"loss": 0.1443, |
|
"step": 17250 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 5.7546835443037974e-06, |
|
"loss": 0.1401, |
|
"step": 17275 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 5.748354430379748e-06, |
|
"loss": 0.1434, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 5.742025316455697e-06, |
|
"loss": 0.1406, |
|
"step": 17325 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 5.7356962025316466e-06, |
|
"loss": 0.1364, |
|
"step": 17350 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 5.729367088607596e-06, |
|
"loss": 0.1382, |
|
"step": 17375 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 5.723037974683545e-06, |
|
"loss": 0.1348, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 5.716708860759495e-06, |
|
"loss": 0.1431, |
|
"step": 17425 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 5.710379746835443e-06, |
|
"loss": 0.1534, |
|
"step": 17450 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 5.704050632911393e-06, |
|
"loss": 0.1421, |
|
"step": 17475 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 5.697721518987342e-06, |
|
"loss": 0.1334, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 5.691392405063292e-06, |
|
"loss": 0.1496, |
|
"step": 17525 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 5.685063291139241e-06, |
|
"loss": 0.143, |
|
"step": 17550 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 5.6787341772151906e-06, |
|
"loss": 0.1432, |
|
"step": 17575 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 5.67240506329114e-06, |
|
"loss": 0.1331, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 5.666075949367089e-06, |
|
"loss": 0.1313, |
|
"step": 17625 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 5.659746835443039e-06, |
|
"loss": 0.1534, |
|
"step": 17650 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 5.6534177215189874e-06, |
|
"loss": 0.128, |
|
"step": 17675 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 5.647088607594937e-06, |
|
"loss": 0.1429, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 5.640759493670886e-06, |
|
"loss": 0.142, |
|
"step": 17725 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 5.634430379746836e-06, |
|
"loss": 0.1366, |
|
"step": 17750 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 5.628101265822785e-06, |
|
"loss": 0.1338, |
|
"step": 17775 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 5.6217721518987346e-06, |
|
"loss": 0.1343, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 5.615443037974684e-06, |
|
"loss": 0.1632, |
|
"step": 17825 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 5.609113924050633e-06, |
|
"loss": 0.1306, |
|
"step": 17850 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 5.602784810126583e-06, |
|
"loss": 0.1436, |
|
"step": 17875 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 5.5964556962025314e-06, |
|
"loss": 0.1321, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 5.590126582278481e-06, |
|
"loss": 0.151, |
|
"step": 17925 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 5.58379746835443e-06, |
|
"loss": 0.151, |
|
"step": 17950 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 5.577721518987343e-06, |
|
"loss": 0.1516, |
|
"step": 17975 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 5.571392405063292e-06, |
|
"loss": 0.1371, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"eval_cer": 7.807183046393315, |
|
"eval_loss": 0.19323883950710297, |
|
"eval_runtime": 2071.9502, |
|
"eval_samples_per_second": 0.91, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 23.607166266113175, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 5.5650632911392414e-06, |
|
"loss": 0.1465, |
|
"step": 18025 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 5.55873417721519e-06, |
|
"loss": 0.1299, |
|
"step": 18050 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 5.5524050632911394e-06, |
|
"loss": 0.1288, |
|
"step": 18075 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 5.546329113924052e-06, |
|
"loss": 0.1412, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 5.540000000000001e-06, |
|
"loss": 0.1323, |
|
"step": 18125 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 5.533670886075951e-06, |
|
"loss": 0.1342, |
|
"step": 18150 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 5.527341772151899e-06, |
|
"loss": 0.1284, |
|
"step": 18175 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 5.521012658227849e-06, |
|
"loss": 0.1415, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 5.514683544303798e-06, |
|
"loss": 0.1454, |
|
"step": 18225 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 5.5083544303797475e-06, |
|
"loss": 0.1445, |
|
"step": 18250 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 5.502025316455697e-06, |
|
"loss": 0.1516, |
|
"step": 18275 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 5.495696202531646e-06, |
|
"loss": 0.1443, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 5.489367088607596e-06, |
|
"loss": 0.1323, |
|
"step": 18325 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 5.483037974683545e-06, |
|
"loss": 0.1452, |
|
"step": 18350 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 5.476708860759495e-06, |
|
"loss": 0.1389, |
|
"step": 18375 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 5.470379746835443e-06, |
|
"loss": 0.1406, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 5.464050632911393e-06, |
|
"loss": 0.1503, |
|
"step": 18425 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 5.457721518987342e-06, |
|
"loss": 0.1405, |
|
"step": 18450 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 5.4513924050632915e-06, |
|
"loss": 0.1391, |
|
"step": 18475 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 5.445063291139241e-06, |
|
"loss": 0.1334, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 5.43873417721519e-06, |
|
"loss": 0.1383, |
|
"step": 18525 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.43240506329114e-06, |
|
"loss": 0.1422, |
|
"step": 18550 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.426075949367089e-06, |
|
"loss": 0.1335, |
|
"step": 18575 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.419746835443039e-06, |
|
"loss": 0.1291, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.413417721518987e-06, |
|
"loss": 0.1428, |
|
"step": 18625 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.407088607594937e-06, |
|
"loss": 0.1248, |
|
"step": 18650 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.400759493670886e-06, |
|
"loss": 0.1362, |
|
"step": 18675 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.3944303797468355e-06, |
|
"loss": 0.1376, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.388101265822785e-06, |
|
"loss": 0.1409, |
|
"step": 18725 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.381772151898734e-06, |
|
"loss": 0.1298, |
|
"step": 18750 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.375443037974684e-06, |
|
"loss": 0.1392, |
|
"step": 18775 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.369113924050633e-06, |
|
"loss": 0.1327, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.362784810126583e-06, |
|
"loss": 0.1239, |
|
"step": 18825 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.356455696202531e-06, |
|
"loss": 0.1375, |
|
"step": 18850 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.350126582278481e-06, |
|
"loss": 0.1516, |
|
"step": 18875 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.343797468354432e-06, |
|
"loss": 0.1355, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.33746835443038e-06, |
|
"loss": 0.1368, |
|
"step": 18925 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.33113924050633e-06, |
|
"loss": 0.1522, |
|
"step": 18950 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.324810126582279e-06, |
|
"loss": 0.1294, |
|
"step": 18975 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.318481012658229e-06, |
|
"loss": 0.1379, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"eval_cer": 7.545230955830329, |
|
"eval_loss": 0.19160360097885132, |
|
"eval_runtime": 2069.0144, |
|
"eval_samples_per_second": 0.912, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 23.131964168669434, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.312151898734178e-06, |
|
"loss": 0.1471, |
|
"step": 19025 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.3058227848101275e-06, |
|
"loss": 0.1262, |
|
"step": 19050 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.299493670886077e-06, |
|
"loss": 0.1417, |
|
"step": 19075 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.293164556962026e-06, |
|
"loss": 0.1264, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.286835443037976e-06, |
|
"loss": 0.1443, |
|
"step": 19125 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.280506329113924e-06, |
|
"loss": 0.1262, |
|
"step": 19150 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.274177215189874e-06, |
|
"loss": 0.1398, |
|
"step": 19175 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.267848101265823e-06, |
|
"loss": 0.1342, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 5.261518987341773e-06, |
|
"loss": 0.1338, |
|
"step": 19225 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 5.255189873417722e-06, |
|
"loss": 0.1351, |
|
"step": 19250 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 5.2488607594936715e-06, |
|
"loss": 0.143, |
|
"step": 19275 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 5.242531645569621e-06, |
|
"loss": 0.1232, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 5.23620253164557e-06, |
|
"loss": 0.1231, |
|
"step": 19325 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 5.22987341772152e-06, |
|
"loss": 0.1432, |
|
"step": 19350 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 5.223544303797468e-06, |
|
"loss": 0.1473, |
|
"step": 19375 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 5.217215189873418e-06, |
|
"loss": 0.1207, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 5.210886075949367e-06, |
|
"loss": 0.1518, |
|
"step": 19425 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 5.204556962025317e-06, |
|
"loss": 0.1409, |
|
"step": 19450 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 5.198227848101266e-06, |
|
"loss": 0.1406, |
|
"step": 19475 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 5.1918987341772155e-06, |
|
"loss": 0.1349, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 5.185569620253165e-06, |
|
"loss": 0.1272, |
|
"step": 19525 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 5.179240506329114e-06, |
|
"loss": 0.1358, |
|
"step": 19550 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 5.172911392405064e-06, |
|
"loss": 0.1254, |
|
"step": 19575 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 5.166582278481012e-06, |
|
"loss": 0.1477, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 5.160253164556962e-06, |
|
"loss": 0.1446, |
|
"step": 19625 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 5.153924050632911e-06, |
|
"loss": 0.1321, |
|
"step": 19650 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 5.147594936708861e-06, |
|
"loss": 0.136, |
|
"step": 19675 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 5.14126582278481e-06, |
|
"loss": 0.1415, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 5.13493670886076e-06, |
|
"loss": 0.1466, |
|
"step": 19725 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 5.12860759493671e-06, |
|
"loss": 0.1425, |
|
"step": 19750 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 5.122278481012659e-06, |
|
"loss": 0.1229, |
|
"step": 19775 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 5.115949367088609e-06, |
|
"loss": 0.1319, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 5.109620253164558e-06, |
|
"loss": 0.1526, |
|
"step": 19825 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 5.1032911392405074e-06, |
|
"loss": 0.1406, |
|
"step": 19850 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 5.096962025316457e-06, |
|
"loss": 0.1395, |
|
"step": 19875 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 5.0906329113924055e-06, |
|
"loss": 0.1206, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 5.084303797468355e-06, |
|
"loss": 0.1348, |
|
"step": 19925 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 5.077974683544304e-06, |
|
"loss": 0.1302, |
|
"step": 19950 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 5.071645569620254e-06, |
|
"loss": 0.1396, |
|
"step": 19975 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 5.065316455696203e-06, |
|
"loss": 0.1305, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"eval_cer": 7.428994979953505, |
|
"eval_loss": 0.18803881108760834, |
|
"eval_runtime": 2070.7699, |
|
"eval_samples_per_second": 0.911, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 23.110115796373172, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 5.058987341772153e-06, |
|
"loss": 0.1211, |
|
"step": 20025 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 5.052658227848102e-06, |
|
"loss": 0.1339, |
|
"step": 20050 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 5.0463291139240514e-06, |
|
"loss": 0.134, |
|
"step": 20075 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 5.04e-06, |
|
"loss": 0.1406, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 5.0336708860759495e-06, |
|
"loss": 0.142, |
|
"step": 20125 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 5.027341772151899e-06, |
|
"loss": 0.118, |
|
"step": 20150 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 5.021012658227848e-06, |
|
"loss": 0.1259, |
|
"step": 20175 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 5.014683544303798e-06, |
|
"loss": 0.1381, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 5.008354430379747e-06, |
|
"loss": 0.1343, |
|
"step": 20225 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 5.002025316455697e-06, |
|
"loss": 0.1354, |
|
"step": 20250 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 4.995696202531646e-06, |
|
"loss": 0.1361, |
|
"step": 20275 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 4.9893670886075954e-06, |
|
"loss": 0.1407, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 4.983037974683544e-06, |
|
"loss": 0.1395, |
|
"step": 20325 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 4.976708860759494e-06, |
|
"loss": 0.1245, |
|
"step": 20350 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 4.970379746835444e-06, |
|
"loss": 0.1501, |
|
"step": 20375 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 4.964050632911393e-06, |
|
"loss": 0.1372, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 4.957721518987343e-06, |
|
"loss": 0.1331, |
|
"step": 20425 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 4.951392405063292e-06, |
|
"loss": 0.1404, |
|
"step": 20450 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 4.945063291139241e-06, |
|
"loss": 0.129, |
|
"step": 20475 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 4.93873417721519e-06, |
|
"loss": 0.1435, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 4.9324050632911394e-06, |
|
"loss": 0.1272, |
|
"step": 20525 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 4.926075949367089e-06, |
|
"loss": 0.1444, |
|
"step": 20550 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 4.919746835443038e-06, |
|
"loss": 0.1428, |
|
"step": 20575 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 4.913417721518988e-06, |
|
"loss": 0.1461, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 4.907088607594937e-06, |
|
"loss": 0.1395, |
|
"step": 20625 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 4.900759493670887e-06, |
|
"loss": 0.1343, |
|
"step": 20650 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 4.894430379746836e-06, |
|
"loss": 0.1518, |
|
"step": 20675 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 4.888101265822785e-06, |
|
"loss": 0.1281, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 4.881772151898734e-06, |
|
"loss": 0.1579, |
|
"step": 20725 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 4.875443037974684e-06, |
|
"loss": 0.1359, |
|
"step": 20750 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 4.869113924050634e-06, |
|
"loss": 0.1486, |
|
"step": 20775 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 4.862784810126583e-06, |
|
"loss": 0.1343, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 4.856455696202533e-06, |
|
"loss": 0.1374, |
|
"step": 20825 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 4.850126582278481e-06, |
|
"loss": 0.1371, |
|
"step": 20850 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 4.843797468354431e-06, |
|
"loss": 0.1395, |
|
"step": 20875 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 4.83746835443038e-06, |
|
"loss": 0.1322, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 4.8311392405063294e-06, |
|
"loss": 0.1442, |
|
"step": 20925 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 4.824810126582279e-06, |
|
"loss": 0.1463, |
|
"step": 20950 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.818481012658228e-06, |
|
"loss": 0.1428, |
|
"step": 20975 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.812151898734178e-06, |
|
"loss": 0.1395, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"eval_cer": 7.463528856844446, |
|
"eval_loss": 0.18767213821411133, |
|
"eval_runtime": 2069.0205, |
|
"eval_samples_per_second": 0.912, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 22.984487655669653, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.806075949367089e-06, |
|
"loss": 0.1501, |
|
"step": 21025 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.799746835443039e-06, |
|
"loss": 0.1353, |
|
"step": 21050 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 4.793417721518987e-06, |
|
"loss": 0.1354, |
|
"step": 21075 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 4.787088607594937e-06, |
|
"loss": 0.1361, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 4.780759493670886e-06, |
|
"loss": 0.1554, |
|
"step": 21125 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 4.774430379746836e-06, |
|
"loss": 0.1263, |
|
"step": 21150 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 4.768101265822786e-06, |
|
"loss": 0.1301, |
|
"step": 21175 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 4.761772151898735e-06, |
|
"loss": 0.1407, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 4.755696202531646e-06, |
|
"loss": 0.1346, |
|
"step": 21225 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 4.749367088607595e-06, |
|
"loss": 0.1409, |
|
"step": 21250 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 4.743037974683545e-06, |
|
"loss": 0.1421, |
|
"step": 21275 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 4.736708860759494e-06, |
|
"loss": 0.1438, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 4.7303797468354435e-06, |
|
"loss": 0.1214, |
|
"step": 21325 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 4.724050632911393e-06, |
|
"loss": 0.1371, |
|
"step": 21350 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 4.717721518987342e-06, |
|
"loss": 0.143, |
|
"step": 21375 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 4.711392405063292e-06, |
|
"loss": 0.1413, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 4.70506329113924e-06, |
|
"loss": 0.1274, |
|
"step": 21425 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 4.69873417721519e-06, |
|
"loss": 0.1295, |
|
"step": 21450 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 4.692405063291139e-06, |
|
"loss": 0.1348, |
|
"step": 21475 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 4.686075949367089e-06, |
|
"loss": 0.1397, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 4.679746835443038e-06, |
|
"loss": 0.1501, |
|
"step": 21525 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 4.673417721518988e-06, |
|
"loss": 0.1466, |
|
"step": 21550 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 4.667088607594937e-06, |
|
"loss": 0.133, |
|
"step": 21575 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 4.660759493670886e-06, |
|
"loss": 0.1401, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 4.654430379746836e-06, |
|
"loss": 0.1361, |
|
"step": 21625 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 4.648101265822785e-06, |
|
"loss": 0.1289, |
|
"step": 21650 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 4.641772151898735e-06, |
|
"loss": 0.1396, |
|
"step": 21675 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 4.635443037974684e-06, |
|
"loss": 0.1352, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 4.6291139240506335e-06, |
|
"loss": 0.1392, |
|
"step": 21725 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 4.622784810126583e-06, |
|
"loss": 0.139, |
|
"step": 21750 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 4.616455696202532e-06, |
|
"loss": 0.1443, |
|
"step": 21775 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 4.610126582278481e-06, |
|
"loss": 0.1239, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 4.60379746835443e-06, |
|
"loss": 0.1336, |
|
"step": 21825 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 4.59746835443038e-06, |
|
"loss": 0.1397, |
|
"step": 21850 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 4.591139240506329e-06, |
|
"loss": 0.1398, |
|
"step": 21875 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 4.584810126582279e-06, |
|
"loss": 0.1448, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 4.578481012658228e-06, |
|
"loss": 0.1271, |
|
"step": 21925 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 4.5721518987341775e-06, |
|
"loss": 0.1348, |
|
"step": 21950 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 4.565822784810127e-06, |
|
"loss": 0.1478, |
|
"step": 21975 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 4.559493670886076e-06, |
|
"loss": 0.1418, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"eval_cer": 7.5907145985647375, |
|
"eval_loss": 0.1861771196126938, |
|
"eval_runtime": 2079.9347, |
|
"eval_samples_per_second": 0.907, |
|
"eval_steps_per_second": 0.113, |
|
"eval_wer": 22.908018352632727, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 4.553164556962026e-06, |
|
"loss": 0.1196, |
|
"step": 22025 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 4.546835443037975e-06, |
|
"loss": 0.141, |
|
"step": 22050 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 4.540506329113925e-06, |
|
"loss": 0.1354, |
|
"step": 22075 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 4.534177215189874e-06, |
|
"loss": 0.1335, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 4.5278481012658235e-06, |
|
"loss": 0.1146, |
|
"step": 22125 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 4.521518987341773e-06, |
|
"loss": 0.1317, |
|
"step": 22150 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 4.5151898734177215e-06, |
|
"loss": 0.13, |
|
"step": 22175 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 4.508860759493671e-06, |
|
"loss": 0.131, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 4.50253164556962e-06, |
|
"loss": 0.1401, |
|
"step": 22225 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 4.49620253164557e-06, |
|
"loss": 0.1273, |
|
"step": 22250 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 4.489873417721519e-06, |
|
"loss": 0.1225, |
|
"step": 22275 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 4.483544303797469e-06, |
|
"loss": 0.1245, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 4.477215189873418e-06, |
|
"loss": 0.1344, |
|
"step": 22325 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 4.4708860759493675e-06, |
|
"loss": 0.1165, |
|
"step": 22350 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 4.464556962025317e-06, |
|
"loss": 0.1465, |
|
"step": 22375 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 4.458227848101266e-06, |
|
"loss": 0.1499, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 4.451898734177216e-06, |
|
"loss": 0.1279, |
|
"step": 22425 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 4.445569620253165e-06, |
|
"loss": 0.1466, |
|
"step": 22450 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.439240506329115e-06, |
|
"loss": 0.1458, |
|
"step": 22475 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.432911392405064e-06, |
|
"loss": 0.1283, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.4265822784810135e-06, |
|
"loss": 0.1324, |
|
"step": 22525 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.420253164556962e-06, |
|
"loss": 0.1148, |
|
"step": 22550 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.4139240506329115e-06, |
|
"loss": 0.1338, |
|
"step": 22575 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.407594936708861e-06, |
|
"loss": 0.1315, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.40126582278481e-06, |
|
"loss": 0.1467, |
|
"step": 22625 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.39493670886076e-06, |
|
"loss": 0.1394, |
|
"step": 22650 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.388607594936709e-06, |
|
"loss": 0.1519, |
|
"step": 22675 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.382278481012659e-06, |
|
"loss": 0.1314, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.375949367088608e-06, |
|
"loss": 0.1277, |
|
"step": 22725 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.3696202531645575e-06, |
|
"loss": 0.1381, |
|
"step": 22750 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.363291139240507e-06, |
|
"loss": 0.1319, |
|
"step": 22775 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 4.356962025316456e-06, |
|
"loss": 0.1386, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 4.350632911392406e-06, |
|
"loss": 0.1331, |
|
"step": 22825 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 4.344303797468355e-06, |
|
"loss": 0.1173, |
|
"step": 22850 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 4.337974683544305e-06, |
|
"loss": 0.1412, |
|
"step": 22875 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 4.331645569620254e-06, |
|
"loss": 0.145, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 4.325316455696203e-06, |
|
"loss": 0.1313, |
|
"step": 22925 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 4.318987341772152e-06, |
|
"loss": 0.1274, |
|
"step": 22950 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 4.3126582278481015e-06, |
|
"loss": 0.1444, |
|
"step": 22975 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 4.306329113924051e-06, |
|
"loss": 0.1432, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"eval_cer": 7.428994979953505, |
|
"eval_loss": 0.1846739947795868, |
|
"eval_runtime": 2064.5206, |
|
"eval_samples_per_second": 0.914, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 22.711383001966354, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 4.3e-06, |
|
"loss": 0.1406, |
|
"step": 23025 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 4.29367088607595e-06, |
|
"loss": 0.117, |
|
"step": 23050 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 4.287341772151899e-06, |
|
"loss": 0.1358, |
|
"step": 23075 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 4.281012658227849e-06, |
|
"loss": 0.1065, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 4.274683544303797e-06, |
|
"loss": 0.0863, |
|
"step": 23125 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 4.268354430379747e-06, |
|
"loss": 0.0824, |
|
"step": 23150 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 4.262025316455696e-06, |
|
"loss": 0.0825, |
|
"step": 23175 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 4.255696202531646e-06, |
|
"loss": 0.07, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 4.249367088607596e-06, |
|
"loss": 0.0733, |
|
"step": 23225 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 4.243037974683545e-06, |
|
"loss": 0.0754, |
|
"step": 23250 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 4.236708860759494e-06, |
|
"loss": 0.0877, |
|
"step": 23275 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 4.230379746835443e-06, |
|
"loss": 0.0876, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 4.224050632911393e-06, |
|
"loss": 0.0749, |
|
"step": 23325 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 4.217721518987342e-06, |
|
"loss": 0.0876, |
|
"step": 23350 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 4.2113924050632915e-06, |
|
"loss": 0.0832, |
|
"step": 23375 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 4.205063291139241e-06, |
|
"loss": 0.0825, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 4.19873417721519e-06, |
|
"loss": 0.0809, |
|
"step": 23425 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 4.19240506329114e-06, |
|
"loss": 0.0779, |
|
"step": 23450 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 4.186075949367089e-06, |
|
"loss": 0.0828, |
|
"step": 23475 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 4.179746835443038e-06, |
|
"loss": 0.0841, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 4.173417721518987e-06, |
|
"loss": 0.0664, |
|
"step": 23525 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 4.167088607594937e-06, |
|
"loss": 0.0735, |
|
"step": 23550 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 4.160759493670886e-06, |
|
"loss": 0.0938, |
|
"step": 23575 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 4.154430379746836e-06, |
|
"loss": 0.0793, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.148101265822786e-06, |
|
"loss": 0.0628, |
|
"step": 23625 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.141772151898734e-06, |
|
"loss": 0.0828, |
|
"step": 23650 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.135443037974684e-06, |
|
"loss": 0.079, |
|
"step": 23675 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.129113924050633e-06, |
|
"loss": 0.0868, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 4.122784810126583e-06, |
|
"loss": 0.0844, |
|
"step": 23725 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 4.116455696202532e-06, |
|
"loss": 0.0771, |
|
"step": 23750 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 4.1101265822784815e-06, |
|
"loss": 0.086, |
|
"step": 23775 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 4.103797468354431e-06, |
|
"loss": 0.0825, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 4.09746835443038e-06, |
|
"loss": 0.0828, |
|
"step": 23825 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 4.09113924050633e-06, |
|
"loss": 0.0877, |
|
"step": 23850 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 4.084810126582278e-06, |
|
"loss": 0.0883, |
|
"step": 23875 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 4.078481012658228e-06, |
|
"loss": 0.0872, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 4.072151898734177e-06, |
|
"loss": 0.0838, |
|
"step": 23925 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 4.065822784810127e-06, |
|
"loss": 0.0795, |
|
"step": 23950 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.059493670886076e-06, |
|
"loss": 0.0764, |
|
"step": 23975 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.0531645569620255e-06, |
|
"loss": 0.0965, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"eval_cer": 7.039857147670227, |
|
"eval_loss": 0.1930975764989853, |
|
"eval_runtime": 2062.1826, |
|
"eval_samples_per_second": 0.915, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.73913043478261, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.046835443037975e-06, |
|
"loss": 0.0773, |
|
"step": 24025 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.040506329113924e-06, |
|
"loss": 0.0816, |
|
"step": 24050 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.034177215189874e-06, |
|
"loss": 0.0786, |
|
"step": 24075 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.027848101265823e-06, |
|
"loss": 0.0829, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.021518987341773e-06, |
|
"loss": 0.0789, |
|
"step": 24125 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.015189873417722e-06, |
|
"loss": 0.0846, |
|
"step": 24150 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.0088607594936715e-06, |
|
"loss": 0.0747, |
|
"step": 24175 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 4.002531645569621e-06, |
|
"loss": 0.0796, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 3.99620253164557e-06, |
|
"loss": 0.0746, |
|
"step": 24225 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 3.989873417721519e-06, |
|
"loss": 0.0763, |
|
"step": 24250 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 3.983544303797468e-06, |
|
"loss": 0.0808, |
|
"step": 24275 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 3.977215189873418e-06, |
|
"loss": 0.0878, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 3.970886075949367e-06, |
|
"loss": 0.0775, |
|
"step": 24325 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 3.964556962025317e-06, |
|
"loss": 0.0907, |
|
"step": 24350 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 3.958227848101266e-06, |
|
"loss": 0.0768, |
|
"step": 24375 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 3.9518987341772155e-06, |
|
"loss": 0.0798, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 3.945569620253165e-06, |
|
"loss": 0.0934, |
|
"step": 24425 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 3.939240506329114e-06, |
|
"loss": 0.0816, |
|
"step": 24450 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 3.932911392405064e-06, |
|
"loss": 0.0844, |
|
"step": 24475 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 3.926582278481013e-06, |
|
"loss": 0.0793, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 3.920253164556963e-06, |
|
"loss": 0.0788, |
|
"step": 24525 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 3.913924050632912e-06, |
|
"loss": 0.0876, |
|
"step": 24550 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 3.9075949367088615e-06, |
|
"loss": 0.0906, |
|
"step": 24575 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 3.901265822784811e-06, |
|
"loss": 0.0837, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 3.8949367088607595e-06, |
|
"loss": 0.0786, |
|
"step": 24625 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 3.888607594936709e-06, |
|
"loss": 0.0858, |
|
"step": 24650 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 3.882278481012658e-06, |
|
"loss": 0.0727, |
|
"step": 24675 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 3.875949367088608e-06, |
|
"loss": 0.0795, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 3.869620253164557e-06, |
|
"loss": 0.0734, |
|
"step": 24725 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 3.863291139240507e-06, |
|
"loss": 0.0773, |
|
"step": 24750 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 3.856962025316456e-06, |
|
"loss": 0.085, |
|
"step": 24775 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 3.8506329113924055e-06, |
|
"loss": 0.0754, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 3.844303797468355e-06, |
|
"loss": 0.0876, |
|
"step": 24825 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 3.837974683544304e-06, |
|
"loss": 0.0842, |
|
"step": 24850 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 3.831645569620254e-06, |
|
"loss": 0.0832, |
|
"step": 24875 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 3.825316455696203e-06, |
|
"loss": 0.082, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 3.818987341772153e-06, |
|
"loss": 0.0717, |
|
"step": 24925 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 3.8126582278481016e-06, |
|
"loss": 0.0803, |
|
"step": 24950 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 3.806329113924051e-06, |
|
"loss": 0.0816, |
|
"step": 24975 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 3.8000000000000005e-06, |
|
"loss": 0.0723, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"eval_cer": 7.269802230383074, |
|
"eval_loss": 0.19612479209899902, |
|
"eval_runtime": 2078.9413, |
|
"eval_samples_per_second": 0.907, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 22.323574393707666, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 3.79367088607595e-06, |
|
"loss": 0.0706, |
|
"step": 25025 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 3.787341772151899e-06, |
|
"loss": 0.0713, |
|
"step": 25050 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 3.7810126582278483e-06, |
|
"loss": 0.0839, |
|
"step": 25075 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 3.7746835443037978e-06, |
|
"loss": 0.076, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 3.768354430379747e-06, |
|
"loss": 0.0863, |
|
"step": 25125 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 3.762025316455696e-06, |
|
"loss": 0.0725, |
|
"step": 25150 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 3.7556962025316456e-06, |
|
"loss": 0.0807, |
|
"step": 25175 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 3.749367088607595e-06, |
|
"loss": 0.087, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 3.7430379746835445e-06, |
|
"loss": 0.075, |
|
"step": 25225 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 3.7367088607594943e-06, |
|
"loss": 0.0816, |
|
"step": 25250 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 3.7303797468354437e-06, |
|
"loss": 0.0706, |
|
"step": 25275 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 3.7240506329113927e-06, |
|
"loss": 0.0807, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 3.717721518987342e-06, |
|
"loss": 0.0853, |
|
"step": 25325 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.7113924050632916e-06, |
|
"loss": 0.0785, |
|
"step": 25350 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.705063291139241e-06, |
|
"loss": 0.0725, |
|
"step": 25375 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.69873417721519e-06, |
|
"loss": 0.0834, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.6924050632911395e-06, |
|
"loss": 0.082, |
|
"step": 25425 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.686075949367089e-06, |
|
"loss": 0.0753, |
|
"step": 25450 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 3.6797468354430383e-06, |
|
"loss": 0.086, |
|
"step": 25475 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 3.6734177215189877e-06, |
|
"loss": 0.0785, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 3.6670886075949368e-06, |
|
"loss": 0.0809, |
|
"step": 25525 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 3.660759493670886e-06, |
|
"loss": 0.0837, |
|
"step": 25550 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 3.6544303797468356e-06, |
|
"loss": 0.0887, |
|
"step": 25575 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 3.648101265822785e-06, |
|
"loss": 0.0776, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 3.641772151898734e-06, |
|
"loss": 0.0828, |
|
"step": 25625 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 3.6354430379746835e-06, |
|
"loss": 0.0762, |
|
"step": 25650 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 3.6291139240506333e-06, |
|
"loss": 0.0833, |
|
"step": 25675 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 3.6227848101265827e-06, |
|
"loss": 0.0818, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 3.616455696202532e-06, |
|
"loss": 0.0835, |
|
"step": 25725 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 3.6101265822784816e-06, |
|
"loss": 0.0739, |
|
"step": 25750 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 3.6037974683544306e-06, |
|
"loss": 0.085, |
|
"step": 25775 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 3.59746835443038e-06, |
|
"loss": 0.0778, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 3.5911392405063295e-06, |
|
"loss": 0.0832, |
|
"step": 25825 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 3.584810126582279e-06, |
|
"loss": 0.0744, |
|
"step": 25850 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 3.5784810126582283e-06, |
|
"loss": 0.0816, |
|
"step": 25875 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 3.5721518987341773e-06, |
|
"loss": 0.0776, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 3.5658227848101267e-06, |
|
"loss": 0.0826, |
|
"step": 25925 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 3.559493670886076e-06, |
|
"loss": 0.0807, |
|
"step": 25950 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 3.5531645569620256e-06, |
|
"loss": 0.0769, |
|
"step": 25975 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 3.5468354430379746e-06, |
|
"loss": 0.0773, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"eval_cer": 7.0752333142414345, |
|
"eval_loss": 0.19767002761363983, |
|
"eval_runtime": 2055.5369, |
|
"eval_samples_per_second": 0.918, |
|
"eval_steps_per_second": 0.115, |
|
"eval_wer": 22.05046974000437, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 3.540506329113924e-06, |
|
"loss": 0.0847, |
|
"step": 26025 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 3.5341772151898735e-06, |
|
"loss": 0.0767, |
|
"step": 26050 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 3.5278481012658233e-06, |
|
"loss": 0.0968, |
|
"step": 26075 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 3.5215189873417727e-06, |
|
"loss": 0.0934, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 3.515189873417722e-06, |
|
"loss": 0.0925, |
|
"step": 26125 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 3.508860759493671e-06, |
|
"loss": 0.0779, |
|
"step": 26150 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 3.5025316455696206e-06, |
|
"loss": 0.093, |
|
"step": 26175 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 3.49620253164557e-06, |
|
"loss": 0.0807, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 3.4898734177215195e-06, |
|
"loss": 0.0822, |
|
"step": 26225 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 3.4835443037974685e-06, |
|
"loss": 0.0871, |
|
"step": 26250 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 3.477215189873418e-06, |
|
"loss": 0.0828, |
|
"step": 26275 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 3.4708860759493673e-06, |
|
"loss": 0.0769, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 3.4645569620253167e-06, |
|
"loss": 0.0897, |
|
"step": 26325 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 3.458227848101266e-06, |
|
"loss": 0.0863, |
|
"step": 26350 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 3.451898734177215e-06, |
|
"loss": 0.0818, |
|
"step": 26375 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 3.4455696202531646e-06, |
|
"loss": 0.0874, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 3.439240506329114e-06, |
|
"loss": 0.0808, |
|
"step": 26425 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 3.4329113924050635e-06, |
|
"loss": 0.0829, |
|
"step": 26450 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 3.4265822784810133e-06, |
|
"loss": 0.078, |
|
"step": 26475 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 3.4202531645569627e-06, |
|
"loss": 0.0805, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 3.4139240506329117e-06, |
|
"loss": 0.0836, |
|
"step": 26525 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 3.407594936708861e-06, |
|
"loss": 0.0817, |
|
"step": 26550 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 3.4012658227848106e-06, |
|
"loss": 0.091, |
|
"step": 26575 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 3.39493670886076e-06, |
|
"loss": 0.0753, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 3.388607594936709e-06, |
|
"loss": 0.0803, |
|
"step": 26625 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 3.3822784810126585e-06, |
|
"loss": 0.0753, |
|
"step": 26650 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 3.375949367088608e-06, |
|
"loss": 0.0946, |
|
"step": 26675 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 3.3696202531645573e-06, |
|
"loss": 0.0903, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 3.3632911392405067e-06, |
|
"loss": 0.0783, |
|
"step": 26725 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 3.3569620253164557e-06, |
|
"loss": 0.091, |
|
"step": 26750 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 3.350632911392405e-06, |
|
"loss": 0.0815, |
|
"step": 26775 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 3.3443037974683546e-06, |
|
"loss": 0.0758, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 3.337974683544304e-06, |
|
"loss": 0.0856, |
|
"step": 26825 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 3.331645569620253e-06, |
|
"loss": 0.0789, |
|
"step": 26850 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 3.3253164556962025e-06, |
|
"loss": 0.0798, |
|
"step": 26875 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 3.3189873417721523e-06, |
|
"loss": 0.0838, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 3.3126582278481017e-06, |
|
"loss": 0.0784, |
|
"step": 26925 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 3.306329113924051e-06, |
|
"loss": 0.085, |
|
"step": 26950 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 3.3000000000000006e-06, |
|
"loss": 0.0749, |
|
"step": 26975 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 3.2936708860759496e-06, |
|
"loss": 0.0862, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"eval_cer": 7.081971631683569, |
|
"eval_loss": 0.19593346118927002, |
|
"eval_runtime": 2062.8535, |
|
"eval_samples_per_second": 0.914, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.952152064671182, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 3.287341772151899e-06, |
|
"loss": 0.08, |
|
"step": 27025 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 3.2810126582278485e-06, |
|
"loss": 0.0793, |
|
"step": 27050 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 3.274683544303798e-06, |
|
"loss": 0.0771, |
|
"step": 27075 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 3.268354430379747e-06, |
|
"loss": 0.0839, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 3.2620253164556963e-06, |
|
"loss": 0.0718, |
|
"step": 27125 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 3.2556962025316457e-06, |
|
"loss": 0.083, |
|
"step": 27150 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 3.249367088607595e-06, |
|
"loss": 0.0802, |
|
"step": 27175 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 3.2430379746835446e-06, |
|
"loss": 0.0703, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 3.2367088607594936e-06, |
|
"loss": 0.0852, |
|
"step": 27225 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 3.230379746835443e-06, |
|
"loss": 0.0816, |
|
"step": 27250 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 3.2240506329113925e-06, |
|
"loss": 0.0866, |
|
"step": 27275 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 3.2177215189873423e-06, |
|
"loss": 0.0846, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 3.2113924050632917e-06, |
|
"loss": 0.0739, |
|
"step": 27325 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 3.205063291139241e-06, |
|
"loss": 0.0741, |
|
"step": 27350 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 3.19873417721519e-06, |
|
"loss": 0.0853, |
|
"step": 27375 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 3.1924050632911396e-06, |
|
"loss": 0.0922, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 3.186075949367089e-06, |
|
"loss": 0.0894, |
|
"step": 27425 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 3.1797468354430384e-06, |
|
"loss": 0.0848, |
|
"step": 27450 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 3.1734177215189875e-06, |
|
"loss": 0.0749, |
|
"step": 27475 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 3.167088607594937e-06, |
|
"loss": 0.0847, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 3.1607594936708863e-06, |
|
"loss": 0.0759, |
|
"step": 27525 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 3.1544303797468357e-06, |
|
"loss": 0.0752, |
|
"step": 27550 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 3.148101265822785e-06, |
|
"loss": 0.0763, |
|
"step": 27575 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 3.141772151898734e-06, |
|
"loss": 0.0865, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 3.1354430379746836e-06, |
|
"loss": 0.0788, |
|
"step": 27625 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 3.129113924050633e-06, |
|
"loss": 0.0791, |
|
"step": 27650 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 3.1227848101265824e-06, |
|
"loss": 0.0814, |
|
"step": 27675 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 3.1164556962025315e-06, |
|
"loss": 0.0772, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 3.1101265822784817e-06, |
|
"loss": 0.085, |
|
"step": 27725 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 3.1037974683544307e-06, |
|
"loss": 0.0805, |
|
"step": 27750 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 3.09746835443038e-06, |
|
"loss": 0.0814, |
|
"step": 27775 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 3.0911392405063296e-06, |
|
"loss": 0.0711, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 3.084810126582279e-06, |
|
"loss": 0.081, |
|
"step": 27825 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 3.078481012658228e-06, |
|
"loss": 0.0774, |
|
"step": 27850 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 3.0721518987341774e-06, |
|
"loss": 0.0752, |
|
"step": 27875 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 3.065822784810127e-06, |
|
"loss": 0.076, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 3.0594936708860763e-06, |
|
"loss": 0.0712, |
|
"step": 27925 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 3.0531645569620253e-06, |
|
"loss": 0.0832, |
|
"step": 27950 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 3.0468354430379747e-06, |
|
"loss": 0.0865, |
|
"step": 27975 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 3.040506329113924e-06, |
|
"loss": 0.0739, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"eval_cer": 7.149354806104916, |
|
"eval_loss": 0.19822219014167786, |
|
"eval_runtime": 2065.3987, |
|
"eval_samples_per_second": 0.913, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.771902993227005, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 3.0341772151898736e-06, |
|
"loss": 0.0749, |
|
"step": 28025 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 3.027848101265823e-06, |
|
"loss": 0.0751, |
|
"step": 28050 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 3.021518987341772e-06, |
|
"loss": 0.0881, |
|
"step": 28075 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 3.0151898734177214e-06, |
|
"loss": 0.0829, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 3.0088607594936713e-06, |
|
"loss": 0.0776, |
|
"step": 28125 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 3.0025316455696207e-06, |
|
"loss": 0.0793, |
|
"step": 28150 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 2.99620253164557e-06, |
|
"loss": 0.086, |
|
"step": 28175 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 2.9898734177215196e-06, |
|
"loss": 0.0888, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 2.9835443037974686e-06, |
|
"loss": 0.0753, |
|
"step": 28225 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 2.977215189873418e-06, |
|
"loss": 0.0964, |
|
"step": 28250 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 2.9708860759493674e-06, |
|
"loss": 0.0856, |
|
"step": 28275 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 2.964556962025317e-06, |
|
"loss": 0.0793, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 2.958227848101266e-06, |
|
"loss": 0.0807, |
|
"step": 28325 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 2.9518987341772153e-06, |
|
"loss": 0.0895, |
|
"step": 28350 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 2.9455696202531647e-06, |
|
"loss": 0.079, |
|
"step": 28375 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 2.939240506329114e-06, |
|
"loss": 0.0797, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 2.9329113924050636e-06, |
|
"loss": 0.0836, |
|
"step": 28425 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 2.9265822784810126e-06, |
|
"loss": 0.0794, |
|
"step": 28450 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 2.920253164556962e-06, |
|
"loss": 0.0826, |
|
"step": 28475 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 2.9139240506329114e-06, |
|
"loss": 0.0822, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 2.907594936708861e-06, |
|
"loss": 0.0888, |
|
"step": 28525 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 2.9012658227848107e-06, |
|
"loss": 0.0752, |
|
"step": 28550 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 2.89493670886076e-06, |
|
"loss": 0.0776, |
|
"step": 28575 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 2.888607594936709e-06, |
|
"loss": 0.0745, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 2.8822784810126586e-06, |
|
"loss": 0.0867, |
|
"step": 28625 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 2.875949367088608e-06, |
|
"loss": 0.0782, |
|
"step": 28650 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 2.8696202531645574e-06, |
|
"loss": 0.0799, |
|
"step": 28675 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 2.8632911392405064e-06, |
|
"loss": 0.08, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 2.856962025316456e-06, |
|
"loss": 0.0808, |
|
"step": 28725 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 2.8506329113924053e-06, |
|
"loss": 0.0866, |
|
"step": 28750 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 2.8443037974683547e-06, |
|
"loss": 0.078, |
|
"step": 28775 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 2.837974683544304e-06, |
|
"loss": 0.0774, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 2.831645569620253e-06, |
|
"loss": 0.0793, |
|
"step": 28825 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 2.8253164556962026e-06, |
|
"loss": 0.0881, |
|
"step": 28850 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 2.818987341772152e-06, |
|
"loss": 0.0668, |
|
"step": 28875 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 2.8126582278481014e-06, |
|
"loss": 0.0688, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 2.8063291139240504e-06, |
|
"loss": 0.0816, |
|
"step": 28925 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 2.8000000000000003e-06, |
|
"loss": 0.0734, |
|
"step": 28950 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 2.7936708860759497e-06, |
|
"loss": 0.0838, |
|
"step": 28975 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 2.787341772151899e-06, |
|
"loss": 0.0843, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"eval_cer": 7.124086115696911, |
|
"eval_loss": 0.19632002711296082, |
|
"eval_runtime": 2068.5466, |
|
"eval_samples_per_second": 0.912, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.89206904085646, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 2.7810126582278486e-06, |
|
"loss": 0.0813, |
|
"step": 29025 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 2.774683544303798e-06, |
|
"loss": 0.0769, |
|
"step": 29050 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 2.768354430379747e-06, |
|
"loss": 0.0806, |
|
"step": 29075 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 2.7620253164556964e-06, |
|
"loss": 0.0804, |
|
"step": 29100 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 2.755696202531646e-06, |
|
"loss": 0.0771, |
|
"step": 29125 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 2.7493670886075953e-06, |
|
"loss": 0.0882, |
|
"step": 29150 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 2.7430379746835443e-06, |
|
"loss": 0.0849, |
|
"step": 29175 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 2.7367088607594937e-06, |
|
"loss": 0.0764, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 2.730379746835443e-06, |
|
"loss": 0.0694, |
|
"step": 29225 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 2.7243037974683546e-06, |
|
"loss": 0.092, |
|
"step": 29250 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 2.717974683544304e-06, |
|
"loss": 0.0706, |
|
"step": 29275 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 2.711645569620253e-06, |
|
"loss": 0.0801, |
|
"step": 29300 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 2.7053164556962025e-06, |
|
"loss": 0.0895, |
|
"step": 29325 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 2.6989873417721523e-06, |
|
"loss": 0.0696, |
|
"step": 29350 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 2.6926582278481017e-06, |
|
"loss": 0.0794, |
|
"step": 29375 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 2.686329113924051e-06, |
|
"loss": 0.0716, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 2.68e-06, |
|
"loss": 0.0936, |
|
"step": 29425 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 2.6736708860759496e-06, |
|
"loss": 0.0785, |
|
"step": 29450 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 2.667341772151899e-06, |
|
"loss": 0.0733, |
|
"step": 29475 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 2.6612658227848105e-06, |
|
"loss": 0.0781, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 2.65493670886076e-06, |
|
"loss": 0.0807, |
|
"step": 29525 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 2.648607594936709e-06, |
|
"loss": 0.0724, |
|
"step": 29550 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 2.6422784810126583e-06, |
|
"loss": 0.0747, |
|
"step": 29575 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 2.6359493670886078e-06, |
|
"loss": 0.0922, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.629620253164557e-06, |
|
"loss": 0.0766, |
|
"step": 29625 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.623291139240506e-06, |
|
"loss": 0.0867, |
|
"step": 29650 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.6169620253164556e-06, |
|
"loss": 0.0786, |
|
"step": 29675 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.610632911392405e-06, |
|
"loss": 0.077, |
|
"step": 29700 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.6043037974683545e-06, |
|
"loss": 0.0832, |
|
"step": 29725 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.5979746835443043e-06, |
|
"loss": 0.0831, |
|
"step": 29750 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.5916455696202538e-06, |
|
"loss": 0.0812, |
|
"step": 29775 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.5853164556962028e-06, |
|
"loss": 0.0688, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.578987341772152e-06, |
|
"loss": 0.0824, |
|
"step": 29825 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.5726582278481016e-06, |
|
"loss": 0.0698, |
|
"step": 29850 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.566329113924051e-06, |
|
"loss": 0.076, |
|
"step": 29875 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.56e-06, |
|
"loss": 0.0956, |
|
"step": 29900 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.5536708860759495e-06, |
|
"loss": 0.0773, |
|
"step": 29925 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.547341772151899e-06, |
|
"loss": 0.0725, |
|
"step": 29950 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.5410126582278483e-06, |
|
"loss": 0.0812, |
|
"step": 29975 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.5346835443037978e-06, |
|
"loss": 0.0734, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"eval_cer": 7.131666722819312, |
|
"eval_loss": 0.1979527324438095, |
|
"eval_runtime": 2065.7489, |
|
"eval_samples_per_second": 0.913, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.7882892724492, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.5283544303797468e-06, |
|
"loss": 0.0791, |
|
"step": 30025 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.522025316455696e-06, |
|
"loss": 0.0629, |
|
"step": 30050 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.5156962025316456e-06, |
|
"loss": 0.0833, |
|
"step": 30075 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.509367088607595e-06, |
|
"loss": 0.0725, |
|
"step": 30100 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.503037974683544e-06, |
|
"loss": 0.0819, |
|
"step": 30125 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.496708860759494e-06, |
|
"loss": 0.0834, |
|
"step": 30150 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.4903797468354433e-06, |
|
"loss": 0.0827, |
|
"step": 30175 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.4840506329113923e-06, |
|
"loss": 0.0722, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.4777215189873418e-06, |
|
"loss": 0.0781, |
|
"step": 30225 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.4713924050632916e-06, |
|
"loss": 0.0712, |
|
"step": 30250 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.4650632911392406e-06, |
|
"loss": 0.0691, |
|
"step": 30275 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.45873417721519e-06, |
|
"loss": 0.0803, |
|
"step": 30300 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 2.4524050632911395e-06, |
|
"loss": 0.0785, |
|
"step": 30325 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 2.446075949367089e-06, |
|
"loss": 0.0843, |
|
"step": 30350 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 2.4397468354430383e-06, |
|
"loss": 0.0818, |
|
"step": 30375 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 2.4334177215189873e-06, |
|
"loss": 0.0713, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.4270886075949368e-06, |
|
"loss": 0.0719, |
|
"step": 30425 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.420759493670886e-06, |
|
"loss": 0.0828, |
|
"step": 30450 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.4144303797468356e-06, |
|
"loss": 0.0767, |
|
"step": 30475 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.408101265822785e-06, |
|
"loss": 0.0776, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.4017721518987345e-06, |
|
"loss": 0.084, |
|
"step": 30525 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.395443037974684e-06, |
|
"loss": 0.0824, |
|
"step": 30550 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.389113924050633e-06, |
|
"loss": 0.0698, |
|
"step": 30575 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.3827848101265823e-06, |
|
"loss": 0.0858, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.3764556962025318e-06, |
|
"loss": 0.0739, |
|
"step": 30625 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.370126582278481e-06, |
|
"loss": 0.0719, |
|
"step": 30650 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.3637974683544306e-06, |
|
"loss": 0.0799, |
|
"step": 30675 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.35746835443038e-06, |
|
"loss": 0.0737, |
|
"step": 30700 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.3511392405063295e-06, |
|
"loss": 0.0745, |
|
"step": 30725 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.344810126582279e-06, |
|
"loss": 0.0698, |
|
"step": 30750 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.338481012658228e-06, |
|
"loss": 0.0872, |
|
"step": 30775 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.3321518987341773e-06, |
|
"loss": 0.0693, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.3258227848101268e-06, |
|
"loss": 0.0702, |
|
"step": 30825 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.319493670886076e-06, |
|
"loss": 0.0874, |
|
"step": 30850 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.3131645569620256e-06, |
|
"loss": 0.0774, |
|
"step": 30875 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.306835443037975e-06, |
|
"loss": 0.0748, |
|
"step": 30900 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.3005063291139245e-06, |
|
"loss": 0.0744, |
|
"step": 30925 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.2941772151898735e-06, |
|
"loss": 0.0762, |
|
"step": 30950 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.287848101265823e-06, |
|
"loss": 0.0795, |
|
"step": 30975 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.2815189873417723e-06, |
|
"loss": 0.0785, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"eval_cer": 7.194838448839325, |
|
"eval_loss": 0.1955455094575882, |
|
"eval_runtime": 2069.5419, |
|
"eval_samples_per_second": 0.911, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.87568276163426, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.2751898734177218e-06, |
|
"loss": 0.0788, |
|
"step": 31025 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.2688607594936708e-06, |
|
"loss": 0.0823, |
|
"step": 31050 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.2625316455696206e-06, |
|
"loss": 0.0756, |
|
"step": 31075 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.25620253164557e-06, |
|
"loss": 0.083, |
|
"step": 31100 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.249873417721519e-06, |
|
"loss": 0.0787, |
|
"step": 31125 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.2435443037974685e-06, |
|
"loss": 0.0737, |
|
"step": 31150 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.237215189873418e-06, |
|
"loss": 0.0781, |
|
"step": 31175 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.2308860759493673e-06, |
|
"loss": 0.0782, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.2245569620253168e-06, |
|
"loss": 0.0827, |
|
"step": 31225 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.2182278481012658e-06, |
|
"loss": 0.0711, |
|
"step": 31250 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.211898734177215e-06, |
|
"loss": 0.0882, |
|
"step": 31275 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.205569620253165e-06, |
|
"loss": 0.079, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.199240506329114e-06, |
|
"loss": 0.0728, |
|
"step": 31325 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 2.1929113924050635e-06, |
|
"loss": 0.0722, |
|
"step": 31350 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 2.186582278481013e-06, |
|
"loss": 0.075, |
|
"step": 31375 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 2.1802531645569623e-06, |
|
"loss": 0.0877, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 2.1739240506329113e-06, |
|
"loss": 0.0756, |
|
"step": 31425 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 2.1675949367088608e-06, |
|
"loss": 0.082, |
|
"step": 31450 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 2.16126582278481e-06, |
|
"loss": 0.0786, |
|
"step": 31475 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 2.1549367088607596e-06, |
|
"loss": 0.0689, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 2.148607594936709e-06, |
|
"loss": 0.0693, |
|
"step": 31525 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 2.1422784810126585e-06, |
|
"loss": 0.0745, |
|
"step": 31550 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 2.135949367088608e-06, |
|
"loss": 0.072, |
|
"step": 31575 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 2.1296202531645573e-06, |
|
"loss": 0.0822, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 2.1232911392405063e-06, |
|
"loss": 0.0783, |
|
"step": 31625 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 2.1169620253164558e-06, |
|
"loss": 0.0755, |
|
"step": 31650 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 2.110632911392405e-06, |
|
"loss": 0.0726, |
|
"step": 31675 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 2.1043037974683546e-06, |
|
"loss": 0.0863, |
|
"step": 31700 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 2.097974683544304e-06, |
|
"loss": 0.0737, |
|
"step": 31725 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 2.0916455696202535e-06, |
|
"loss": 0.0803, |
|
"step": 31750 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 2.085316455696203e-06, |
|
"loss": 0.0716, |
|
"step": 31775 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 2.078987341772152e-06, |
|
"loss": 0.083, |
|
"step": 31800 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 2.0729113924050633e-06, |
|
"loss": 0.0747, |
|
"step": 31825 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 2.0665822784810128e-06, |
|
"loss": 0.0808, |
|
"step": 31850 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 2.060253164556962e-06, |
|
"loss": 0.0715, |
|
"step": 31875 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 2.0539240506329116e-06, |
|
"loss": 0.077, |
|
"step": 31900 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 2.047594936708861e-06, |
|
"loss": 0.0736, |
|
"step": 31925 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 2.0412658227848105e-06, |
|
"loss": 0.0794, |
|
"step": 31950 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 2.0349367088607595e-06, |
|
"loss": 0.0736, |
|
"step": 31975 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 2.028607594936709e-06, |
|
"loss": 0.0691, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"eval_cer": 7.0937636872073035, |
|
"eval_loss": 0.19776789844036102, |
|
"eval_runtime": 2069.1254, |
|
"eval_samples_per_second": 0.911, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.744592527856675, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 2.0222784810126583e-06, |
|
"loss": 0.0739, |
|
"step": 32025 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 2.0159493670886078e-06, |
|
"loss": 0.0852, |
|
"step": 32050 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 2.009620253164557e-06, |
|
"loss": 0.0732, |
|
"step": 32075 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 2.0032911392405066e-06, |
|
"loss": 0.0835, |
|
"step": 32100 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.996962025316456e-06, |
|
"loss": 0.068, |
|
"step": 32125 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.9906329113924055e-06, |
|
"loss": 0.0812, |
|
"step": 32150 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.9843037974683545e-06, |
|
"loss": 0.08, |
|
"step": 32175 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.977974683544304e-06, |
|
"loss": 0.0717, |
|
"step": 32200 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.9716455696202533e-06, |
|
"loss": 0.0655, |
|
"step": 32225 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.9653164556962028e-06, |
|
"loss": 0.0794, |
|
"step": 32250 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.9589873417721518e-06, |
|
"loss": 0.086, |
|
"step": 32275 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.9526582278481016e-06, |
|
"loss": 0.0827, |
|
"step": 32300 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.946329113924051e-06, |
|
"loss": 0.0782, |
|
"step": 32325 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.94e-06, |
|
"loss": 0.0779, |
|
"step": 32350 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.9336708860759495e-06, |
|
"loss": 0.0729, |
|
"step": 32375 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.927341772151899e-06, |
|
"loss": 0.0857, |
|
"step": 32400 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.9210126582278483e-06, |
|
"loss": 0.0739, |
|
"step": 32425 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.9146835443037973e-06, |
|
"loss": 0.0697, |
|
"step": 32450 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.9083544303797468e-06, |
|
"loss": 0.0762, |
|
"step": 32475 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.9020253164556964e-06, |
|
"loss": 0.0628, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.8956962025316458e-06, |
|
"loss": 0.0745, |
|
"step": 32525 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.889367088607595e-06, |
|
"loss": 0.0628, |
|
"step": 32550 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.8830379746835445e-06, |
|
"loss": 0.0714, |
|
"step": 32575 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.876708860759494e-06, |
|
"loss": 0.0763, |
|
"step": 32600 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.8703797468354431e-06, |
|
"loss": 0.0797, |
|
"step": 32625 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.8640506329113926e-06, |
|
"loss": 0.0776, |
|
"step": 32650 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.8577215189873418e-06, |
|
"loss": 0.0796, |
|
"step": 32675 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.8513924050632912e-06, |
|
"loss": 0.0758, |
|
"step": 32700 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.8450632911392408e-06, |
|
"loss": 0.0699, |
|
"step": 32725 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.83873417721519e-06, |
|
"loss": 0.0813, |
|
"step": 32750 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.8324050632911395e-06, |
|
"loss": 0.0822, |
|
"step": 32775 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.8260759493670887e-06, |
|
"loss": 0.0684, |
|
"step": 32800 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.8197468354430381e-06, |
|
"loss": 0.0943, |
|
"step": 32825 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.8134177215189873e-06, |
|
"loss": 0.0747, |
|
"step": 32850 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.8070886075949368e-06, |
|
"loss": 0.0831, |
|
"step": 32875 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.8007594936708862e-06, |
|
"loss": 0.0722, |
|
"step": 32900 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.7944303797468356e-06, |
|
"loss": 0.0819, |
|
"step": 32925 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.788101265822785e-06, |
|
"loss": 0.0784, |
|
"step": 32950 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.7817721518987343e-06, |
|
"loss": 0.0787, |
|
"step": 32975 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.7754430379746837e-06, |
|
"loss": 0.0834, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"eval_cer": 7.012061588221422, |
|
"eval_loss": 0.1952751874923706, |
|
"eval_runtime": 2064.5918, |
|
"eval_samples_per_second": 0.913, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.324011361153595, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.7691139240506331e-06, |
|
"loss": 0.079, |
|
"step": 33025 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.7627848101265823e-06, |
|
"loss": 0.0881, |
|
"step": 33050 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.7564556962025318e-06, |
|
"loss": 0.0669, |
|
"step": 33075 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.750126582278481e-06, |
|
"loss": 0.0803, |
|
"step": 33100 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.7437974683544306e-06, |
|
"loss": 0.0823, |
|
"step": 33125 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.73746835443038e-06, |
|
"loss": 0.068, |
|
"step": 33150 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.7311392405063293e-06, |
|
"loss": 0.0737, |
|
"step": 33175 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.7248101265822787e-06, |
|
"loss": 0.0885, |
|
"step": 33200 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.718481012658228e-06, |
|
"loss": 0.0747, |
|
"step": 33225 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.7121518987341773e-06, |
|
"loss": 0.0714, |
|
"step": 33250 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.7058227848101266e-06, |
|
"loss": 0.0638, |
|
"step": 33275 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.699493670886076e-06, |
|
"loss": 0.0744, |
|
"step": 33300 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.6931645569620256e-06, |
|
"loss": 0.0741, |
|
"step": 33325 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.6868354430379748e-06, |
|
"loss": 0.0693, |
|
"step": 33350 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.6805063291139243e-06, |
|
"loss": 0.0749, |
|
"step": 33375 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.6741772151898735e-06, |
|
"loss": 0.0753, |
|
"step": 33400 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.667848101265823e-06, |
|
"loss": 0.0799, |
|
"step": 33425 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.6615189873417723e-06, |
|
"loss": 0.0785, |
|
"step": 33450 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.6551898734177216e-06, |
|
"loss": 0.0802, |
|
"step": 33475 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.648860759493671e-06, |
|
"loss": 0.0833, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.6425316455696206e-06, |
|
"loss": 0.0807, |
|
"step": 33525 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.6362025316455698e-06, |
|
"loss": 0.0898, |
|
"step": 33550 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.6298734177215193e-06, |
|
"loss": 0.0692, |
|
"step": 33575 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.6235443037974685e-06, |
|
"loss": 0.0807, |
|
"step": 33600 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.617215189873418e-06, |
|
"loss": 0.0797, |
|
"step": 33625 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.6108860759493671e-06, |
|
"loss": 0.0649, |
|
"step": 33650 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.6045569620253166e-06, |
|
"loss": 0.0703, |
|
"step": 33675 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.5982278481012658e-06, |
|
"loss": 0.0743, |
|
"step": 33700 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.5918987341772152e-06, |
|
"loss": 0.0715, |
|
"step": 33725 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.5855696202531648e-06, |
|
"loss": 0.065, |
|
"step": 33750 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.579240506329114e-06, |
|
"loss": 0.0744, |
|
"step": 33775 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.5729113924050635e-06, |
|
"loss": 0.0765, |
|
"step": 33800 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.5665822784810127e-06, |
|
"loss": 0.0728, |
|
"step": 33825 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.5602531645569621e-06, |
|
"loss": 0.0728, |
|
"step": 33850 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.5539240506329115e-06, |
|
"loss": 0.0747, |
|
"step": 33875 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.5475949367088608e-06, |
|
"loss": 0.0834, |
|
"step": 33900 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.5412658227848102e-06, |
|
"loss": 0.0719, |
|
"step": 33925 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.5349367088607598e-06, |
|
"loss": 0.0762, |
|
"step": 33950 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.528607594936709e-06, |
|
"loss": 0.0888, |
|
"step": 33975 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.5222784810126585e-06, |
|
"loss": 0.0675, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"eval_cer": 7.076917893601968, |
|
"eval_loss": 0.19584281742572784, |
|
"eval_runtime": 2068.6884, |
|
"eval_samples_per_second": 0.912, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.771902993227005, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.5159493670886077e-06, |
|
"loss": 0.0758, |
|
"step": 34025 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.5096202531645571e-06, |
|
"loss": 0.0767, |
|
"step": 34050 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.5032911392405063e-06, |
|
"loss": 0.0651, |
|
"step": 34075 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.4969620253164558e-06, |
|
"loss": 0.0772, |
|
"step": 34100 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.490632911392405e-06, |
|
"loss": 0.0759, |
|
"step": 34125 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.4843037974683546e-06, |
|
"loss": 0.0697, |
|
"step": 34150 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.477974683544304e-06, |
|
"loss": 0.0716, |
|
"step": 34175 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.4716455696202533e-06, |
|
"loss": 0.0808, |
|
"step": 34200 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.4653164556962027e-06, |
|
"loss": 0.0806, |
|
"step": 34225 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.458987341772152e-06, |
|
"loss": 0.0805, |
|
"step": 34250 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.4526582278481013e-06, |
|
"loss": 0.0664, |
|
"step": 34275 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.4463291139240508e-06, |
|
"loss": 0.0812, |
|
"step": 34300 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.44e-06, |
|
"loss": 0.0818, |
|
"step": 34325 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.4336708860759496e-06, |
|
"loss": 0.0711, |
|
"step": 34350 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.427341772151899e-06, |
|
"loss": 0.08, |
|
"step": 34375 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.4210126582278483e-06, |
|
"loss": 0.0713, |
|
"step": 34400 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.4146835443037977e-06, |
|
"loss": 0.0735, |
|
"step": 34425 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.408354430379747e-06, |
|
"loss": 0.0883, |
|
"step": 34450 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.4020253164556963e-06, |
|
"loss": 0.0759, |
|
"step": 34475 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.3956962025316455e-06, |
|
"loss": 0.0679, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.389367088607595e-06, |
|
"loss": 0.0683, |
|
"step": 34525 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.3830379746835442e-06, |
|
"loss": 0.0773, |
|
"step": 34550 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.3767088607594938e-06, |
|
"loss": 0.0772, |
|
"step": 34575 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.3703797468354433e-06, |
|
"loss": 0.0685, |
|
"step": 34600 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.3640506329113925e-06, |
|
"loss": 0.0785, |
|
"step": 34625 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.357721518987342e-06, |
|
"loss": 0.0557, |
|
"step": 34650 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.3513924050632913e-06, |
|
"loss": 0.0439, |
|
"step": 34675 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 1.3450632911392405e-06, |
|
"loss": 0.0406, |
|
"step": 34700 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 1.33873417721519e-06, |
|
"loss": 0.0415, |
|
"step": 34725 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 1.3324050632911392e-06, |
|
"loss": 0.039, |
|
"step": 34750 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 1.3260759493670888e-06, |
|
"loss": 0.0423, |
|
"step": 34775 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 1.3197468354430383e-06, |
|
"loss": 0.0381, |
|
"step": 34800 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 1.3134177215189875e-06, |
|
"loss": 0.0383, |
|
"step": 34825 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 1.307088607594937e-06, |
|
"loss": 0.0491, |
|
"step": 34850 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 1.3007594936708861e-06, |
|
"loss": 0.0389, |
|
"step": 34875 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 1.2944303797468355e-06, |
|
"loss": 0.037, |
|
"step": 34900 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 1.2881012658227848e-06, |
|
"loss": 0.0432, |
|
"step": 34925 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 1.2817721518987342e-06, |
|
"loss": 0.04, |
|
"step": 34950 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 1.2754430379746838e-06, |
|
"loss": 0.0435, |
|
"step": 34975 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 1.269113924050633e-06, |
|
"loss": 0.042, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"eval_cer": 6.962366497085677, |
|
"eval_loss": 0.20532827079296112, |
|
"eval_runtime": 2061.136, |
|
"eval_samples_per_second": 0.915, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.34039764037579, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 1.2627848101265825e-06, |
|
"loss": 0.0362, |
|
"step": 35025 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 1.2564556962025317e-06, |
|
"loss": 0.0382, |
|
"step": 35050 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 1.2501265822784811e-06, |
|
"loss": 0.0407, |
|
"step": 35075 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 1.2437974683544305e-06, |
|
"loss": 0.0435, |
|
"step": 35100 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 1.23746835443038e-06, |
|
"loss": 0.0406, |
|
"step": 35125 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.2311392405063292e-06, |
|
"loss": 0.0439, |
|
"step": 35150 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.2248101265822786e-06, |
|
"loss": 0.0415, |
|
"step": 35175 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.2184810126582278e-06, |
|
"loss": 0.0371, |
|
"step": 35200 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.2121518987341775e-06, |
|
"loss": 0.0392, |
|
"step": 35225 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.2060759493670887e-06, |
|
"loss": 0.0419, |
|
"step": 35250 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 1.1997468354430381e-06, |
|
"loss": 0.0421, |
|
"step": 35275 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 1.1934177215189874e-06, |
|
"loss": 0.0353, |
|
"step": 35300 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 1.1870886075949368e-06, |
|
"loss": 0.0382, |
|
"step": 35325 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 1.1807594936708862e-06, |
|
"loss": 0.0309, |
|
"step": 35350 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 1.1744303797468354e-06, |
|
"loss": 0.0413, |
|
"step": 35375 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 1.1681012658227848e-06, |
|
"loss": 0.0416, |
|
"step": 35400 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 1.1617721518987343e-06, |
|
"loss": 0.0407, |
|
"step": 35425 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 1.1554430379746837e-06, |
|
"loss": 0.0399, |
|
"step": 35450 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 1.149113924050633e-06, |
|
"loss": 0.0339, |
|
"step": 35475 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 1.1427848101265823e-06, |
|
"loss": 0.0337, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 1.1364556962025318e-06, |
|
"loss": 0.0423, |
|
"step": 35525 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 1.1301265822784812e-06, |
|
"loss": 0.0423, |
|
"step": 35550 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 1.1237974683544304e-06, |
|
"loss": 0.0429, |
|
"step": 35575 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 1.1174683544303798e-06, |
|
"loss": 0.039, |
|
"step": 35600 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.1111392405063293e-06, |
|
"loss": 0.0369, |
|
"step": 35625 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.1048101265822787e-06, |
|
"loss": 0.0547, |
|
"step": 35650 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.098481012658228e-06, |
|
"loss": 0.0383, |
|
"step": 35675 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.0921518987341773e-06, |
|
"loss": 0.0355, |
|
"step": 35700 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.0858227848101268e-06, |
|
"loss": 0.0345, |
|
"step": 35725 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 1.079493670886076e-06, |
|
"loss": 0.0413, |
|
"step": 35750 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 1.0731645569620254e-06, |
|
"loss": 0.0402, |
|
"step": 35775 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 1.0668354430379746e-06, |
|
"loss": 0.0443, |
|
"step": 35800 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 1.0605063291139243e-06, |
|
"loss": 0.0348, |
|
"step": 35825 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 1.0541772151898735e-06, |
|
"loss": 0.038, |
|
"step": 35850 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 1.047848101265823e-06, |
|
"loss": 0.0377, |
|
"step": 35875 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 1.0415189873417721e-06, |
|
"loss": 0.0389, |
|
"step": 35900 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 1.0351898734177216e-06, |
|
"loss": 0.0427, |
|
"step": 35925 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 1.028860759493671e-06, |
|
"loss": 0.0366, |
|
"step": 35950 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 1.0225316455696204e-06, |
|
"loss": 0.0328, |
|
"step": 35975 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 1.0162025316455696e-06, |
|
"loss": 0.0474, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"eval_cer": 7.030591961187292, |
|
"eval_loss": 0.20971696078777313, |
|
"eval_runtime": 2070.0469, |
|
"eval_samples_per_second": 0.911, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.553419270264364, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 1.009873417721519e-06, |
|
"loss": 0.0327, |
|
"step": 36025 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 1.0035443037974685e-06, |
|
"loss": 0.0395, |
|
"step": 36050 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 9.97215189873418e-07, |
|
"loss": 0.0414, |
|
"step": 36075 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 9.908860759493671e-07, |
|
"loss": 0.0408, |
|
"step": 36100 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 9.845569620253166e-07, |
|
"loss": 0.0423, |
|
"step": 36125 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 9.78227848101266e-07, |
|
"loss": 0.0449, |
|
"step": 36150 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 9.718987341772152e-07, |
|
"loss": 0.0359, |
|
"step": 36175 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 9.655696202531646e-07, |
|
"loss": 0.0418, |
|
"step": 36200 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 9.592405063291138e-07, |
|
"loss": 0.0405, |
|
"step": 36225 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 9.529113924050634e-07, |
|
"loss": 0.0436, |
|
"step": 36250 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 9.465822784810127e-07, |
|
"loss": 0.0375, |
|
"step": 36275 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 9.402531645569621e-07, |
|
"loss": 0.0405, |
|
"step": 36300 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 9.339240506329115e-07, |
|
"loss": 0.0373, |
|
"step": 36325 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 9.275949367088609e-07, |
|
"loss": 0.0402, |
|
"step": 36350 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 9.212658227848102e-07, |
|
"loss": 0.0423, |
|
"step": 36375 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 9.149367088607595e-07, |
|
"loss": 0.0417, |
|
"step": 36400 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 9.086075949367088e-07, |
|
"loss": 0.0421, |
|
"step": 36425 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 9.022784810126584e-07, |
|
"loss": 0.0399, |
|
"step": 36450 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 8.959493670886077e-07, |
|
"loss": 0.0389, |
|
"step": 36475 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 8.89620253164557e-07, |
|
"loss": 0.0391, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 8.832911392405063e-07, |
|
"loss": 0.035, |
|
"step": 36525 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 8.769620253164559e-07, |
|
"loss": 0.043, |
|
"step": 36550 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 8.706329113924052e-07, |
|
"loss": 0.0439, |
|
"step": 36575 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 8.643037974683545e-07, |
|
"loss": 0.0442, |
|
"step": 36600 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 8.579746835443038e-07, |
|
"loss": 0.0332, |
|
"step": 36625 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 8.516455696202533e-07, |
|
"loss": 0.0413, |
|
"step": 36650 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 8.453164556962026e-07, |
|
"loss": 0.0417, |
|
"step": 36675 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 8.389873417721519e-07, |
|
"loss": 0.0406, |
|
"step": 36700 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 8.326582278481013e-07, |
|
"loss": 0.0379, |
|
"step": 36725 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 8.263291139240508e-07, |
|
"loss": 0.0417, |
|
"step": 36750 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 8.200000000000001e-07, |
|
"loss": 0.0423, |
|
"step": 36775 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 8.136708860759494e-07, |
|
"loss": 0.0402, |
|
"step": 36800 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 8.073417721518987e-07, |
|
"loss": 0.0385, |
|
"step": 36825 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 8.010126582278481e-07, |
|
"loss": 0.0471, |
|
"step": 36850 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 7.946835443037976e-07, |
|
"loss": 0.0371, |
|
"step": 36875 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 7.883544303797469e-07, |
|
"loss": 0.0424, |
|
"step": 36900 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 7.820253164556962e-07, |
|
"loss": 0.0394, |
|
"step": 36925 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 7.756962025316456e-07, |
|
"loss": 0.0377, |
|
"step": 36950 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 7.693670886075951e-07, |
|
"loss": 0.0409, |
|
"step": 36975 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 7.630379746835444e-07, |
|
"loss": 0.0428, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"eval_cer": 6.980896870051548, |
|
"eval_loss": 0.2106560468673706, |
|
"eval_runtime": 2064.6958, |
|
"eval_samples_per_second": 0.913, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.31854926807953, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 7.567088607594937e-07, |
|
"loss": 0.0348, |
|
"step": 37025 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 7.503797468354431e-07, |
|
"loss": 0.0407, |
|
"step": 37050 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 7.440506329113925e-07, |
|
"loss": 0.0426, |
|
"step": 37075 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 7.377215189873418e-07, |
|
"loss": 0.0488, |
|
"step": 37100 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 7.313924050632912e-07, |
|
"loss": 0.0421, |
|
"step": 37125 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 7.250632911392406e-07, |
|
"loss": 0.0357, |
|
"step": 37150 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 7.1873417721519e-07, |
|
"loss": 0.0411, |
|
"step": 37175 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 7.124050632911393e-07, |
|
"loss": 0.0428, |
|
"step": 37200 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 7.060759493670886e-07, |
|
"loss": 0.0354, |
|
"step": 37225 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 6.997468354430379e-07, |
|
"loss": 0.0428, |
|
"step": 37250 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 6.934177215189875e-07, |
|
"loss": 0.0352, |
|
"step": 37275 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 6.870886075949368e-07, |
|
"loss": 0.0393, |
|
"step": 37300 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 6.807594936708861e-07, |
|
"loss": 0.0369, |
|
"step": 37325 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 6.744303797468354e-07, |
|
"loss": 0.0367, |
|
"step": 37350 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 6.681012658227849e-07, |
|
"loss": 0.0417, |
|
"step": 37375 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 6.617721518987343e-07, |
|
"loss": 0.0397, |
|
"step": 37400 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 6.554430379746836e-07, |
|
"loss": 0.0401, |
|
"step": 37425 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 6.491139240506329e-07, |
|
"loss": 0.0351, |
|
"step": 37450 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 6.427848101265824e-07, |
|
"loss": 0.0444, |
|
"step": 37475 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 6.364556962025317e-07, |
|
"loss": 0.0366, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 6.30126582278481e-07, |
|
"loss": 0.0397, |
|
"step": 37525 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 6.237974683544304e-07, |
|
"loss": 0.0343, |
|
"step": 37550 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 6.174683544303798e-07, |
|
"loss": 0.0365, |
|
"step": 37575 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 6.111392405063292e-07, |
|
"loss": 0.0387, |
|
"step": 37600 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 6.048101265822785e-07, |
|
"loss": 0.0433, |
|
"step": 37625 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 5.984810126582279e-07, |
|
"loss": 0.0408, |
|
"step": 37650 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 5.921518987341773e-07, |
|
"loss": 0.0391, |
|
"step": 37675 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 5.858227848101266e-07, |
|
"loss": 0.0389, |
|
"step": 37700 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 5.79493670886076e-07, |
|
"loss": 0.0401, |
|
"step": 37725 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 5.731645569620253e-07, |
|
"loss": 0.0331, |
|
"step": 37750 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 5.668354430379748e-07, |
|
"loss": 0.0419, |
|
"step": 37775 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 5.605063291139241e-07, |
|
"loss": 0.0323, |
|
"step": 37800 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 5.541772151898735e-07, |
|
"loss": 0.0356, |
|
"step": 37825 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 5.478481012658228e-07, |
|
"loss": 0.0416, |
|
"step": 37850 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 5.415189873417722e-07, |
|
"loss": 0.0409, |
|
"step": 37875 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 5.351898734177216e-07, |
|
"loss": 0.0449, |
|
"step": 37900 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 5.288607594936709e-07, |
|
"loss": 0.043, |
|
"step": 37925 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 5.225316455696202e-07, |
|
"loss": 0.0375, |
|
"step": 37950 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 5.162025316455697e-07, |
|
"loss": 0.0365, |
|
"step": 37975 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 5.09873417721519e-07, |
|
"loss": 0.0343, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"eval_cer": 6.951416731242209, |
|
"eval_loss": 0.21110448241233826, |
|
"eval_runtime": 2062.8842, |
|
"eval_samples_per_second": 0.914, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.389556478042383, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 5.035443037974684e-07, |
|
"loss": 0.0448, |
|
"step": 38025 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 4.972151898734177e-07, |
|
"loss": 0.0394, |
|
"step": 38050 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 4.908860759493672e-07, |
|
"loss": 0.043, |
|
"step": 38075 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 4.845569620253165e-07, |
|
"loss": 0.0356, |
|
"step": 38100 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 4.782278481012659e-07, |
|
"loss": 0.0426, |
|
"step": 38125 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 4.718987341772152e-07, |
|
"loss": 0.0484, |
|
"step": 38150 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 4.655696202531646e-07, |
|
"loss": 0.0458, |
|
"step": 38175 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 4.592405063291139e-07, |
|
"loss": 0.0381, |
|
"step": 38200 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 4.5291139240506335e-07, |
|
"loss": 0.0417, |
|
"step": 38225 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 4.4658227848101267e-07, |
|
"loss": 0.0382, |
|
"step": 38250 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 4.402531645569621e-07, |
|
"loss": 0.0345, |
|
"step": 38275 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 4.339240506329114e-07, |
|
"loss": 0.0361, |
|
"step": 38300 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 4.275949367088608e-07, |
|
"loss": 0.0401, |
|
"step": 38325 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 4.2126582278481017e-07, |
|
"loss": 0.0388, |
|
"step": 38350 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 4.1493670886075954e-07, |
|
"loss": 0.0297, |
|
"step": 38375 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 4.0860759493670887e-07, |
|
"loss": 0.0407, |
|
"step": 38400 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 4.022784810126583e-07, |
|
"loss": 0.0401, |
|
"step": 38425 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 3.959493670886076e-07, |
|
"loss": 0.0388, |
|
"step": 38450 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 3.8962025316455704e-07, |
|
"loss": 0.0448, |
|
"step": 38475 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 3.8329113924050637e-07, |
|
"loss": 0.037, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 3.7696202531645574e-07, |
|
"loss": 0.0367, |
|
"step": 38525 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 3.706329113924051e-07, |
|
"loss": 0.0451, |
|
"step": 38550 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 3.643037974683545e-07, |
|
"loss": 0.0397, |
|
"step": 38575 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 3.579746835443038e-07, |
|
"loss": 0.0474, |
|
"step": 38600 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 3.516455696202532e-07, |
|
"loss": 0.0344, |
|
"step": 38625 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 3.4531645569620256e-07, |
|
"loss": 0.042, |
|
"step": 38650 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 3.389873417721519e-07, |
|
"loss": 0.0376, |
|
"step": 38675 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 3.326582278481013e-07, |
|
"loss": 0.0347, |
|
"step": 38700 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 3.2632911392405063e-07, |
|
"loss": 0.0467, |
|
"step": 38725 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 3.2e-07, |
|
"loss": 0.0357, |
|
"step": 38750 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 3.136708860759494e-07, |
|
"loss": 0.0389, |
|
"step": 38775 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 3.0734177215189876e-07, |
|
"loss": 0.0387, |
|
"step": 38800 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 3.0101265822784813e-07, |
|
"loss": 0.039, |
|
"step": 38825 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 2.946835443037975e-07, |
|
"loss": 0.0355, |
|
"step": 38850 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 2.883544303797469e-07, |
|
"loss": 0.0406, |
|
"step": 38875 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 2.8202531645569626e-07, |
|
"loss": 0.0397, |
|
"step": 38900 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 2.7569620253164563e-07, |
|
"loss": 0.0408, |
|
"step": 38925 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 2.6936708860759495e-07, |
|
"loss": 0.0471, |
|
"step": 38950 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 2.6303797468354433e-07, |
|
"loss": 0.0359, |
|
"step": 38975 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 2.567088607594937e-07, |
|
"loss": 0.0378, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"eval_cer": 6.875610660018193, |
|
"eval_loss": 0.2102653980255127, |
|
"eval_runtime": 2061.0305, |
|
"eval_samples_per_second": 0.915, |
|
"eval_steps_per_second": 0.115, |
|
"eval_wer": 21.258466244264802, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 2.503797468354431e-07, |
|
"loss": 0.0405, |
|
"step": 39025 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 2.440506329113924e-07, |
|
"loss": 0.0385, |
|
"step": 39050 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 2.3772151898734177e-07, |
|
"loss": 0.0472, |
|
"step": 39075 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 2.3139240506329115e-07, |
|
"loss": 0.0374, |
|
"step": 39100 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 2.2506329113924052e-07, |
|
"loss": 0.0361, |
|
"step": 39125 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 2.1873417721518987e-07, |
|
"loss": 0.0338, |
|
"step": 39150 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 2.1240506329113925e-07, |
|
"loss": 0.044, |
|
"step": 39175 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 2.0607594936708862e-07, |
|
"loss": 0.0394, |
|
"step": 39200 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 1.99746835443038e-07, |
|
"loss": 0.0412, |
|
"step": 39225 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 1.9341772151898734e-07, |
|
"loss": 0.038, |
|
"step": 39250 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 1.8708860759493672e-07, |
|
"loss": 0.0423, |
|
"step": 39275 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 1.807594936708861e-07, |
|
"loss": 0.0401, |
|
"step": 39300 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 1.7443037974683544e-07, |
|
"loss": 0.0391, |
|
"step": 39325 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 1.6835443037974686e-07, |
|
"loss": 0.0408, |
|
"step": 39350 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 1.620253164556962e-07, |
|
"loss": 0.0385, |
|
"step": 39375 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 1.556962025316456e-07, |
|
"loss": 0.0361, |
|
"step": 39400 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 1.4936708860759494e-07, |
|
"loss": 0.0397, |
|
"step": 39425 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 1.430379746835443e-07, |
|
"loss": 0.0382, |
|
"step": 39450 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 1.3670886075949369e-07, |
|
"loss": 0.0312, |
|
"step": 39475 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 1.3037974683544303e-07, |
|
"loss": 0.0432, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 1.240506329113924e-07, |
|
"loss": 0.0463, |
|
"step": 39525 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 1.1772151898734178e-07, |
|
"loss": 0.0402, |
|
"step": 39550 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 1.1139240506329114e-07, |
|
"loss": 0.0384, |
|
"step": 39575 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 1.0506329113924052e-07, |
|
"loss": 0.0342, |
|
"step": 39600 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 9.873417721518988e-08, |
|
"loss": 0.0367, |
|
"step": 39625 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 9.240506329113926e-08, |
|
"loss": 0.0398, |
|
"step": 39650 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 8.607594936708862e-08, |
|
"loss": 0.0408, |
|
"step": 39675 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 7.974683544303799e-08, |
|
"loss": 0.0356, |
|
"step": 39700 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 7.341772151898734e-08, |
|
"loss": 0.0434, |
|
"step": 39725 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 6.708860759493672e-08, |
|
"loss": 0.0424, |
|
"step": 39750 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 6.075949367088608e-08, |
|
"loss": 0.0425, |
|
"step": 39775 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 5.4430379746835445e-08, |
|
"loss": 0.0429, |
|
"step": 39800 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 4.810126582278481e-08, |
|
"loss": 0.0394, |
|
"step": 39825 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 4.177215189873418e-08, |
|
"loss": 0.0399, |
|
"step": 39850 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 3.544303797468355e-08, |
|
"loss": 0.036, |
|
"step": 39875 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 2.9113924050632914e-08, |
|
"loss": 0.0372, |
|
"step": 39900 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 2.2784810126582282e-08, |
|
"loss": 0.0377, |
|
"step": 39925 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 1.6455696202531647e-08, |
|
"loss": 0.0327, |
|
"step": 39950 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 1.0126582278481013e-08, |
|
"loss": 0.0362, |
|
"step": 39975 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 3.79746835443038e-09, |
|
"loss": 0.0361, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"eval_cer": 6.900879350426199, |
|
"eval_loss": 0.21062006056308746, |
|
"eval_runtime": 2061.1354, |
|
"eval_samples_per_second": 0.915, |
|
"eval_steps_per_second": 0.115, |
|
"eval_wer": 21.36770810574612, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"step": 40000, |
|
"total_flos": 3.2659050859167744e+20, |
|
"train_loss": 0.05977862850278616, |
|
"train_runtime": 98510.5981, |
|
"train_samples_per_second": 3.248, |
|
"train_steps_per_second": 0.406 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"eval_cer": 6.875610660018193, |
|
"eval_loss": 0.2102653980255127, |
|
"eval_runtime": 2061.4518, |
|
"eval_samples_per_second": 0.915, |
|
"eval_steps_per_second": 0.114, |
|
"eval_wer": 21.258466244264802, |
|
"step": 40000 |
|
} |
|
], |
|
"max_steps": 40000, |
|
"num_train_epochs": 4, |
|
"total_flos": 3.2659050859167744e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|