|
{ |
|
"best_metric": 0.43305763602256775, |
|
"best_model_checkpoint": "model/E5/typhoon_E5_typhoon_shuffle_augment_gpt4/checkpoint-8439", |
|
"epoch": 9.996445076430856, |
|
"eval_steps": 500, |
|
"global_step": 14060, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.777777777777778e-05, |
|
"loss": 3.0727, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.999998996341642e-05, |
|
"loss": 2.1876, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.999963868383706e-05, |
|
"loss": 1.9128, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.9998785583137e-05, |
|
"loss": 1.8314, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.999743067844064e-05, |
|
"loss": 1.773, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.999557399694518e-05, |
|
"loss": 1.7195, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.9993215575920024e-05, |
|
"loss": 1.6424, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.999035546270608e-05, |
|
"loss": 1.6174, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.998699371471479e-05, |
|
"loss": 1.5565, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.9983130399426966e-05, |
|
"loss": 1.5504, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.9978765594391474e-05, |
|
"loss": 1.5254, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.9973899387223616e-05, |
|
"loss": 1.4849, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.996853187560343e-05, |
|
"loss": 1.472, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.996266316727371e-05, |
|
"loss": 1.4301, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.995629338003782e-05, |
|
"loss": 1.4209, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.994942264175737e-05, |
|
"loss": 1.4046, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.9942051090349606e-05, |
|
"loss": 1.4021, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.9934178873784674e-05, |
|
"loss": 1.3303, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.992580615008264e-05, |
|
"loss": 1.3696, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.991693308731033e-05, |
|
"loss": 1.3401, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.990755986357791e-05, |
|
"loss": 1.332, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.989768666703538e-05, |
|
"loss": 1.2977, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.988731369586874e-05, |
|
"loss": 1.3081, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.987644115829604e-05, |
|
"loss": 1.2828, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.9865069272563195e-05, |
|
"loss": 1.2835, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.98531982669396e-05, |
|
"loss": 1.2671, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.9840828379713556e-05, |
|
"loss": 1.2453, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.9827959859187476e-05, |
|
"loss": 1.2327, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.9814592963672915e-05, |
|
"loss": 1.2441, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.980072796148535e-05, |
|
"loss": 1.2119, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.978636513093887e-05, |
|
"loss": 1.217, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.9771504760340494e-05, |
|
"loss": 1.1747, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.975614714798445e-05, |
|
"loss": 1.1939, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.9740292602146154e-05, |
|
"loss": 1.1603, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.972394144107606e-05, |
|
"loss": 1.1204, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.970709399299322e-05, |
|
"loss": 1.1431, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.968975059607874e-05, |
|
"loss": 1.1524, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.967191159846896e-05, |
|
"loss": 1.1309, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.9653577358248484e-05, |
|
"loss": 1.1314, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.9634748243442994e-05, |
|
"loss": 1.0861, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.9615424632011857e-05, |
|
"loss": 1.1026, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.959560691184052e-05, |
|
"loss": 1.108, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 4.957529548073276e-05, |
|
"loss": 1.0604, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 4.9554490746402696e-05, |
|
"loss": 1.1033, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 4.953319312646653e-05, |
|
"loss": 1.0724, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 4.951140304843428e-05, |
|
"loss": 1.0607, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 4.948912094970113e-05, |
|
"loss": 1.0467, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.946634727753864e-05, |
|
"loss": 1.0388, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.9443082489085814e-05, |
|
"loss": 0.9867, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.9419327051339883e-05, |
|
"loss": 1.0129, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.939508144114696e-05, |
|
"loss": 1.0029, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 4.937034614519245e-05, |
|
"loss": 1.0039, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.934512165999128e-05, |
|
"loss": 0.9735, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.931940849187795e-05, |
|
"loss": 0.9852, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.9293207156996354e-05, |
|
"loss": 0.9656, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.9266518181289414e-05, |
|
"loss": 1.0178, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 4.923934210048856e-05, |
|
"loss": 0.966, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 4.921167946010291e-05, |
|
"loss": 0.9721, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 4.9183530815408386e-05, |
|
"loss": 0.9451, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 4.9154896731436526e-05, |
|
"loss": 1.0005, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.9125777782963165e-05, |
|
"loss": 0.9435, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 4.909617455449689e-05, |
|
"loss": 0.9306, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.906608764026729e-05, |
|
"loss": 0.9644, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.903551764421307e-05, |
|
"loss": 0.9203, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.900446517996987e-05, |
|
"loss": 0.9102, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 4.8972930870857994e-05, |
|
"loss": 0.9134, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 4.89409153498699e-05, |
|
"loss": 0.8876, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.890841925965744e-05, |
|
"loss": 0.9273, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.8875443252519035e-05, |
|
"loss": 0.9117, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 4.884198799038652e-05, |
|
"loss": 0.8953, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.8869273066520691, |
|
"eval_runtime": 171.3717, |
|
"eval_samples_per_second": 32.462, |
|
"eval_steps_per_second": 8.117, |
|
"step": 1406 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.880805414481189e-05, |
|
"loss": 0.8234, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 4.8773642396953796e-05, |
|
"loss": 0.7706, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 4.87387534375639e-05, |
|
"loss": 0.8079, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 4.8703387966973e-05, |
|
"loss": 0.8082, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 4.866754669507696e-05, |
|
"loss": 0.7857, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 4.8631230341322455e-05, |
|
"loss": 0.7698, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 4.859443963469256e-05, |
|
"loss": 0.7786, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 4.855717531369208e-05, |
|
"loss": 0.7453, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 4.851943812633279e-05, |
|
"loss": 0.7654, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 4.848122883011832e-05, |
|
"loss": 0.7666, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 4.844254819202904e-05, |
|
"loss": 0.7849, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 4.840339698850661e-05, |
|
"loss": 0.7768, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 4.836377600543842e-05, |
|
"loss": 0.7652, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 4.832368603814182e-05, |
|
"loss": 0.7788, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 4.8283127891348124e-05, |
|
"loss": 0.7398, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.824210237918649e-05, |
|
"loss": 0.7389, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 4.820061032516756e-05, |
|
"loss": 0.74, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 4.815865256216693e-05, |
|
"loss": 0.7435, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 4.811622993240844e-05, |
|
"loss": 0.7257, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 4.807334328744726e-05, |
|
"loss": 0.7655, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 4.8029993488152806e-05, |
|
"loss": 0.7437, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 4.798618140469143e-05, |
|
"loss": 0.7188, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 4.794190791650903e-05, |
|
"loss": 0.7187, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 4.789717391231328e-05, |
|
"loss": 0.7358, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 4.7851980290055896e-05, |
|
"loss": 0.6868, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 4.7806327956914544e-05, |
|
"loss": 0.6775, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 4.7760217829274675e-05, |
|
"loss": 0.6768, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 4.771365083271112e-05, |
|
"loss": 0.7081, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 4.7666627901969454e-05, |
|
"loss": 0.6925, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 4.761914998094732e-05, |
|
"loss": 0.6841, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 4.7571218022675443e-05, |
|
"loss": 0.6839, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 4.7522832989298486e-05, |
|
"loss": 0.6912, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 4.747399585205575e-05, |
|
"loss": 0.6909, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 4.7424707591261685e-05, |
|
"loss": 0.6693, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 4.737496919628619e-05, |
|
"loss": 0.6909, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.732478166553479e-05, |
|
"loss": 0.6276, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 4.727414600642857e-05, |
|
"loss": 0.6702, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.722306323538392e-05, |
|
"loss": 0.6458, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.717153437779221e-05, |
|
"loss": 0.6527, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.711956046799917e-05, |
|
"loss": 0.6522, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.7067142549284085e-05, |
|
"loss": 0.6553, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.7014281673838904e-05, |
|
"loss": 0.6554, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 4.6960978902747135e-05, |
|
"loss": 0.6392, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 4.6907235305962476e-05, |
|
"loss": 0.6207, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 4.6853051962287405e-05, |
|
"loss": 0.6147, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 4.679842995935149e-05, |
|
"loss": 0.6345, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 4.674337039358957e-05, |
|
"loss": 0.6585, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.668787437021973e-05, |
|
"loss": 0.6293, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.6631943003221145e-05, |
|
"loss": 0.6337, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.6575577415311684e-05, |
|
"loss": 0.6347, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.6518778737925406e-05, |
|
"loss": 0.6265, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 4.646154811118982e-05, |
|
"loss": 0.6139, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 4.640388668390302e-05, |
|
"loss": 0.6191, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 4.6345795613510625e-05, |
|
"loss": 0.6035, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 4.6287276066082516e-05, |
|
"loss": 0.6031, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 4.6228329216289475e-05, |
|
"loss": 0.6121, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 4.616895624737957e-05, |
|
"loss": 0.5995, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.6109158351154416e-05, |
|
"loss": 0.5869, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 4.6048936727945255e-05, |
|
"loss": 0.5725, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 4.598829258658885e-05, |
|
"loss": 0.5962, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 4.592722714440324e-05, |
|
"loss": 0.599, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 4.586574162716328e-05, |
|
"loss": 0.604, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 4.5803837269076073e-05, |
|
"loss": 0.5809, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 4.5741515312756125e-05, |
|
"loss": 0.5661, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 4.567877700920049e-05, |
|
"loss": 0.5885, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 4.5615623617763606e-05, |
|
"loss": 0.5612, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.5552056406132003e-05, |
|
"loss": 0.5735, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.548807665029892e-05, |
|
"loss": 0.5808, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 4.542368563453861e-05, |
|
"loss": 0.5607, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 4.535888465138063e-05, |
|
"loss": 0.5605, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.6370431780815125, |
|
"eval_runtime": 170.9185, |
|
"eval_samples_per_second": 32.548, |
|
"eval_steps_per_second": 8.138, |
|
"step": 2813 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 4.529367500158386e-05, |
|
"loss": 0.5473, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 4.522805799411039e-05, |
|
"loss": 0.503, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 4.5162034946099277e-05, |
|
"loss": 0.5132, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.509560718284007e-05, |
|
"loss": 0.5009, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 4.502877603774622e-05, |
|
"loss": 0.5176, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.496154285232833e-05, |
|
"loss": 0.5032, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.489390897616719e-05, |
|
"loss": 0.5074, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 4.482587576688673e-05, |
|
"loss": 0.4971, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 4.4757444590126736e-05, |
|
"loss": 0.4858, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 4.4688616819515464e-05, |
|
"loss": 0.5065, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 4.461939383664202e-05, |
|
"loss": 0.5034, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 4.45497770310287e-05, |
|
"loss": 0.5047, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 4.4479767800103036e-05, |
|
"loss": 0.5026, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 4.4409367549169764e-05, |
|
"loss": 0.5083, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 4.433857769138261e-05, |
|
"loss": 0.4818, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 4.426739964771595e-05, |
|
"loss": 0.4851, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 4.4195834846936264e-05, |
|
"loss": 0.5099, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 4.4123884725573446e-05, |
|
"loss": 0.4986, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 4.4051550727892e-05, |
|
"loss": 0.4804, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.3978834305862004e-05, |
|
"loss": 0.4846, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.3905736919130034e-05, |
|
"loss": 0.4855, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 4.383226003498978e-05, |
|
"loss": 0.4869, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.375840512835266e-05, |
|
"loss": 0.4896, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.368417368171819e-05, |
|
"loss": 0.4845, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 4.3609567185144184e-05, |
|
"loss": 0.4822, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.3534587136216944e-05, |
|
"loss": 0.491, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.345923504002111e-05, |
|
"loss": 0.4819, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.338351240910945e-05, |
|
"loss": 0.4907, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.330742076347258e-05, |
|
"loss": 0.4777, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 4.3230961630508354e-05, |
|
"loss": 0.4864, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 4.315413654499128e-05, |
|
"loss": 0.4826, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 4.307694704904165e-05, |
|
"loss": 0.4728, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 4.299939469209463e-05, |
|
"loss": 0.4664, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 4.292148103086917e-05, |
|
"loss": 0.4747, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 4.2843207629336694e-05, |
|
"loss": 0.474, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 4.2764576058689735e-05, |
|
"loss": 0.4791, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 4.268558789731044e-05, |
|
"loss": 0.4768, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 4.260624473073883e-05, |
|
"loss": 0.4741, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 4.2526548151640986e-05, |
|
"loss": 0.4573, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 4.24464997597771e-05, |
|
"loss": 0.4798, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 4.236610116196934e-05, |
|
"loss": 0.4589, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 4.228535397206962e-05, |
|
"loss": 0.4737, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 4.220425981092716e-05, |
|
"loss": 0.4638, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 4.212282030635601e-05, |
|
"loss": 0.4621, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 4.204103709310234e-05, |
|
"loss": 0.4658, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 4.195891181281161e-05, |
|
"loss": 0.4637, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 4.187644611399566e-05, |
|
"loss": 0.4731, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 4.17936416519996e-05, |
|
"loss": 0.4719, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 4.171050008896855e-05, |
|
"loss": 0.4825, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 4.162702309381434e-05, |
|
"loss": 0.4491, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.1543212342181956e-05, |
|
"loss": 0.467, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.1459069516415916e-05, |
|
"loss": 0.4716, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 4.137459630552652e-05, |
|
"loss": 0.478, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 4.128979440515594e-05, |
|
"loss": 0.4498, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 4.1204665517544144e-05, |
|
"loss": 0.4706, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 4.1119211351494795e-05, |
|
"loss": 0.4628, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 4.103343362234089e-05, |
|
"loss": 0.4568, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 4.0947334051910367e-05, |
|
"loss": 0.4552, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 4.086091436849153e-05, |
|
"loss": 0.4421, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 4.077417630679833e-05, |
|
"loss": 0.4589, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 4.068712160793559e-05, |
|
"loss": 0.4503, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 4.0599752019364026e-05, |
|
"loss": 0.4597, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 4.0512069294865176e-05, |
|
"loss": 0.4387, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 4.042407519450619e-05, |
|
"loss": 0.4456, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 4.033577148460456e-05, |
|
"loss": 0.4391, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 4.024715993769253e-05, |
|
"loss": 0.4618, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 4.0158242332481654e-05, |
|
"loss": 0.4621, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 4.006902045382701e-05, |
|
"loss": 0.4364, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 3.997949609269143e-05, |
|
"loss": 0.4468, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 3.9889671046109464e-05, |
|
"loss": 0.4554, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.5355364680290222, |
|
"eval_runtime": 170.7962, |
|
"eval_samples_per_second": 32.571, |
|
"eval_steps_per_second": 8.144, |
|
"step": 4219 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 3.979954711715141e-05, |
|
"loss": 0.4513, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 3.9709126114887056e-05, |
|
"loss": 0.401, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 3.961840985434938e-05, |
|
"loss": 0.3981, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 3.952740015649812e-05, |
|
"loss": 0.4054, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 3.9436098848183226e-05, |
|
"loss": 0.3924, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 3.9344507762108165e-05, |
|
"loss": 0.4086, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 3.925262873679319e-05, |
|
"loss": 0.4095, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 3.916046361653836e-05, |
|
"loss": 0.3991, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 3.906801425138656e-05, |
|
"loss": 0.4067, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 3.89752824970864e-05, |
|
"loss": 0.402, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 3.888227021505486e-05, |
|
"loss": 0.4126, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 3.8788979272340066e-05, |
|
"loss": 0.4206, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 3.869541154158368e-05, |
|
"loss": 0.4187, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 3.860156890098339e-05, |
|
"loss": 0.4107, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 3.8507453234255176e-05, |
|
"loss": 0.4126, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 3.841306643059552e-05, |
|
"loss": 0.4081, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 3.8318410384643485e-05, |
|
"loss": 0.4109, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 3.822348699644264e-05, |
|
"loss": 0.4095, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 3.812829817140295e-05, |
|
"loss": 0.4068, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 3.8032845820262575e-05, |
|
"loss": 0.4002, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 3.793713185904942e-05, |
|
"loss": 0.4075, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 3.7841158209042756e-05, |
|
"loss": 0.4088, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 3.7744926796734596e-05, |
|
"loss": 0.3946, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 3.764843955379107e-05, |
|
"loss": 0.4004, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 3.7551698417013635e-05, |
|
"loss": 0.4077, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 3.7454705328300164e-05, |
|
"loss": 0.4015, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 3.735746223460604e-05, |
|
"loss": 0.4129, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 3.7259971087904984e-05, |
|
"loss": 0.4039, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 3.7162233845149944e-05, |
|
"loss": 0.3983, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 3.706425246823378e-05, |
|
"loss": 0.4059, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 3.69660289239499e-05, |
|
"loss": 0.4091, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 3.6867565183952764e-05, |
|
"loss": 0.401, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 3.67688632247183e-05, |
|
"loss": 0.3997, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 3.666992502750426e-05, |
|
"loss": 0.4017, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 3.657075257831043e-05, |
|
"loss": 0.3891, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 3.6471347867838766e-05, |
|
"loss": 0.4148, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 3.6371712891453424e-05, |
|
"loss": 0.4028, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 3.627184964914074e-05, |
|
"loss": 0.4149, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 3.617176014546906e-05, |
|
"loss": 0.3956, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 3.607144638954847e-05, |
|
"loss": 0.403, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 3.597091039499055e-05, |
|
"loss": 0.4101, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 3.587015417986788e-05, |
|
"loss": 0.4055, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 3.576917976667357e-05, |
|
"loss": 0.3912, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 3.566798918228062e-05, |
|
"loss": 0.3942, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 3.5566584457901304e-05, |
|
"loss": 0.3802, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 3.546496762904633e-05, |
|
"loss": 0.4022, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 3.536314073548402e-05, |
|
"loss": 0.4009, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 3.5261105821199344e-05, |
|
"loss": 0.4041, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 3.515886493435291e-05, |
|
"loss": 0.4123, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 3.505642012723983e-05, |
|
"loss": 0.3953, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 3.495377345624854e-05, |
|
"loss": 0.3994, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 3.4850926981819525e-05, |
|
"loss": 0.3978, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 3.4747882768403947e-05, |
|
"loss": 0.3955, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 3.464464288442219e-05, |
|
"loss": 0.4025, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 3.4541209402222396e-05, |
|
"loss": 0.3929, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 3.443758439803879e-05, |
|
"loss": 0.3876, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 3.433376995195008e-05, |
|
"loss": 0.3939, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 3.422976814783765e-05, |
|
"loss": 0.3927, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 3.4125581073343735e-05, |
|
"loss": 0.3945, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 3.4021210819829555e-05, |
|
"loss": 0.3889, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 3.391665948233328e-05, |
|
"loss": 0.4027, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 3.3811929159528024e-05, |
|
"loss": 0.3915, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 3.370702195367967e-05, |
|
"loss": 0.398, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 3.360193997060475e-05, |
|
"loss": 0.3965, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 3.349668531962807e-05, |
|
"loss": 0.3929, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 3.339126011354044e-05, |
|
"loss": 0.391, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 3.328566646855625e-05, |
|
"loss": 0.3796, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 3.3179906504270996e-05, |
|
"loss": 0.394, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 3.30739823436187e-05, |
|
"loss": 0.3899, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 3.2967896112829324e-05, |
|
"loss": 0.3927, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 3.286164994138612e-05, |
|
"loss": 0.3893, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.4602062702178955, |
|
"eval_runtime": 171.0396, |
|
"eval_samples_per_second": 32.525, |
|
"eval_steps_per_second": 8.133, |
|
"step": 5626 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 3.27552459619828e-05, |
|
"loss": 0.3656, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 3.26486863104808e-05, |
|
"loss": 0.3517, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 3.25419731258664e-05, |
|
"loss": 0.3473, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 3.2435108550207746e-05, |
|
"loss": 0.3588, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 3.232809472861189e-05, |
|
"loss": 0.3478, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 3.22209338091817e-05, |
|
"loss": 0.3504, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 3.211362794297278e-05, |
|
"loss": 0.357, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 3.200617928395028e-05, |
|
"loss": 0.3609, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 3.1898589988945596e-05, |
|
"loss": 0.3501, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 3.179086221761319e-05, |
|
"loss": 0.3547, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 3.1682998132387146e-05, |
|
"loss": 0.3587, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 3.15749998984378e-05, |
|
"loss": 0.3496, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 3.146686968362827e-05, |
|
"loss": 0.3547, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 3.135860965847096e-05, |
|
"loss": 0.356, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 3.125022199608396e-05, |
|
"loss": 0.3488, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 3.114170887214744e-05, |
|
"loss": 0.357, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 3.103307246485997e-05, |
|
"loss": 0.3486, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 3.092431495489484e-05, |
|
"loss": 0.3592, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 3.0815438525356194e-05, |
|
"loss": 0.3515, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 3.070644536173531e-05, |
|
"loss": 0.3495, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"learning_rate": 3.059733765186666e-05, |
|
"loss": 0.369, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 3.0488117585884037e-05, |
|
"loss": 0.3476, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 3.0378787356176557e-05, |
|
"loss": 0.3559, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 3.0269349157344667e-05, |
|
"loss": 0.3502, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 3.015980518615611e-05, |
|
"loss": 0.3507, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 3.0050157641501803e-05, |
|
"loss": 0.3452, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"learning_rate": 2.9940408724351694e-05, |
|
"loss": 0.3487, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 2.9830560637710614e-05, |
|
"loss": 0.3491, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 2.972061558657403e-05, |
|
"loss": 0.3516, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 2.9610575777883785e-05, |
|
"loss": 0.3527, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 2.9500443420483815e-05, |
|
"loss": 0.3599, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"learning_rate": 2.9390220725075778e-05, |
|
"loss": 0.3485, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 2.9279909904174717e-05, |
|
"loss": 0.3483, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 2.9169513172064634e-05, |
|
"loss": 0.3462, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 2.9059032744754022e-05, |
|
"loss": 0.3492, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 2.8948470839931403e-05, |
|
"loss": 0.3512, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 2.883782967692082e-05, |
|
"loss": 0.3501, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 2.872711147663726e-05, |
|
"loss": 0.352, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 2.8616318461542102e-05, |
|
"loss": 0.3523, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 2.8505452855598492e-05, |
|
"loss": 0.3498, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 2.8394516884226683e-05, |
|
"loss": 0.349, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 2.8283512774259414e-05, |
|
"loss": 0.3625, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 2.817244275389716e-05, |
|
"loss": 0.3449, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 2.806130905266342e-05, |
|
"loss": 0.3548, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 2.7950113901359974e-05, |
|
"loss": 0.3465, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 2.7838859532022116e-05, |
|
"loss": 0.3465, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 2.7727548177873798e-05, |
|
"loss": 0.3583, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 2.7616182073282854e-05, |
|
"loss": 0.3512, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"learning_rate": 2.7504763453716132e-05, |
|
"loss": 0.3607, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 2.7393294555694614e-05, |
|
"loss": 0.3508, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 2.728177761674854e-05, |
|
"loss": 0.3382, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 2.717021487537246e-05, |
|
"loss": 0.3477, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 2.7058608570980343e-05, |
|
"loss": 0.3514, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 4.76, |
|
"learning_rate": 2.6946960943860596e-05, |
|
"loss": 0.3496, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 2.6835274235131107e-05, |
|
"loss": 0.3557, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 2.6723550686694245e-05, |
|
"loss": 0.3527, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 2.661179254119187e-05, |
|
"loss": 0.3508, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 2.6500002041960338e-05, |
|
"loss": 0.3534, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 2.6388181432985405e-05, |
|
"loss": 0.3437, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 2.6276332958857246e-05, |
|
"loss": 0.3453, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 2.6164458864725384e-05, |
|
"loss": 0.3489, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 4.88, |
|
"learning_rate": 2.6052561396253595e-05, |
|
"loss": 0.3483, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 2.5940642799574876e-05, |
|
"loss": 0.3455, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 2.5828705321246304e-05, |
|
"loss": 0.3603, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"learning_rate": 2.5716751208204e-05, |
|
"loss": 0.3491, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 2.560478270771798e-05, |
|
"loss": 0.3383, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 2.549280206734705e-05, |
|
"loss": 0.3501, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 2.538081153489373e-05, |
|
"loss": 0.3462, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"learning_rate": 2.5268813358359084e-05, |
|
"loss": 0.3493, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 2.5156809785897623e-05, |
|
"loss": 0.3544, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.48198580741882324, |
|
"eval_runtime": 171.3048, |
|
"eval_samples_per_second": 32.474, |
|
"eval_steps_per_second": 8.12, |
|
"step": 7032 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 2.5044803065772165e-05, |
|
"loss": 0.3358, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 2.4932795446308734e-05, |
|
"loss": 0.3109, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 2.482078917585136e-05, |
|
"loss": 0.3119, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 2.4708786502717054e-05, |
|
"loss": 0.314, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 2.4596789675150577e-05, |
|
"loss": 0.3164, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"learning_rate": 2.4484800941279355e-05, |
|
"loss": 0.3078, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"learning_rate": 2.4372822549068354e-05, |
|
"loss": 0.3058, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 2.4260856746274963e-05, |
|
"loss": 0.3026, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 5.12, |
|
"learning_rate": 2.4148905780403844e-05, |
|
"loss": 0.3045, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"learning_rate": 2.4036971898661832e-05, |
|
"loss": 0.3097, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 5.15, |
|
"learning_rate": 2.392505734791285e-05, |
|
"loss": 0.3157, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 2.3813164374632775e-05, |
|
"loss": 0.3022, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 2.3701295224864356e-05, |
|
"loss": 0.307, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 5.19, |
|
"learning_rate": 2.3589452144172137e-05, |
|
"loss": 0.3149, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"learning_rate": 2.347763737759736e-05, |
|
"loss": 0.3075, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 5.22, |
|
"learning_rate": 2.336585316961292e-05, |
|
"loss": 0.312, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 5.23, |
|
"learning_rate": 2.325410176407833e-05, |
|
"loss": 0.3114, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 5.25, |
|
"learning_rate": 2.314238540419461e-05, |
|
"loss": 0.3183, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 2.303070633245933e-05, |
|
"loss": 0.3089, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 2.2919066790621575e-05, |
|
"loss": 0.312, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 5.29, |
|
"learning_rate": 2.280746901963693e-05, |
|
"loss": 0.3094, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 2.26959152596225e-05, |
|
"loss": 0.3099, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 2.2584407749811985e-05, |
|
"loss": 0.3111, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 2.2472948728510664e-05, |
|
"loss": 0.3133, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"learning_rate": 2.2361540433050492e-05, |
|
"loss": 0.3049, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 5.36, |
|
"learning_rate": 2.2250185099745253e-05, |
|
"loss": 0.3056, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 5.38, |
|
"learning_rate": 2.213888496384556e-05, |
|
"loss": 0.3085, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 5.39, |
|
"learning_rate": 2.2027642259494046e-05, |
|
"loss": 0.311, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 2.1916459219680557e-05, |
|
"loss": 0.3064, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 2.1805338076197234e-05, |
|
"loss": 0.3091, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 2.169428105959378e-05, |
|
"loss": 0.3031, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"learning_rate": 2.1583290399132695e-05, |
|
"loss": 0.3115, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 5.46, |
|
"learning_rate": 2.147236832274447e-05, |
|
"loss": 0.3091, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 5.47, |
|
"learning_rate": 2.1361517056982903e-05, |
|
"loss": 0.3062, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 2.1250738826980432e-05, |
|
"loss": 0.2961, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 2.1140035856403405e-05, |
|
"loss": 0.306, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 5.52, |
|
"learning_rate": 2.1029410367407476e-05, |
|
"loss": 0.3104, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 2.0918864580593034e-05, |
|
"loss": 0.3127, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 2.0808400714960567e-05, |
|
"loss": 0.3205, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"learning_rate": 2.0698020987866153e-05, |
|
"loss": 0.3037, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 2.058772761497694e-05, |
|
"loss": 0.3058, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 2.047752281022671e-05, |
|
"loss": 0.3088, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 2.0367408785771353e-05, |
|
"loss": 0.3092, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"learning_rate": 2.0257387751944556e-05, |
|
"loss": 0.3084, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"learning_rate": 2.014746191721337e-05, |
|
"loss": 0.3144, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 5.65, |
|
"learning_rate": 2.003763348813391e-05, |
|
"loss": 0.2987, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"learning_rate": 1.992790466930706e-05, |
|
"loss": 0.311, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"learning_rate": 1.98182776633342e-05, |
|
"loss": 0.3113, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 1.9708754670773005e-05, |
|
"loss": 0.3048, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 1.9599337890093302e-05, |
|
"loss": 0.3103, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 5.72, |
|
"learning_rate": 1.9490029517632884e-05, |
|
"loss": 0.3112, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 1.9380831747553458e-05, |
|
"loss": 0.31, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"learning_rate": 1.9271746771796607e-05, |
|
"loss": 0.3103, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 5.76, |
|
"learning_rate": 1.9162776780039766e-05, |
|
"loss": 0.3113, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 5.77, |
|
"learning_rate": 1.905392395965227e-05, |
|
"loss": 0.3167, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 5.79, |
|
"learning_rate": 1.8945190495651492e-05, |
|
"loss": 0.3082, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 1.8836578570658926e-05, |
|
"loss": 0.3016, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 1.872809036485637e-05, |
|
"loss": 0.3061, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"learning_rate": 1.8619728055942254e-05, |
|
"loss": 0.3113, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 5.84, |
|
"learning_rate": 1.851149381908781e-05, |
|
"loss": 0.3114, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 1.8403389826893476e-05, |
|
"loss": 0.3059, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 5.87, |
|
"learning_rate": 1.8295418249345283e-05, |
|
"loss": 0.311, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 1.8187581253771274e-05, |
|
"loss": 0.3145, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 5.9, |
|
"learning_rate": 1.8079881004798005e-05, |
|
"loss": 0.3126, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 5.92, |
|
"learning_rate": 1.797231966430712e-05, |
|
"loss": 0.3077, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 1.7864899391391915e-05, |
|
"loss": 0.3087, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 1.775762234231401e-05, |
|
"loss": 0.3021, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 1.7650490670460113e-05, |
|
"loss": 0.3076, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"learning_rate": 1.7543506526298713e-05, |
|
"loss": 0.3115, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 5.99, |
|
"learning_rate": 1.7436672057336967e-05, |
|
"loss": 0.3075, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 0.43305763602256775, |
|
"eval_runtime": 171.036, |
|
"eval_samples_per_second": 32.525, |
|
"eval_steps_per_second": 8.133, |
|
"step": 8439 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 1.7329989408077596e-05, |
|
"loss": 0.3047, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 1.722346071997582e-05, |
|
"loss": 0.2703, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 1.7117088131396355e-05, |
|
"loss": 0.2708, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 1.701087377757053e-05, |
|
"loss": 0.2652, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"learning_rate": 1.6904819790553407e-05, |
|
"loss": 0.2755, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 1.6798928299180978e-05, |
|
"loss": 0.2772, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 1.6693201429027427e-05, |
|
"loss": 0.2724, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 6.1, |
|
"learning_rate": 1.65876413023625e-05, |
|
"loss": 0.2712, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 1.6482250038108852e-05, |
|
"loss": 0.2676, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 1.6377029751799554e-05, |
|
"loss": 0.2785, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 6.14, |
|
"learning_rate": 1.627198255553562e-05, |
|
"loss": 0.2726, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 6.16, |
|
"learning_rate": 1.6167110557943588e-05, |
|
"loss": 0.2753, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 6.17, |
|
"learning_rate": 1.6062415864133213e-05, |
|
"loss": 0.2707, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"learning_rate": 1.595790057565522e-05, |
|
"loss": 0.2648, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"learning_rate": 1.5853566790459102e-05, |
|
"loss": 0.2739, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 1.574941660285098e-05, |
|
"loss": 0.2706, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 6.23, |
|
"learning_rate": 1.5645452103451657e-05, |
|
"loss": 0.2715, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 6.24, |
|
"learning_rate": 1.5541675379154548e-05, |
|
"loss": 0.2714, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 6.26, |
|
"learning_rate": 1.5438088513083826e-05, |
|
"loss": 0.2785, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 6.27, |
|
"learning_rate": 1.5334693584552655e-05, |
|
"loss": 0.2719, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 6.29, |
|
"learning_rate": 1.523149266902138e-05, |
|
"loss": 0.2741, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 6.3, |
|
"learning_rate": 1.5128487838055887e-05, |
|
"loss": 0.2622, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 6.31, |
|
"learning_rate": 1.5025681159286076e-05, |
|
"loss": 0.2703, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"learning_rate": 1.4923074696364265e-05, |
|
"loss": 0.2669, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"learning_rate": 1.4820670508923825e-05, |
|
"loss": 0.2743, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 6.36, |
|
"learning_rate": 1.4718470652537846e-05, |
|
"loss": 0.2729, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 6.37, |
|
"learning_rate": 1.461647717867783e-05, |
|
"loss": 0.2697, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 1.4514692134672523e-05, |
|
"loss": 0.275, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 1.4413117563666873e-05, |
|
"loss": 0.275, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 6.41, |
|
"learning_rate": 1.431175550458094e-05, |
|
"loss": 0.2734, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 6.43, |
|
"learning_rate": 1.4210607992069003e-05, |
|
"loss": 0.2654, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 1.4109677056478748e-05, |
|
"loss": 0.2687, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 1.4008964723810459e-05, |
|
"loss": 0.2777, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 6.47, |
|
"learning_rate": 1.3908473015676359e-05, |
|
"loss": 0.2708, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 6.48, |
|
"learning_rate": 1.3808203949260098e-05, |
|
"loss": 0.2746, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 1.3708159537276161e-05, |
|
"loss": 0.2745, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 6.51, |
|
"learning_rate": 1.3608341787929518e-05, |
|
"loss": 0.2779, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 1.3508752704875344e-05, |
|
"loss": 0.2713, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 1.3409394287178727e-05, |
|
"loss": 0.2731, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 1.331026852927459e-05, |
|
"loss": 0.2743, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 6.57, |
|
"learning_rate": 1.3211377420927657e-05, |
|
"loss": 0.2706, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 6.58, |
|
"learning_rate": 1.311272294719249e-05, |
|
"loss": 0.2724, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 1.3014307088373637e-05, |
|
"loss": 0.2792, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 6.61, |
|
"learning_rate": 1.2916131819985933e-05, |
|
"loss": 0.2705, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 6.63, |
|
"learning_rate": 1.2818199112714779e-05, |
|
"loss": 0.2742, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 6.64, |
|
"learning_rate": 1.2720510932376611e-05, |
|
"loss": 0.2717, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 1.2623069239879476e-05, |
|
"loss": 0.2759, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"learning_rate": 1.2525875991183606e-05, |
|
"loss": 0.2797, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 6.68, |
|
"learning_rate": 1.2428933137262196e-05, |
|
"loss": 0.277, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 6.7, |
|
"learning_rate": 1.2332242624062225e-05, |
|
"loss": 0.2741, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 1.2235806392465435e-05, |
|
"loss": 0.2632, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 1.2139626378249299e-05, |
|
"loss": 0.2686, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 6.74, |
|
"learning_rate": 1.2043704512048217e-05, |
|
"loss": 0.274, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 1.194804271931477e-05, |
|
"loss": 0.2744, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 1.1852642920281021e-05, |
|
"loss": 0.2718, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"learning_rate": 1.1757507029920009e-05, |
|
"loss": 0.2697, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"learning_rate": 1.1662636957907291e-05, |
|
"loss": 0.2765, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 1.1568034608582642e-05, |
|
"loss": 0.2693, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 6.83, |
|
"learning_rate": 1.1473701880911774e-05, |
|
"loss": 0.2701, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 1.1379640668448263e-05, |
|
"loss": 0.2749, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 6.85, |
|
"learning_rate": 1.1285852859295506e-05, |
|
"loss": 0.2749, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"learning_rate": 1.1192340336068874e-05, |
|
"loss": 0.2673, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"learning_rate": 1.1099104975857852e-05, |
|
"loss": 0.2716, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 6.9, |
|
"learning_rate": 1.1006148650188409e-05, |
|
"loss": 0.2657, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 1.09134732249854e-05, |
|
"loss": 0.2716, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 6.92, |
|
"learning_rate": 1.082108056053516e-05, |
|
"loss": 0.2706, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 1.0728972511448104e-05, |
|
"loss": 0.2756, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 6.95, |
|
"learning_rate": 1.063715092662152e-05, |
|
"loss": 0.2693, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 1.0545617649202486e-05, |
|
"loss": 0.2725, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"learning_rate": 1.0454374516550825e-05, |
|
"loss": 0.2697, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 1.036342336020224e-05, |
|
"loss": 0.2685, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 0.441493421792984, |
|
"eval_runtime": 171.0135, |
|
"eval_samples_per_second": 32.53, |
|
"eval_steps_per_second": 8.134, |
|
"step": 9845 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 1.0272766005831583e-05, |
|
"loss": 0.2493, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 1.0182404273216154e-05, |
|
"loss": 0.2425, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 1.0092339976199192e-05, |
|
"loss": 0.2442, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"learning_rate": 1.0002574922653506e-05, |
|
"loss": 0.2374, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"learning_rate": 9.91311091444512e-06, |
|
"loss": 0.2363, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"learning_rate": 9.823949747397134e-06, |
|
"loss": 0.2427, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 7.1, |
|
"learning_rate": 9.735093211253698e-06, |
|
"loss": 0.2417, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 7.11, |
|
"learning_rate": 9.64654308964405e-06, |
|
"loss": 0.236, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 7.12, |
|
"learning_rate": 9.558301160046717e-06, |
|
"loss": 0.2411, |
|
"step": 10020 |
|
}, |
|
{ |
|
"epoch": 7.14, |
|
"learning_rate": 9.470369193753877e-06, |
|
"loss": 0.245, |
|
"step": 10040 |
|
}, |
|
{ |
|
"epoch": 7.15, |
|
"learning_rate": 9.38274895583575e-06, |
|
"loss": 0.2474, |
|
"step": 10060 |
|
}, |
|
{ |
|
"epoch": 7.17, |
|
"learning_rate": 9.295442205105178e-06, |
|
"loss": 0.2398, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"learning_rate": 9.208450694082373e-06, |
|
"loss": 0.2383, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 9.121776168959667e-06, |
|
"loss": 0.2429, |
|
"step": 10120 |
|
}, |
|
{ |
|
"epoch": 7.21, |
|
"learning_rate": 9.035420369566485e-06, |
|
"loss": 0.2413, |
|
"step": 10140 |
|
}, |
|
{ |
|
"epoch": 7.22, |
|
"learning_rate": 8.949385029334459e-06, |
|
"loss": 0.2395, |
|
"step": 10160 |
|
}, |
|
{ |
|
"epoch": 7.24, |
|
"learning_rate": 8.863671875262577e-06, |
|
"loss": 0.2384, |
|
"step": 10180 |
|
}, |
|
{ |
|
"epoch": 7.25, |
|
"learning_rate": 8.778282627882536e-06, |
|
"loss": 0.2399, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 7.27, |
|
"learning_rate": 8.693219001224239e-06, |
|
"loss": 0.2408, |
|
"step": 10220 |
|
}, |
|
{ |
|
"epoch": 7.28, |
|
"learning_rate": 8.608482702781332e-06, |
|
"loss": 0.2382, |
|
"step": 10240 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 8.524075433476963e-06, |
|
"loss": 0.24, |
|
"step": 10260 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 8.439998887629649e-06, |
|
"loss": 0.2434, |
|
"step": 10280 |
|
}, |
|
{ |
|
"epoch": 7.32, |
|
"learning_rate": 8.356254752919241e-06, |
|
"loss": 0.2411, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 7.34, |
|
"learning_rate": 8.272844710353036e-06, |
|
"loss": 0.2417, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 7.35, |
|
"learning_rate": 8.189770434232096e-06, |
|
"loss": 0.2389, |
|
"step": 10340 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 8.10703359211757e-06, |
|
"loss": 0.2352, |
|
"step": 10360 |
|
}, |
|
{ |
|
"epoch": 7.38, |
|
"learning_rate": 8.02463584479724e-06, |
|
"loss": 0.234, |
|
"step": 10380 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 7.942578846252227e-06, |
|
"loss": 0.2399, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 7.41, |
|
"learning_rate": 7.860864243623726e-06, |
|
"loss": 0.2465, |
|
"step": 10420 |
|
}, |
|
{ |
|
"epoch": 7.42, |
|
"learning_rate": 7.779493677179971e-06, |
|
"loss": 0.2455, |
|
"step": 10440 |
|
}, |
|
{ |
|
"epoch": 7.44, |
|
"learning_rate": 7.698468780283344e-06, |
|
"loss": 0.2416, |
|
"step": 10460 |
|
}, |
|
{ |
|
"epoch": 7.45, |
|
"learning_rate": 7.617791179357522e-06, |
|
"loss": 0.2419, |
|
"step": 10480 |
|
}, |
|
{ |
|
"epoch": 7.47, |
|
"learning_rate": 7.537462493854866e-06, |
|
"loss": 0.2501, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 7.48, |
|
"learning_rate": 7.457484336223939e-06, |
|
"loss": 0.2349, |
|
"step": 10520 |
|
}, |
|
{ |
|
"epoch": 7.49, |
|
"learning_rate": 7.377858311877081e-06, |
|
"loss": 0.2488, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 7.51, |
|
"learning_rate": 7.298586019158216e-06, |
|
"loss": 0.2435, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 7.52, |
|
"learning_rate": 7.219669049310784e-06, |
|
"loss": 0.2453, |
|
"step": 10580 |
|
}, |
|
{ |
|
"epoch": 7.54, |
|
"learning_rate": 7.141108986445768e-06, |
|
"loss": 0.2441, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 7.55, |
|
"learning_rate": 7.062907407509903e-06, |
|
"loss": 0.2388, |
|
"step": 10620 |
|
}, |
|
{ |
|
"epoch": 7.56, |
|
"learning_rate": 6.985065882254046e-06, |
|
"loss": 0.2396, |
|
"step": 10640 |
|
}, |
|
{ |
|
"epoch": 7.58, |
|
"learning_rate": 6.907585973201633e-06, |
|
"loss": 0.2357, |
|
"step": 10660 |
|
}, |
|
{ |
|
"epoch": 7.59, |
|
"learning_rate": 6.830469235617323e-06, |
|
"loss": 0.2455, |
|
"step": 10680 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 6.7537172174758135e-06, |
|
"loss": 0.2471, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 7.62, |
|
"learning_rate": 6.677331459430713e-06, |
|
"loss": 0.2445, |
|
"step": 10720 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 6.601313494783648e-06, |
|
"loss": 0.2427, |
|
"step": 10740 |
|
}, |
|
{ |
|
"epoch": 7.65, |
|
"learning_rate": 6.525664849453478e-06, |
|
"loss": 0.239, |
|
"step": 10760 |
|
}, |
|
{ |
|
"epoch": 7.66, |
|
"learning_rate": 6.450387041945677e-06, |
|
"loss": 0.2453, |
|
"step": 10780 |
|
}, |
|
{ |
|
"epoch": 7.68, |
|
"learning_rate": 6.375481583321829e-06, |
|
"loss": 0.2401, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 7.69, |
|
"learning_rate": 6.3009499771693156e-06, |
|
"loss": 0.2417, |
|
"step": 10820 |
|
}, |
|
{ |
|
"epoch": 7.71, |
|
"learning_rate": 6.226793719571111e-06, |
|
"loss": 0.2463, |
|
"step": 10840 |
|
}, |
|
{ |
|
"epoch": 7.72, |
|
"learning_rate": 6.153014299075799e-06, |
|
"loss": 0.2366, |
|
"step": 10860 |
|
}, |
|
{ |
|
"epoch": 7.74, |
|
"learning_rate": 6.0796131966676324e-06, |
|
"loss": 0.237, |
|
"step": 10880 |
|
}, |
|
{ |
|
"epoch": 7.75, |
|
"learning_rate": 6.006591885736851e-06, |
|
"loss": 0.2364, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 7.76, |
|
"learning_rate": 5.9339518320500665e-06, |
|
"loss": 0.235, |
|
"step": 10920 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 5.861694493720898e-06, |
|
"loss": 0.2448, |
|
"step": 10940 |
|
}, |
|
{ |
|
"epoch": 7.79, |
|
"learning_rate": 5.789821321180639e-06, |
|
"loss": 0.2471, |
|
"step": 10960 |
|
}, |
|
{ |
|
"epoch": 7.81, |
|
"learning_rate": 5.718333757149183e-06, |
|
"loss": 0.2391, |
|
"step": 10980 |
|
}, |
|
{ |
|
"epoch": 7.82, |
|
"learning_rate": 5.647233236606037e-06, |
|
"loss": 0.241, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 7.84, |
|
"learning_rate": 5.576521186761563e-06, |
|
"loss": 0.2383, |
|
"step": 11020 |
|
}, |
|
{ |
|
"epoch": 7.85, |
|
"learning_rate": 5.506199027028272e-06, |
|
"loss": 0.2455, |
|
"step": 11040 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 5.436268168992356e-06, |
|
"loss": 0.2358, |
|
"step": 11060 |
|
}, |
|
{ |
|
"epoch": 7.88, |
|
"learning_rate": 5.36673001638538e-06, |
|
"loss": 0.2439, |
|
"step": 11080 |
|
}, |
|
{ |
|
"epoch": 7.89, |
|
"learning_rate": 5.297585965056056e-06, |
|
"loss": 0.245, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 7.91, |
|
"learning_rate": 5.228837402942252e-06, |
|
"loss": 0.2461, |
|
"step": 11120 |
|
}, |
|
{ |
|
"epoch": 7.92, |
|
"learning_rate": 5.1604857100431445e-06, |
|
"loss": 0.2315, |
|
"step": 11140 |
|
}, |
|
{ |
|
"epoch": 7.93, |
|
"learning_rate": 5.092532258391483e-06, |
|
"loss": 0.2451, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 7.95, |
|
"learning_rate": 5.0249784120260626e-06, |
|
"loss": 0.2364, |
|
"step": 11180 |
|
}, |
|
{ |
|
"epoch": 7.96, |
|
"learning_rate": 4.957825526964371e-06, |
|
"loss": 0.2425, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 7.98, |
|
"learning_rate": 4.891074951175328e-06, |
|
"loss": 0.2481, |
|
"step": 11220 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"learning_rate": 4.824728024552239e-06, |
|
"loss": 0.2408, |
|
"step": 11240 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 0.46929147839546204, |
|
"eval_runtime": 171.1743, |
|
"eval_samples_per_second": 32.499, |
|
"eval_steps_per_second": 8.126, |
|
"step": 11252 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 4.758786078885927e-06, |
|
"loss": 0.2229, |
|
"step": 11260 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 4.69325043783796e-06, |
|
"loss": 0.2182, |
|
"step": 11280 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 4.628122416914099e-06, |
|
"loss": 0.2152, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 8.05, |
|
"learning_rate": 4.563403323437909e-06, |
|
"loss": 0.2172, |
|
"step": 11320 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 4.499094456524478e-06, |
|
"loss": 0.2185, |
|
"step": 11340 |
|
}, |
|
{ |
|
"epoch": 8.08, |
|
"learning_rate": 4.435197107054364e-06, |
|
"loss": 0.2163, |
|
"step": 11360 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"learning_rate": 4.371712557647698e-06, |
|
"loss": 0.2195, |
|
"step": 11380 |
|
}, |
|
{ |
|
"epoch": 8.11, |
|
"learning_rate": 4.308642082638401e-06, |
|
"loss": 0.2256, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"learning_rate": 4.245986948048619e-06, |
|
"loss": 0.2176, |
|
"step": 11420 |
|
}, |
|
{ |
|
"epoch": 8.13, |
|
"learning_rate": 4.18374841156334e-06, |
|
"loss": 0.2234, |
|
"step": 11440 |
|
}, |
|
{ |
|
"epoch": 8.15, |
|
"learning_rate": 4.121927722505095e-06, |
|
"loss": 0.2227, |
|
"step": 11460 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"learning_rate": 4.060526121808916e-06, |
|
"loss": 0.2255, |
|
"step": 11480 |
|
}, |
|
{ |
|
"epoch": 8.18, |
|
"learning_rate": 3.999544841997427e-06, |
|
"loss": 0.2238, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 3.938985107156082e-06, |
|
"loss": 0.2179, |
|
"step": 11520 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 3.878848132908605e-06, |
|
"loss": 0.2217, |
|
"step": 11540 |
|
}, |
|
{ |
|
"epoch": 8.22, |
|
"learning_rate": 3.819135126392606e-06, |
|
"loss": 0.2218, |
|
"step": 11560 |
|
}, |
|
{ |
|
"epoch": 8.23, |
|
"learning_rate": 3.7598472862353157e-06, |
|
"loss": 0.2205, |
|
"step": 11580 |
|
}, |
|
{ |
|
"epoch": 8.25, |
|
"learning_rate": 3.700985802529544e-06, |
|
"loss": 0.2109, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 8.26, |
|
"learning_rate": 3.6425518568098087e-06, |
|
"loss": 0.2193, |
|
"step": 11620 |
|
}, |
|
{ |
|
"epoch": 8.28, |
|
"learning_rate": 3.584546622028581e-06, |
|
"loss": 0.218, |
|
"step": 11640 |
|
}, |
|
{ |
|
"epoch": 8.29, |
|
"learning_rate": 3.526971262532758e-06, |
|
"loss": 0.2161, |
|
"step": 11660 |
|
}, |
|
{ |
|
"epoch": 8.3, |
|
"learning_rate": 3.4698269340403157e-06, |
|
"loss": 0.2149, |
|
"step": 11680 |
|
}, |
|
{ |
|
"epoch": 8.32, |
|
"learning_rate": 3.4131147836170634e-06, |
|
"loss": 0.2141, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 8.33, |
|
"learning_rate": 3.356835949653642e-06, |
|
"loss": 0.2174, |
|
"step": 11720 |
|
}, |
|
{ |
|
"epoch": 8.35, |
|
"learning_rate": 3.3009915618426894e-06, |
|
"loss": 0.2254, |
|
"step": 11740 |
|
}, |
|
{ |
|
"epoch": 8.36, |
|
"learning_rate": 3.2455827411561364e-06, |
|
"loss": 0.22, |
|
"step": 11760 |
|
}, |
|
{ |
|
"epoch": 8.38, |
|
"learning_rate": 3.1906105998227104e-06, |
|
"loss": 0.2205, |
|
"step": 11780 |
|
}, |
|
{ |
|
"epoch": 8.39, |
|
"learning_rate": 3.136076241305633e-06, |
|
"loss": 0.2206, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"learning_rate": 3.081980760280437e-06, |
|
"loss": 0.2198, |
|
"step": 11820 |
|
}, |
|
{ |
|
"epoch": 8.42, |
|
"learning_rate": 3.0283252426130034e-06, |
|
"loss": 0.2262, |
|
"step": 11840 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 2.9751107653377934e-06, |
|
"loss": 0.2234, |
|
"step": 11860 |
|
}, |
|
{ |
|
"epoch": 8.45, |
|
"learning_rate": 2.9223383966361818e-06, |
|
"loss": 0.2215, |
|
"step": 11880 |
|
}, |
|
{ |
|
"epoch": 8.46, |
|
"learning_rate": 2.870009195815046e-06, |
|
"loss": 0.2214, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 8.47, |
|
"learning_rate": 2.8181242132854973e-06, |
|
"loss": 0.2202, |
|
"step": 11920 |
|
}, |
|
{ |
|
"epoch": 8.49, |
|
"learning_rate": 2.766684490541796e-06, |
|
"loss": 0.2136, |
|
"step": 11940 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 2.715691060140424e-06, |
|
"loss": 0.2136, |
|
"step": 11960 |
|
}, |
|
{ |
|
"epoch": 8.52, |
|
"learning_rate": 2.665144945679407e-06, |
|
"loss": 0.2146, |
|
"step": 11980 |
|
}, |
|
{ |
|
"epoch": 8.53, |
|
"learning_rate": 2.6150471617777116e-06, |
|
"loss": 0.2231, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 8.55, |
|
"learning_rate": 2.565398714054917e-06, |
|
"loss": 0.2234, |
|
"step": 12020 |
|
}, |
|
{ |
|
"epoch": 8.56, |
|
"learning_rate": 2.51620059911101e-06, |
|
"loss": 0.2208, |
|
"step": 12040 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"learning_rate": 2.4674538045063976e-06, |
|
"loss": 0.2227, |
|
"step": 12060 |
|
}, |
|
{ |
|
"epoch": 8.59, |
|
"learning_rate": 2.4191593087420613e-06, |
|
"loss": 0.2274, |
|
"step": 12080 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"learning_rate": 2.3713180812399317e-06, |
|
"loss": 0.2202, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 8.62, |
|
"learning_rate": 2.3239310823234215e-06, |
|
"loss": 0.2156, |
|
"step": 12120 |
|
}, |
|
{ |
|
"epoch": 8.63, |
|
"learning_rate": 2.2769992631981595e-06, |
|
"loss": 0.2208, |
|
"step": 12140 |
|
}, |
|
{ |
|
"epoch": 8.65, |
|
"learning_rate": 2.230523565932882e-06, |
|
"loss": 0.2178, |
|
"step": 12160 |
|
}, |
|
{ |
|
"epoch": 8.66, |
|
"learning_rate": 2.1845049234405306e-06, |
|
"loss": 0.2228, |
|
"step": 12180 |
|
}, |
|
{ |
|
"epoch": 8.67, |
|
"learning_rate": 2.1389442594595214e-06, |
|
"loss": 0.2213, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 2.093842488535219e-06, |
|
"loss": 0.2229, |
|
"step": 12220 |
|
}, |
|
{ |
|
"epoch": 8.7, |
|
"learning_rate": 2.049200516001554e-06, |
|
"loss": 0.225, |
|
"step": 12240 |
|
}, |
|
{ |
|
"epoch": 8.72, |
|
"learning_rate": 2.0050192379628656e-06, |
|
"loss": 0.2194, |
|
"step": 12260 |
|
}, |
|
{ |
|
"epoch": 8.73, |
|
"learning_rate": 1.9612995412759016e-06, |
|
"loss": 0.2232, |
|
"step": 12280 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 1.9180423035320416e-06, |
|
"loss": 0.2112, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 8.76, |
|
"learning_rate": 1.875248393039658e-06, |
|
"loss": 0.2132, |
|
"step": 12320 |
|
}, |
|
{ |
|
"epoch": 8.77, |
|
"learning_rate": 1.8329186688066797e-06, |
|
"loss": 0.2229, |
|
"step": 12340 |
|
}, |
|
{ |
|
"epoch": 8.79, |
|
"learning_rate": 1.7910539805233827e-06, |
|
"loss": 0.2132, |
|
"step": 12360 |
|
}, |
|
{ |
|
"epoch": 8.8, |
|
"learning_rate": 1.7496551685453028e-06, |
|
"loss": 0.2222, |
|
"step": 12380 |
|
}, |
|
{ |
|
"epoch": 8.82, |
|
"learning_rate": 1.7087230638763745e-06, |
|
"loss": 0.2242, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"learning_rate": 1.6682584881522634e-06, |
|
"loss": 0.2254, |
|
"step": 12420 |
|
}, |
|
{ |
|
"epoch": 8.84, |
|
"learning_rate": 1.6282622536238551e-06, |
|
"loss": 0.2208, |
|
"step": 12440 |
|
}, |
|
{ |
|
"epoch": 8.86, |
|
"learning_rate": 1.5887351631409614e-06, |
|
"loss": 0.2238, |
|
"step": 12460 |
|
}, |
|
{ |
|
"epoch": 8.87, |
|
"learning_rate": 1.5496780101362074e-06, |
|
"loss": 0.2186, |
|
"step": 12480 |
|
}, |
|
{ |
|
"epoch": 8.89, |
|
"learning_rate": 1.5110915786090918e-06, |
|
"loss": 0.2192, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 8.9, |
|
"learning_rate": 1.4729766431102604e-06, |
|
"loss": 0.2164, |
|
"step": 12520 |
|
}, |
|
{ |
|
"epoch": 8.92, |
|
"learning_rate": 1.4353339687259632e-06, |
|
"loss": 0.2198, |
|
"step": 12540 |
|
}, |
|
{ |
|
"epoch": 8.93, |
|
"learning_rate": 1.3981643110626775e-06, |
|
"loss": 0.2188, |
|
"step": 12560 |
|
}, |
|
{ |
|
"epoch": 8.94, |
|
"learning_rate": 1.3614684162319564e-06, |
|
"loss": 0.2228, |
|
"step": 12580 |
|
}, |
|
{ |
|
"epoch": 8.96, |
|
"learning_rate": 1.3252470208354518e-06, |
|
"loss": 0.2183, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 1.2895008519501206e-06, |
|
"loss": 0.2183, |
|
"step": 12620 |
|
}, |
|
{ |
|
"epoch": 8.99, |
|
"learning_rate": 1.2542306271136284e-06, |
|
"loss": 0.218, |
|
"step": 12640 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 0.47556841373443604, |
|
"eval_runtime": 170.6905, |
|
"eval_samples_per_second": 32.591, |
|
"eval_steps_per_second": 8.149, |
|
"step": 12658 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 1.2194370543099659e-06, |
|
"loss": 0.2195, |
|
"step": 12660 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 1.1851208319552109e-06, |
|
"loss": 0.2087, |
|
"step": 12680 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 1.1512826488835227e-06, |
|
"loss": 0.2115, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 1.1179231843333248e-06, |
|
"loss": 0.207, |
|
"step": 12720 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 1.085043107933642e-06, |
|
"loss": 0.21, |
|
"step": 12740 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"learning_rate": 1.0526430796906878e-06, |
|
"loss": 0.2057, |
|
"step": 12760 |
|
}, |
|
{ |
|
"epoch": 9.09, |
|
"learning_rate": 1.0207237499746002e-06, |
|
"loss": 0.2121, |
|
"step": 12780 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"learning_rate": 9.892857595063947e-07, |
|
"loss": 0.2129, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 9.11, |
|
"learning_rate": 9.583297393450929e-07, |
|
"loss": 0.2131, |
|
"step": 12820 |
|
}, |
|
{ |
|
"epoch": 9.13, |
|
"learning_rate": 9.278563108750665e-07, |
|
"loss": 0.2127, |
|
"step": 12840 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 8.978660857935555e-07, |
|
"loss": 0.2134, |
|
"step": 12860 |
|
}, |
|
{ |
|
"epoch": 9.16, |
|
"learning_rate": 8.68359666098395e-07, |
|
"loss": 0.2099, |
|
"step": 12880 |
|
}, |
|
{ |
|
"epoch": 9.17, |
|
"learning_rate": 8.393376440759326e-07, |
|
"loss": 0.209, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 9.19, |
|
"learning_rate": 8.108006022891274e-07, |
|
"loss": 0.2084, |
|
"step": 12920 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 7.827491135658726e-07, |
|
"loss": 0.2091, |
|
"step": 12940 |
|
}, |
|
{ |
|
"epoch": 9.21, |
|
"learning_rate": 7.551837409874862e-07, |
|
"loss": 0.2116, |
|
"step": 12960 |
|
}, |
|
{ |
|
"epoch": 9.23, |
|
"learning_rate": 7.281050378774135e-07, |
|
"loss": 0.2031, |
|
"step": 12980 |
|
}, |
|
{ |
|
"epoch": 9.24, |
|
"learning_rate": 7.015135477901086e-07, |
|
"loss": 0.2038, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 9.26, |
|
"learning_rate": 6.754098045001517e-07, |
|
"loss": 0.202, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 9.27, |
|
"learning_rate": 6.497943319914962e-07, |
|
"loss": 0.2145, |
|
"step": 13040 |
|
}, |
|
{ |
|
"epoch": 9.29, |
|
"learning_rate": 6.246676444469774e-07, |
|
"loss": 0.2117, |
|
"step": 13060 |
|
}, |
|
{ |
|
"epoch": 9.3, |
|
"learning_rate": 6.000302462379898e-07, |
|
"loss": 0.2067, |
|
"step": 13080 |
|
}, |
|
{ |
|
"epoch": 9.31, |
|
"learning_rate": 5.758826319143512e-07, |
|
"loss": 0.2014, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 9.33, |
|
"learning_rate": 5.5222528619438e-07, |
|
"loss": 0.209, |
|
"step": 13120 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"learning_rate": 5.29058683955172e-07, |
|
"loss": 0.2062, |
|
"step": 13140 |
|
}, |
|
{ |
|
"epoch": 9.36, |
|
"learning_rate": 5.063832902230586e-07, |
|
"loss": 0.2086, |
|
"step": 13160 |
|
}, |
|
{ |
|
"epoch": 9.37, |
|
"learning_rate": 4.841995601642751e-07, |
|
"loss": 0.2134, |
|
"step": 13180 |
|
}, |
|
{ |
|
"epoch": 9.38, |
|
"learning_rate": 4.625079390758319e-07, |
|
"loss": 0.2027, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 9.4, |
|
"learning_rate": 4.41308862376566e-07, |
|
"loss": 0.2098, |
|
"step": 13220 |
|
}, |
|
{ |
|
"epoch": 9.41, |
|
"learning_rate": 4.2060275559840377e-07, |
|
"loss": 0.2086, |
|
"step": 13240 |
|
}, |
|
{ |
|
"epoch": 9.43, |
|
"learning_rate": 4.0039003437782055e-07, |
|
"loss": 0.2082, |
|
"step": 13260 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 3.80671104447497e-07, |
|
"loss": 0.2104, |
|
"step": 13280 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 3.61446361628176e-07, |
|
"loss": 0.2075, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 9.47, |
|
"learning_rate": 3.427161918207106e-07, |
|
"loss": 0.2133, |
|
"step": 13320 |
|
}, |
|
{ |
|
"epoch": 9.48, |
|
"learning_rate": 3.2448097099833095e-07, |
|
"loss": 0.2087, |
|
"step": 13340 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"learning_rate": 3.0674106519908155e-07, |
|
"loss": 0.2049, |
|
"step": 13360 |
|
}, |
|
{ |
|
"epoch": 9.51, |
|
"learning_rate": 2.8949683051848754e-07, |
|
"loss": 0.2087, |
|
"step": 13380 |
|
}, |
|
{ |
|
"epoch": 9.53, |
|
"learning_rate": 2.727486131023971e-07, |
|
"loss": 0.2017, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 9.54, |
|
"learning_rate": 2.564967491400394e-07, |
|
"loss": 0.2105, |
|
"step": 13420 |
|
}, |
|
{ |
|
"epoch": 9.56, |
|
"learning_rate": 2.4074156485727197e-07, |
|
"loss": 0.2055, |
|
"step": 13440 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 2.2548337651003837e-07, |
|
"loss": 0.213, |
|
"step": 13460 |
|
}, |
|
{ |
|
"epoch": 9.58, |
|
"learning_rate": 2.1072249037800418e-07, |
|
"loss": 0.2072, |
|
"step": 13480 |
|
}, |
|
{ |
|
"epoch": 9.6, |
|
"learning_rate": 1.9645920275843943e-07, |
|
"loss": 0.2115, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"learning_rate": 1.8269379996023183e-07, |
|
"loss": 0.2093, |
|
"step": 13520 |
|
}, |
|
{ |
|
"epoch": 9.63, |
|
"learning_rate": 1.6942655829817189e-07, |
|
"loss": 0.2134, |
|
"step": 13540 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"learning_rate": 1.566577440873962e-07, |
|
"loss": 0.2098, |
|
"step": 13560 |
|
}, |
|
{ |
|
"epoch": 9.66, |
|
"learning_rate": 1.4438761363803067e-07, |
|
"loss": 0.209, |
|
"step": 13580 |
|
}, |
|
{ |
|
"epoch": 9.67, |
|
"learning_rate": 1.3261641325006124e-07, |
|
"loss": 0.2119, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"learning_rate": 1.213443792083796e-07, |
|
"loss": 0.2116, |
|
"step": 13620 |
|
}, |
|
{ |
|
"epoch": 9.7, |
|
"learning_rate": 1.1057173777804797e-07, |
|
"loss": 0.2091, |
|
"step": 13640 |
|
}, |
|
{ |
|
"epoch": 9.71, |
|
"learning_rate": 1.0029870519975004e-07, |
|
"loss": 0.2074, |
|
"step": 13660 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"learning_rate": 9.052548768545832e-08, |
|
"loss": 0.2092, |
|
"step": 13680 |
|
}, |
|
{ |
|
"epoch": 9.74, |
|
"learning_rate": 8.125228141428465e-08, |
|
"loss": 0.2043, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 9.75, |
|
"learning_rate": 7.247927252854725e-08, |
|
"loss": 0.215, |
|
"step": 13720 |
|
}, |
|
{ |
|
"epoch": 9.77, |
|
"learning_rate": 6.420663713004038e-08, |
|
"loss": 0.2143, |
|
"step": 13740 |
|
}, |
|
{ |
|
"epoch": 9.78, |
|
"learning_rate": 5.643454127648995e-08, |
|
"loss": 0.2043, |
|
"step": 13760 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"learning_rate": 4.9163140978225605e-08, |
|
"loss": 0.2103, |
|
"step": 13780 |
|
}, |
|
{ |
|
"epoch": 9.81, |
|
"learning_rate": 4.239258219504716e-08, |
|
"loss": 0.2062, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 9.83, |
|
"learning_rate": 3.612300083329079e-08, |
|
"loss": 0.2076, |
|
"step": 13820 |
|
}, |
|
{ |
|
"epoch": 9.84, |
|
"learning_rate": 3.035452274311457e-08, |
|
"loss": 0.2104, |
|
"step": 13840 |
|
}, |
|
{ |
|
"epoch": 9.85, |
|
"learning_rate": 2.5087263715953268e-08, |
|
"loss": 0.2117, |
|
"step": 13860 |
|
}, |
|
{ |
|
"epoch": 9.87, |
|
"learning_rate": 2.0321329482209107e-08, |
|
"loss": 0.206, |
|
"step": 13880 |
|
}, |
|
{ |
|
"epoch": 9.88, |
|
"learning_rate": 1.605681570912565e-08, |
|
"loss": 0.2169, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"learning_rate": 1.2293807998858819e-08, |
|
"loss": 0.2124, |
|
"step": 13920 |
|
}, |
|
{ |
|
"epoch": 9.91, |
|
"learning_rate": 9.03238188677269e-09, |
|
"loss": 0.2097, |
|
"step": 13940 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"learning_rate": 6.272602839915709e-09, |
|
"loss": 0.2076, |
|
"step": 13960 |
|
}, |
|
{ |
|
"epoch": 9.94, |
|
"learning_rate": 4.014526255702311e-09, |
|
"loss": 0.2033, |
|
"step": 13980 |
|
}, |
|
{ |
|
"epoch": 9.95, |
|
"learning_rate": 2.2581974608082425e-09, |
|
"loss": 0.2207, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"learning_rate": 1.0036517102601784e-09, |
|
"loss": 0.2111, |
|
"step": 14020 |
|
}, |
|
{ |
|
"epoch": 9.98, |
|
"learning_rate": 2.509141867224063e-10, |
|
"loss": 0.2116, |
|
"step": 14040 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.2079, |
|
"step": 14060 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 0.48169249296188354, |
|
"eval_runtime": 171.0447, |
|
"eval_samples_per_second": 32.524, |
|
"eval_steps_per_second": 8.132, |
|
"step": 14060 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"step": 14060, |
|
"total_flos": 4.511702957167411e+18, |
|
"train_loss": 0.43939053083042673, |
|
"train_runtime": 38556.6993, |
|
"train_samples_per_second": 11.673, |
|
"train_steps_per_second": 0.365 |
|
} |
|
], |
|
"logging_steps": 20, |
|
"max_steps": 14060, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"total_flos": 4.511702957167411e+18, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|