|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 392, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.002551020408163265, |
|
"grad_norm": 14000.175878727785, |
|
"learning_rate": 7.499999999999999e-06, |
|
"loss": 48.0695, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.012755102040816327, |
|
"grad_norm": 3007.3104608903254, |
|
"learning_rate": 3.75e-05, |
|
"loss": 32.7037, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.025510204081632654, |
|
"grad_norm": 4753.9239460531635, |
|
"learning_rate": 7.5e-05, |
|
"loss": 27.6128, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03826530612244898, |
|
"grad_norm": 7672.954437161885, |
|
"learning_rate": 0.0001125, |
|
"loss": 118.424, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.05102040816326531, |
|
"grad_norm": 2658.2281406370403, |
|
"learning_rate": 0.00015, |
|
"loss": 36.7995, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.06377551020408163, |
|
"grad_norm": 301.82927722978826, |
|
"learning_rate": 0.00018749999999999998, |
|
"loss": 25.8987, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.07653061224489796, |
|
"grad_norm": 261.2743806851815, |
|
"learning_rate": 0.000225, |
|
"loss": 19.7068, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.08928571428571429, |
|
"grad_norm": 170.80865284544745, |
|
"learning_rate": 0.0002625, |
|
"loss": 16.2639, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.10204081632653061, |
|
"grad_norm": 371.9839372489695, |
|
"learning_rate": 0.0003, |
|
"loss": 14.8666, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.11479591836734694, |
|
"grad_norm": 55.46043538057413, |
|
"learning_rate": 0.00029985067118498503, |
|
"loss": 13.883, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.12755102040816327, |
|
"grad_norm": 61.65936269712346, |
|
"learning_rate": 0.00029940298206120687, |
|
"loss": 11.3142, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.14030612244897958, |
|
"grad_norm": 49.78316202658068, |
|
"learning_rate": 0.0002986578240004834, |
|
"loss": 10.2446, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.15306122448979592, |
|
"grad_norm": 95.22883433955214, |
|
"learning_rate": 0.0002976166806504174, |
|
"loss": 9.2125, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.16581632653061223, |
|
"grad_norm": 87.6593544514716, |
|
"learning_rate": 0.00029628162498037856, |
|
"loss": 8.0024, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.17857142857142858, |
|
"grad_norm": 29.457366532379673, |
|
"learning_rate": 0.000294655315154116, |
|
"loss": 7.0794, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.1913265306122449, |
|
"grad_norm": 23.54866864598235, |
|
"learning_rate": 0.0002927409892372191, |
|
"loss": 6.428, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.20408163265306123, |
|
"grad_norm": 14.487277559435018, |
|
"learning_rate": 0.00029054245874996426, |
|
"loss": 6.0644, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.21683673469387754, |
|
"grad_norm": 11.612016698553127, |
|
"learning_rate": 0.00028806410107838377, |
|
"loss": 5.9737, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.22959183673469388, |
|
"grad_norm": 22.24847439069348, |
|
"learning_rate": 0.00028531085075866815, |
|
"loss": 5.6293, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.2423469387755102, |
|
"grad_norm": 25.32273350126865, |
|
"learning_rate": 0.0002822881896522532, |
|
"loss": 5.4368, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.25510204081632654, |
|
"grad_norm": 29.73658242449881, |
|
"learning_rate": 0.000279002136031155, |
|
"loss": 5.2007, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.26785714285714285, |
|
"grad_norm": 17.009329865460675, |
|
"learning_rate": 0.0002754592325952845, |
|
"loss": 4.9784, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.28061224489795916, |
|
"grad_norm": 17.65449666930728, |
|
"learning_rate": 0.00027166653344559883, |
|
"loss": 5.1761, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.29336734693877553, |
|
"grad_norm": 25.60749450078011, |
|
"learning_rate": 0.0002676315900390278, |
|
"loss": 4.981, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.30612244897959184, |
|
"grad_norm": 19.37666875758449, |
|
"learning_rate": 0.00026336243615313873, |
|
"loss": 4.883, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.31887755102040816, |
|
"grad_norm": 25.223445103235044, |
|
"learning_rate": 0.00025886757189047695, |
|
"loss": 4.6906, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.33163265306122447, |
|
"grad_norm": 24.827221258799884, |
|
"learning_rate": 0.0002541559467544291, |
|
"loss": 4.5962, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.34438775510204084, |
|
"grad_norm": 19.172749908950284, |
|
"learning_rate": 0.00024923694183030657, |
|
"loss": 4.4971, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.35714285714285715, |
|
"grad_norm": 17.803290735515485, |
|
"learning_rate": 0.0002441203511071278, |
|
"loss": 4.3923, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.36989795918367346, |
|
"grad_norm": 15.152588588737128, |
|
"learning_rate": 0.00023881636197728743, |
|
"loss": 4.2507, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.3826530612244898, |
|
"grad_norm": 28.266308285286772, |
|
"learning_rate": 0.0002333355349529403, |
|
"loss": 4.1071, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.39540816326530615, |
|
"grad_norm": 13.058293075351123, |
|
"learning_rate": 0.00022768878263948445, |
|
"loss": 4.0515, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.40816326530612246, |
|
"grad_norm": 14.87484760188348, |
|
"learning_rate": 0.00022188734800800852, |
|
"loss": 3.9059, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.42091836734693877, |
|
"grad_norm": 12.57655967452742, |
|
"learning_rate": 0.00021594278200996457, |
|
"loss": 3.876, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.4336734693877551, |
|
"grad_norm": 18.644576504823885, |
|
"learning_rate": 0.00020986692057863607, |
|
"loss": 3.703, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.44642857142857145, |
|
"grad_norm": 6.3705323701789265, |
|
"learning_rate": 0.000203671861063193, |
|
"loss": 3.6606, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.45918367346938777, |
|
"grad_norm": 9.945740907460738, |
|
"learning_rate": 0.00019736993814225374, |
|
"loss": 3.6781, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.4719387755102041, |
|
"grad_norm": 5.28597104720817, |
|
"learning_rate": 0.00019097369926491297, |
|
"loss": 3.5963, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.4846938775510204, |
|
"grad_norm": 5.570474507029548, |
|
"learning_rate": 0.00018449587966813208, |
|
"loss": 3.5206, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.49744897959183676, |
|
"grad_norm": 5.769299754282368, |
|
"learning_rate": 0.00017794937702023467, |
|
"loss": 3.3751, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.5102040816326531, |
|
"grad_norm": 7.514106159557929, |
|
"learning_rate": 0.00017134722574099276, |
|
"loss": 3.3315, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.5229591836734694, |
|
"grad_norm": 5.467040990097222, |
|
"learning_rate": 0.0001647025710494341, |
|
"loss": 3.2297, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.5357142857142857, |
|
"grad_norm": 7.752282985206345, |
|
"learning_rate": 0.00015802864279104223, |
|
"loss": 3.1714, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.548469387755102, |
|
"grad_norm": 4.902568385413757, |
|
"learning_rate": 0.0001513387290964616, |
|
"loss": 3.076, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.5612244897959183, |
|
"grad_norm": 8.809433652257214, |
|
"learning_rate": 0.00014464614992415294, |
|
"loss": 3.0378, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.5739795918367347, |
|
"grad_norm": 5.62075198192736, |
|
"learning_rate": 0.00013796423053967887, |
|
"loss": 2.946, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.5867346938775511, |
|
"grad_norm": 7.091999165572766, |
|
"learning_rate": 0.0001313062749844221, |
|
"loss": 2.9117, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.5994897959183674, |
|
"grad_norm": 9.664621289964327, |
|
"learning_rate": 0.00012468553958656264, |
|
"loss": 2.7779, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.6122448979591837, |
|
"grad_norm": 8.079291607422778, |
|
"learning_rate": 0.00011811520656705348, |
|
"loss": 2.7627, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.625, |
|
"grad_norm": 9.132106023308898, |
|
"learning_rate": 0.00011160835779314889, |
|
"loss": 2.7464, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.6377551020408163, |
|
"grad_norm": 12.688497775305251, |
|
"learning_rate": 0.00010517794873174064, |
|
"loss": 2.6784, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.6505102040816326, |
|
"grad_norm": 7.508821446542865, |
|
"learning_rate": 9.883678265436473e-05, |
|
"loss": 2.6542, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.6632653061224489, |
|
"grad_norm": 7.288150995036509, |
|
"learning_rate": 9.259748514523653e-05, |
|
"loss": 2.5675, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.6760204081632653, |
|
"grad_norm": 2.598306498532322, |
|
"learning_rate": 8.647247896307018e-05, |
|
"loss": 2.5179, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.6887755102040817, |
|
"grad_norm": 2.9340097554582947, |
|
"learning_rate": 8.047395930673417e-05, |
|
"loss": 2.4673, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.701530612244898, |
|
"grad_norm": 3.520741961544552, |
|
"learning_rate": 7.46138695339903e-05, |
|
"loss": 2.5285, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.7142857142857143, |
|
"grad_norm": 4.635037745202294, |
|
"learning_rate": 6.890387738166041e-05, |
|
"loss": 2.3872, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.7270408163265306, |
|
"grad_norm": 3.987422126482862, |
|
"learning_rate": 6.335535173456914e-05, |
|
"loss": 2.3409, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.7397959183673469, |
|
"grad_norm": 3.0549621746230122, |
|
"learning_rate": 5.79793399895161e-05, |
|
"loss": 2.3254, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.7525510204081632, |
|
"grad_norm": 2.527266957929868, |
|
"learning_rate": 5.278654605934644e-05, |
|
"loss": 2.3429, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.7653061224489796, |
|
"grad_norm": 3.9761126497464083, |
|
"learning_rate": 4.778730906091632e-05, |
|
"loss": 2.2348, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.7780612244897959, |
|
"grad_norm": 2.498797778043332, |
|
"learning_rate": 4.2991582729385174e-05, |
|
"loss": 2.2421, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.7908163265306123, |
|
"grad_norm": 2.5462712929101894, |
|
"learning_rate": 3.840891559982256e-05, |
|
"loss": 2.2751, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.8035714285714286, |
|
"grad_norm": 2.307072702085812, |
|
"learning_rate": 3.404843199558945e-05, |
|
"loss": 2.1867, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.8163265306122449, |
|
"grad_norm": 5.263562323908358, |
|
"learning_rate": 2.9918813861345952e-05, |
|
"loss": 2.181, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.8290816326530612, |
|
"grad_norm": 2.0011040214883575, |
|
"learning_rate": 2.6028283476858038e-05, |
|
"loss": 2.1732, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.8418367346938775, |
|
"grad_norm": 2.2471501177416116, |
|
"learning_rate": 2.238458708602039e-05, |
|
"loss": 2.1108, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.8545918367346939, |
|
"grad_norm": 1.66846134386125, |
|
"learning_rate": 1.8994979473690537e-05, |
|
"loss": 2.1007, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.8673469387755102, |
|
"grad_norm": 2.0450866177774976, |
|
"learning_rate": 1.5866209521043304e-05, |
|
"loss": 2.0334, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.8801020408163265, |
|
"grad_norm": 2.1608778408402807, |
|
"learning_rate": 1.3004506768205226e-05, |
|
"loss": 2.0735, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.8928571428571429, |
|
"grad_norm": 2.1504307790627912, |
|
"learning_rate": 1.0415569010922963e-05, |
|
"loss": 2.0473, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.9056122448979592, |
|
"grad_norm": 1.6880900496760922, |
|
"learning_rate": 8.104550955962469e-06, |
|
"loss": 2.029, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.9183673469387755, |
|
"grad_norm": 1.7900392749998477, |
|
"learning_rate": 6.076053957825411e-06, |
|
"loss": 2.0533, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.9311224489795918, |
|
"grad_norm": 1.5178075882516209, |
|
"learning_rate": 4.334116857218317e-06, |
|
"loss": 1.9693, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.9438775510204082, |
|
"grad_norm": 1.6874487405997367, |
|
"learning_rate": 2.882207939515435e-06, |
|
"loss": 1.9791, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.9566326530612245, |
|
"grad_norm": 1.4244094850155935, |
|
"learning_rate": 1.7232180292259369e-06, |
|
"loss": 1.9566, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.9693877551020408, |
|
"grad_norm": 1.302432101303747, |
|
"learning_rate": 8.594547342153979e-07, |
|
"loss": 1.9451, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.9821428571428571, |
|
"grad_norm": 1.2049628347309629, |
|
"learning_rate": 2.926378511411198e-07, |
|
"loss": 1.9952, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.9948979591836735, |
|
"grad_norm": 1.2981702353996074, |
|
"learning_rate": 2.3895941249507665e-08, |
|
"loss": 1.9687, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 6.496990203857422, |
|
"eval_runtime": 2.5179, |
|
"eval_samples_per_second": 3.972, |
|
"eval_steps_per_second": 0.397, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 392, |
|
"total_flos": 13478110691328.0, |
|
"train_loss": 7.205524507833987, |
|
"train_runtime": 9216.1272, |
|
"train_samples_per_second": 1.36, |
|
"train_steps_per_second": 0.043 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 392, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 13478110691328.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|