|
{ |
|
"best_metric": 0.0013208846794441342, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-25", |
|
"epoch": 0.5988023952095808, |
|
"eval_steps": 5, |
|
"global_step": 25, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.023952095808383235, |
|
"grad_norm": 3.7283265590667725, |
|
"learning_rate": 2e-05, |
|
"loss": 0.8484, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.023952095808383235, |
|
"eval_loss": 0.9376181364059448, |
|
"eval_runtime": 2.5188, |
|
"eval_samples_per_second": 7.146, |
|
"eval_steps_per_second": 3.573, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.04790419161676647, |
|
"grad_norm": 3.78507137298584, |
|
"learning_rate": 4e-05, |
|
"loss": 0.8896, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0718562874251497, |
|
"grad_norm": 3.4431090354919434, |
|
"learning_rate": 6e-05, |
|
"loss": 0.7761, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.09580838323353294, |
|
"grad_norm": 4.195216655731201, |
|
"learning_rate": 8e-05, |
|
"loss": 0.9114, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.11976047904191617, |
|
"grad_norm": 3.787885904312134, |
|
"learning_rate": 0.0001, |
|
"loss": 0.7343, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.11976047904191617, |
|
"eval_loss": 0.5795788168907166, |
|
"eval_runtime": 1.9599, |
|
"eval_samples_per_second": 9.184, |
|
"eval_steps_per_second": 4.592, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.1437125748502994, |
|
"grad_norm": 3.610590696334839, |
|
"learning_rate": 0.00012, |
|
"loss": 0.5398, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.16766467065868262, |
|
"grad_norm": 3.8623299598693848, |
|
"learning_rate": 0.00014, |
|
"loss": 0.3491, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.19161676646706588, |
|
"grad_norm": 2.895916223526001, |
|
"learning_rate": 0.00016, |
|
"loss": 0.1587, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.2155688622754491, |
|
"grad_norm": 1.3432469367980957, |
|
"learning_rate": 0.00018, |
|
"loss": 0.0762, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.23952095808383234, |
|
"grad_norm": 0.9612584114074707, |
|
"learning_rate": 0.0002, |
|
"loss": 0.0333, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.23952095808383234, |
|
"eval_loss": 0.014482591301202774, |
|
"eval_runtime": 1.9555, |
|
"eval_samples_per_second": 9.205, |
|
"eval_steps_per_second": 4.602, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.2634730538922156, |
|
"grad_norm": 0.4900355339050293, |
|
"learning_rate": 0.00019781476007338058, |
|
"loss": 0.0134, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.2874251497005988, |
|
"grad_norm": 0.32205134630203247, |
|
"learning_rate": 0.0001913545457642601, |
|
"loss": 0.0068, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.31137724550898205, |
|
"grad_norm": 0.15824849903583527, |
|
"learning_rate": 0.00018090169943749476, |
|
"loss": 0.0037, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.33532934131736525, |
|
"grad_norm": 0.07333680987358093, |
|
"learning_rate": 0.00016691306063588583, |
|
"loss": 0.0018, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.3592814371257485, |
|
"grad_norm": 0.07736411690711975, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 0.002, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.3592814371257485, |
|
"eval_loss": 0.0016586449928581715, |
|
"eval_runtime": 1.9794, |
|
"eval_samples_per_second": 9.093, |
|
"eval_steps_per_second": 4.547, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.38323353293413176, |
|
"grad_norm": 0.058378469198942184, |
|
"learning_rate": 0.00013090169943749476, |
|
"loss": 0.0016, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.40718562874251496, |
|
"grad_norm": 0.052952490746974945, |
|
"learning_rate": 0.00011045284632676536, |
|
"loss": 0.0017, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.4311377245508982, |
|
"grad_norm": 0.03747759759426117, |
|
"learning_rate": 8.954715367323468e-05, |
|
"loss": 0.0016, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.4550898203592814, |
|
"grad_norm": 0.04296877980232239, |
|
"learning_rate": 6.909830056250527e-05, |
|
"loss": 0.0016, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.47904191616766467, |
|
"grad_norm": 0.03620421141386032, |
|
"learning_rate": 5.000000000000002e-05, |
|
"loss": 0.0015, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.47904191616766467, |
|
"eval_loss": 0.0013291877694427967, |
|
"eval_runtime": 2.3203, |
|
"eval_samples_per_second": 7.758, |
|
"eval_steps_per_second": 3.879, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.5029940119760479, |
|
"grad_norm": 0.029291512444615364, |
|
"learning_rate": 3.308693936411421e-05, |
|
"loss": 0.0013, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.5269461077844312, |
|
"grad_norm": 0.027527112513780594, |
|
"learning_rate": 1.9098300562505266e-05, |
|
"loss": 0.0012, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.5508982035928144, |
|
"grad_norm": 0.02056039869785309, |
|
"learning_rate": 8.645454235739903e-06, |
|
"loss": 0.001, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.5748502994011976, |
|
"grad_norm": 0.02181430160999298, |
|
"learning_rate": 2.1852399266194314e-06, |
|
"loss": 0.001, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.5988023952095808, |
|
"grad_norm": 0.02083355188369751, |
|
"learning_rate": 0.0, |
|
"loss": 0.001, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.5988023952095808, |
|
"eval_loss": 0.0013208846794441342, |
|
"eval_runtime": 1.9734, |
|
"eval_samples_per_second": 9.121, |
|
"eval_steps_per_second": 4.561, |
|
"step": 25 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 25, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 10, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 2, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 222595409510400.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|