|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 4.99558693733451, |
|
"eval_steps": 500, |
|
"global_step": 1415, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00353045013239188, |
|
"grad_norm": 20.178845075428434, |
|
"learning_rate": 1.1764705882352942e-07, |
|
"loss": 0.7173, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.00706090026478376, |
|
"grad_norm": 21.408372160532007, |
|
"learning_rate": 2.3529411764705883e-07, |
|
"loss": 0.8369, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.01059135039717564, |
|
"grad_norm": 22.339887896461786, |
|
"learning_rate": 3.529411764705883e-07, |
|
"loss": 0.7742, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01412180052956752, |
|
"grad_norm": 20.114111492682788, |
|
"learning_rate": 4.7058823529411767e-07, |
|
"loss": 0.6685, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0176522506619594, |
|
"grad_norm": 19.395831354668825, |
|
"learning_rate": 5.882352941176471e-07, |
|
"loss": 0.724, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.02118270079435128, |
|
"grad_norm": 20.441308282025314, |
|
"learning_rate": 7.058823529411766e-07, |
|
"loss": 0.8051, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.02471315092674316, |
|
"grad_norm": 21.439520175660693, |
|
"learning_rate": 8.235294117647059e-07, |
|
"loss": 0.7853, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.02824360105913504, |
|
"grad_norm": 16.361184069343224, |
|
"learning_rate": 9.411764705882353e-07, |
|
"loss": 0.6029, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.03177405119152692, |
|
"grad_norm": 16.50966475117204, |
|
"learning_rate": 1.0588235294117648e-06, |
|
"loss": 0.607, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.0353045013239188, |
|
"grad_norm": 10.733426578588741, |
|
"learning_rate": 1.1764705882352942e-06, |
|
"loss": 0.5416, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.038834951456310676, |
|
"grad_norm": 10.395447257616585, |
|
"learning_rate": 1.2941176470588237e-06, |
|
"loss": 0.4926, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.04236540158870256, |
|
"grad_norm": 12.789389493634516, |
|
"learning_rate": 1.4117647058823531e-06, |
|
"loss": 0.5638, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.04589585172109444, |
|
"grad_norm": 8.94856882496586, |
|
"learning_rate": 1.5294117647058826e-06, |
|
"loss": 0.459, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.04942630185348632, |
|
"grad_norm": 4.421740571415892, |
|
"learning_rate": 1.6470588235294118e-06, |
|
"loss": 0.322, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.0529567519858782, |
|
"grad_norm": 4.035083034631525, |
|
"learning_rate": 1.7647058823529414e-06, |
|
"loss": 0.3224, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.05648720211827008, |
|
"grad_norm": 4.319827920176901, |
|
"learning_rate": 1.8823529411764707e-06, |
|
"loss": 0.3079, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.06001765225066196, |
|
"grad_norm": 4.9583914837299, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.3102, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.06354810238305383, |
|
"grad_norm": 4.011276537282153, |
|
"learning_rate": 2.1176470588235296e-06, |
|
"loss": 0.2933, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.06707855251544571, |
|
"grad_norm": 3.1423743425131607, |
|
"learning_rate": 2.2352941176470592e-06, |
|
"loss": 0.2419, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.0706090026478376, |
|
"grad_norm": 3.526182340817241, |
|
"learning_rate": 2.3529411764705885e-06, |
|
"loss": 0.2649, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.07413945278022947, |
|
"grad_norm": 2.1822830906532205, |
|
"learning_rate": 2.470588235294118e-06, |
|
"loss": 0.241, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.07766990291262135, |
|
"grad_norm": 3.01235390581002, |
|
"learning_rate": 2.5882352941176473e-06, |
|
"loss": 0.2352, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.08120035304501325, |
|
"grad_norm": 3.34743138297086, |
|
"learning_rate": 2.7058823529411766e-06, |
|
"loss": 0.2196, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.08473080317740513, |
|
"grad_norm": 2.453669230978404, |
|
"learning_rate": 2.8235294117647062e-06, |
|
"loss": 0.2577, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.088261253309797, |
|
"grad_norm": 2.0737233723937765, |
|
"learning_rate": 2.9411764705882355e-06, |
|
"loss": 0.2099, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.09179170344218888, |
|
"grad_norm": 2.3481110465941653, |
|
"learning_rate": 3.058823529411765e-06, |
|
"loss": 0.2302, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.09532215357458076, |
|
"grad_norm": 2.4508501080365086, |
|
"learning_rate": 3.1764705882352943e-06, |
|
"loss": 0.2183, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.09885260370697264, |
|
"grad_norm": 1.9310659307898554, |
|
"learning_rate": 3.2941176470588236e-06, |
|
"loss": 0.22, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.10238305383936452, |
|
"grad_norm": 1.731861214941761, |
|
"learning_rate": 3.4117647058823532e-06, |
|
"loss": 0.1702, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.1059135039717564, |
|
"grad_norm": 1.7963688467168324, |
|
"learning_rate": 3.529411764705883e-06, |
|
"loss": 0.2141, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.10944395410414828, |
|
"grad_norm": 1.8118405787472915, |
|
"learning_rate": 3.6470588235294117e-06, |
|
"loss": 0.2155, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.11297440423654016, |
|
"grad_norm": 2.1536428426895764, |
|
"learning_rate": 3.7647058823529414e-06, |
|
"loss": 0.2348, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.11650485436893204, |
|
"grad_norm": 1.3361162643597526, |
|
"learning_rate": 3.882352941176471e-06, |
|
"loss": 0.211, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.12003530450132392, |
|
"grad_norm": 1.3679923730470394, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.2006, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.1235657546337158, |
|
"grad_norm": 1.6951013269663189, |
|
"learning_rate": 4.11764705882353e-06, |
|
"loss": 0.2173, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.12709620476610767, |
|
"grad_norm": 1.1243675718278554, |
|
"learning_rate": 4.235294117647059e-06, |
|
"loss": 0.21, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.13062665489849956, |
|
"grad_norm": 1.798056949309266, |
|
"learning_rate": 4.352941176470588e-06, |
|
"loss": 0.2151, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.13415710503089143, |
|
"grad_norm": 1.5535634835962755, |
|
"learning_rate": 4.4705882352941184e-06, |
|
"loss": 0.2136, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.13768755516328332, |
|
"grad_norm": 1.3322693781225434, |
|
"learning_rate": 4.588235294117647e-06, |
|
"loss": 0.2016, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.1412180052956752, |
|
"grad_norm": 1.2353366230952958, |
|
"learning_rate": 4.705882352941177e-06, |
|
"loss": 0.2158, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.14474845542806708, |
|
"grad_norm": 1.649116854499638, |
|
"learning_rate": 4.823529411764706e-06, |
|
"loss": 0.2085, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.14827890556045895, |
|
"grad_norm": 1.5327812092407431, |
|
"learning_rate": 4.941176470588236e-06, |
|
"loss": 0.223, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.15180935569285084, |
|
"grad_norm": 1.3654123529518498, |
|
"learning_rate": 5.058823529411765e-06, |
|
"loss": 0.1897, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.1553398058252427, |
|
"grad_norm": 1.2874875896720268, |
|
"learning_rate": 5.176470588235295e-06, |
|
"loss": 0.208, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.1588702559576346, |
|
"grad_norm": 1.5894455663917926, |
|
"learning_rate": 5.294117647058824e-06, |
|
"loss": 0.2114, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.1624007060900265, |
|
"grad_norm": 1.4199861690645073, |
|
"learning_rate": 5.411764705882353e-06, |
|
"loss": 0.2204, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.16593115622241836, |
|
"grad_norm": 1.6111716054466962, |
|
"learning_rate": 5.529411764705883e-06, |
|
"loss": 0.2296, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.16946160635481025, |
|
"grad_norm": 1.4732964348658326, |
|
"learning_rate": 5.6470588235294125e-06, |
|
"loss": 0.2236, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.17299205648720212, |
|
"grad_norm": 1.7125830950347147, |
|
"learning_rate": 5.764705882352941e-06, |
|
"loss": 0.1889, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.176522506619594, |
|
"grad_norm": 1.4473702644438498, |
|
"learning_rate": 5.882352941176471e-06, |
|
"loss": 0.2002, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.18005295675198588, |
|
"grad_norm": 1.5337680099050695, |
|
"learning_rate": 6e-06, |
|
"loss": 0.2162, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.18358340688437777, |
|
"grad_norm": 1.6166188689363534, |
|
"learning_rate": 6.11764705882353e-06, |
|
"loss": 0.1984, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.18711385701676964, |
|
"grad_norm": 1.193587355481407, |
|
"learning_rate": 6.2352941176470595e-06, |
|
"loss": 0.1952, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.19064430714916153, |
|
"grad_norm": 1.150057201671324, |
|
"learning_rate": 6.352941176470589e-06, |
|
"loss": 0.2181, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.1941747572815534, |
|
"grad_norm": 1.6229574548509673, |
|
"learning_rate": 6.470588235294119e-06, |
|
"loss": 0.2064, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.1977052074139453, |
|
"grad_norm": 1.3194440409324901, |
|
"learning_rate": 6.588235294117647e-06, |
|
"loss": 0.2024, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.20123565754633715, |
|
"grad_norm": 1.5072216772819462, |
|
"learning_rate": 6.705882352941176e-06, |
|
"loss": 0.2041, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.20476610767872905, |
|
"grad_norm": 1.2184767982361704, |
|
"learning_rate": 6.8235294117647065e-06, |
|
"loss": 0.2165, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.2082965578111209, |
|
"grad_norm": 1.4524431095544894, |
|
"learning_rate": 6.941176470588236e-06, |
|
"loss": 0.2108, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.2118270079435128, |
|
"grad_norm": 1.132320106750761, |
|
"learning_rate": 7.058823529411766e-06, |
|
"loss": 0.1904, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.21535745807590467, |
|
"grad_norm": 1.215224093914693, |
|
"learning_rate": 7.176470588235295e-06, |
|
"loss": 0.1994, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.21888790820829657, |
|
"grad_norm": 1.0022175847945094, |
|
"learning_rate": 7.294117647058823e-06, |
|
"loss": 0.186, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.22241835834068843, |
|
"grad_norm": 1.3628751075073207, |
|
"learning_rate": 7.4117647058823535e-06, |
|
"loss": 0.1911, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.22594880847308033, |
|
"grad_norm": 1.2505219061147377, |
|
"learning_rate": 7.529411764705883e-06, |
|
"loss": 0.2018, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.2294792586054722, |
|
"grad_norm": 1.3707206838948995, |
|
"learning_rate": 7.647058823529411e-06, |
|
"loss": 0.1997, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.23300970873786409, |
|
"grad_norm": 1.3412107722466498, |
|
"learning_rate": 7.764705882352941e-06, |
|
"loss": 0.2174, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.23654015887025595, |
|
"grad_norm": 1.3129029846369458, |
|
"learning_rate": 7.882352941176471e-06, |
|
"loss": 0.1947, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.24007060900264784, |
|
"grad_norm": 1.6510384082728948, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.1724, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.2436010591350397, |
|
"grad_norm": 1.2774008213531902, |
|
"learning_rate": 8.11764705882353e-06, |
|
"loss": 0.2002, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.2471315092674316, |
|
"grad_norm": 1.3287314560957624, |
|
"learning_rate": 8.23529411764706e-06, |
|
"loss": 0.1928, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.2506619593998235, |
|
"grad_norm": 1.0681415016582574, |
|
"learning_rate": 8.35294117647059e-06, |
|
"loss": 0.1721, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.25419240953221534, |
|
"grad_norm": 1.21518473330371, |
|
"learning_rate": 8.470588235294118e-06, |
|
"loss": 0.1985, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.25772285966460723, |
|
"grad_norm": 1.4756769784911512, |
|
"learning_rate": 8.588235294117647e-06, |
|
"loss": 0.2281, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.2612533097969991, |
|
"grad_norm": 0.8760360831121267, |
|
"learning_rate": 8.705882352941177e-06, |
|
"loss": 0.1737, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.264783759929391, |
|
"grad_norm": 1.336571680306999, |
|
"learning_rate": 8.823529411764707e-06, |
|
"loss": 0.1901, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.26831421006178285, |
|
"grad_norm": 1.4034937627691704, |
|
"learning_rate": 8.941176470588237e-06, |
|
"loss": 0.2238, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.27184466019417475, |
|
"grad_norm": 1.1091052692159387, |
|
"learning_rate": 9.058823529411765e-06, |
|
"loss": 0.2, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.27537511032656664, |
|
"grad_norm": 1.5465224562234414, |
|
"learning_rate": 9.176470588235294e-06, |
|
"loss": 0.192, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.27890556045895853, |
|
"grad_norm": 1.4355050182008238, |
|
"learning_rate": 9.294117647058824e-06, |
|
"loss": 0.1853, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.2824360105913504, |
|
"grad_norm": 1.6148908553292696, |
|
"learning_rate": 9.411764705882354e-06, |
|
"loss": 0.1944, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.28596646072374227, |
|
"grad_norm": 1.006979010091179, |
|
"learning_rate": 9.529411764705882e-06, |
|
"loss": 0.1842, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.28949691085613416, |
|
"grad_norm": 1.2631451183457176, |
|
"learning_rate": 9.647058823529412e-06, |
|
"loss": 0.1889, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.29302736098852605, |
|
"grad_norm": 1.0554142641488224, |
|
"learning_rate": 9.764705882352942e-06, |
|
"loss": 0.1954, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.2965578111209179, |
|
"grad_norm": 1.2871157434331197, |
|
"learning_rate": 9.882352941176472e-06, |
|
"loss": 0.1783, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.3000882612533098, |
|
"grad_norm": 1.8307918651683295, |
|
"learning_rate": 1e-05, |
|
"loss": 0.2095, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.3036187113857017, |
|
"grad_norm": 1.27120661506548, |
|
"learning_rate": 9.999996725424495e-06, |
|
"loss": 0.1818, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.30714916151809357, |
|
"grad_norm": 1.3010925765539272, |
|
"learning_rate": 9.999986901702262e-06, |
|
"loss": 0.1967, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.3106796116504854, |
|
"grad_norm": 1.012603025593678, |
|
"learning_rate": 9.999970528846173e-06, |
|
"loss": 0.2084, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.3142100617828773, |
|
"grad_norm": 1.042462350269043, |
|
"learning_rate": 9.99994760687767e-06, |
|
"loss": 0.2113, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.3177405119152692, |
|
"grad_norm": 1.479637631853533, |
|
"learning_rate": 9.999918135826783e-06, |
|
"loss": 0.1923, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.3212709620476611, |
|
"grad_norm": 1.6536152705406355, |
|
"learning_rate": 9.999882115732109e-06, |
|
"loss": 0.1757, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.324801412180053, |
|
"grad_norm": 1.0022195112884287, |
|
"learning_rate": 9.999839546640827e-06, |
|
"loss": 0.2014, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.3283318623124448, |
|
"grad_norm": 1.0752815706731398, |
|
"learning_rate": 9.9997904286087e-06, |
|
"loss": 0.1947, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.3318623124448367, |
|
"grad_norm": 0.9640451554876094, |
|
"learning_rate": 9.999734761700061e-06, |
|
"loss": 0.1947, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.3353927625772286, |
|
"grad_norm": 0.9472529792678961, |
|
"learning_rate": 9.999672545987826e-06, |
|
"loss": 0.1933, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.3389232127096205, |
|
"grad_norm": 0.8262321839814756, |
|
"learning_rate": 9.999603781553487e-06, |
|
"loss": 0.1824, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.34245366284201234, |
|
"grad_norm": 1.302452468053817, |
|
"learning_rate": 9.999528468487113e-06, |
|
"loss": 0.2068, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.34598411297440423, |
|
"grad_norm": 0.903902282892577, |
|
"learning_rate": 9.999446606887349e-06, |
|
"loss": 0.1978, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.34951456310679613, |
|
"grad_norm": 1.0828870366995145, |
|
"learning_rate": 9.999358196861422e-06, |
|
"loss": 0.218, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.353045013239188, |
|
"grad_norm": 1.1262059857251303, |
|
"learning_rate": 9.999263238525135e-06, |
|
"loss": 0.2054, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.35657546337157986, |
|
"grad_norm": 0.9433649602156967, |
|
"learning_rate": 9.999161732002867e-06, |
|
"loss": 0.1811, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.36010591350397175, |
|
"grad_norm": 1.2486258616646937, |
|
"learning_rate": 9.999053677427573e-06, |
|
"loss": 0.2028, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.36363636363636365, |
|
"grad_norm": 1.0344680877191332, |
|
"learning_rate": 9.998939074940788e-06, |
|
"loss": 0.2063, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.36716681376875554, |
|
"grad_norm": 1.3236038216383461, |
|
"learning_rate": 9.99881792469262e-06, |
|
"loss": 0.2014, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.3706972639011474, |
|
"grad_norm": 1.0486814111295473, |
|
"learning_rate": 9.998690226841756e-06, |
|
"loss": 0.2005, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.37422771403353927, |
|
"grad_norm": 0.9052783364228919, |
|
"learning_rate": 9.998555981555459e-06, |
|
"loss": 0.1806, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.37775816416593117, |
|
"grad_norm": 0.8762864385995432, |
|
"learning_rate": 9.998415189009566e-06, |
|
"loss": 0.1963, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.38128861429832306, |
|
"grad_norm": 1.0133330127905669, |
|
"learning_rate": 9.998267849388494e-06, |
|
"loss": 0.2063, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.3848190644307149, |
|
"grad_norm": 1.1032798889585722, |
|
"learning_rate": 9.99811396288523e-06, |
|
"loss": 0.2136, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.3883495145631068, |
|
"grad_norm": 0.9526777003854597, |
|
"learning_rate": 9.99795352970134e-06, |
|
"loss": 0.1997, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.3918799646954987, |
|
"grad_norm": 1.3675080989097932, |
|
"learning_rate": 9.997786550046965e-06, |
|
"loss": 0.1866, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.3954104148278906, |
|
"grad_norm": 1.0304325204839548, |
|
"learning_rate": 9.99761302414082e-06, |
|
"loss": 0.2001, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.3989408649602824, |
|
"grad_norm": 0.8029445134568592, |
|
"learning_rate": 9.997432952210193e-06, |
|
"loss": 0.1806, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.4024713150926743, |
|
"grad_norm": 0.9662432601060303, |
|
"learning_rate": 9.997246334490952e-06, |
|
"loss": 0.1983, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.4060017652250662, |
|
"grad_norm": 0.6650992554809413, |
|
"learning_rate": 9.997053171227527e-06, |
|
"loss": 0.1904, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.4095322153574581, |
|
"grad_norm": 1.2497143912397224, |
|
"learning_rate": 9.996853462672935e-06, |
|
"loss": 0.1921, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.41306266548984993, |
|
"grad_norm": 1.2105700846879488, |
|
"learning_rate": 9.996647209088759e-06, |
|
"loss": 0.1913, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.4165931156222418, |
|
"grad_norm": 0.8745391363159488, |
|
"learning_rate": 9.996434410745158e-06, |
|
"loss": 0.1877, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.4201235657546337, |
|
"grad_norm": 1.155984605743064, |
|
"learning_rate": 9.996215067920855e-06, |
|
"loss": 0.1931, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.4236540158870256, |
|
"grad_norm": 0.9959150019854762, |
|
"learning_rate": 9.99598918090316e-06, |
|
"loss": 0.1901, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.42718446601941745, |
|
"grad_norm": 1.051448796745704, |
|
"learning_rate": 9.995756749987942e-06, |
|
"loss": 0.1642, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.43071491615180935, |
|
"grad_norm": 0.935738059968614, |
|
"learning_rate": 9.995517775479647e-06, |
|
"loss": 0.1702, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.43424536628420124, |
|
"grad_norm": 1.085401876613817, |
|
"learning_rate": 9.995272257691291e-06, |
|
"loss": 0.1992, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.43777581641659313, |
|
"grad_norm": 0.9924411904534596, |
|
"learning_rate": 9.99502019694446e-06, |
|
"loss": 0.1922, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.44130626654898497, |
|
"grad_norm": 1.3346252020931497, |
|
"learning_rate": 9.994761593569312e-06, |
|
"loss": 0.1894, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.44483671668137686, |
|
"grad_norm": 0.996928968088298, |
|
"learning_rate": 9.994496447904573e-06, |
|
"loss": 0.18, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.44836716681376876, |
|
"grad_norm": 1.185034118977359, |
|
"learning_rate": 9.99422476029754e-06, |
|
"loss": 0.2015, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.45189761694616065, |
|
"grad_norm": 1.079465604221569, |
|
"learning_rate": 9.993946531104075e-06, |
|
"loss": 0.1795, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.4554280670785525, |
|
"grad_norm": 1.0665272432739623, |
|
"learning_rate": 9.993661760688611e-06, |
|
"loss": 0.199, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.4589585172109444, |
|
"grad_norm": 1.2364225145584533, |
|
"learning_rate": 9.993370449424153e-06, |
|
"loss": 0.2198, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.4624889673433363, |
|
"grad_norm": 0.755126320981263, |
|
"learning_rate": 9.993072597692266e-06, |
|
"loss": 0.2006, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.46601941747572817, |
|
"grad_norm": 1.3871006955467502, |
|
"learning_rate": 9.992768205883085e-06, |
|
"loss": 0.2037, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.46954986760812, |
|
"grad_norm": 0.7426689064084516, |
|
"learning_rate": 9.992457274395312e-06, |
|
"loss": 0.1774, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.4730803177405119, |
|
"grad_norm": 0.9196839690873588, |
|
"learning_rate": 9.992139803636218e-06, |
|
"loss": 0.1973, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.4766107678729038, |
|
"grad_norm": 0.7786869026927636, |
|
"learning_rate": 9.99181579402163e-06, |
|
"loss": 0.1842, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.4801412180052957, |
|
"grad_norm": 1.03244467425382, |
|
"learning_rate": 9.991485245975949e-06, |
|
"loss": 0.1663, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.4836716681376876, |
|
"grad_norm": 1.0713507973139096, |
|
"learning_rate": 9.991148159932134e-06, |
|
"loss": 0.1941, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.4872021182700794, |
|
"grad_norm": 0.9453234857914946, |
|
"learning_rate": 9.990804536331715e-06, |
|
"loss": 0.2007, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.4907325684024713, |
|
"grad_norm": 0.7607319035095186, |
|
"learning_rate": 9.990454375624778e-06, |
|
"loss": 0.18, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.4942630185348632, |
|
"grad_norm": 0.9915191845505621, |
|
"learning_rate": 9.990097678269972e-06, |
|
"loss": 0.1832, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.4977934686672551, |
|
"grad_norm": 0.6818554079855815, |
|
"learning_rate": 9.989734444734513e-06, |
|
"loss": 0.1779, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.501323918799647, |
|
"grad_norm": 0.8679707616739033, |
|
"learning_rate": 9.989364675494175e-06, |
|
"loss": 0.1585, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.5048543689320388, |
|
"grad_norm": 0.834867345483289, |
|
"learning_rate": 9.988988371033292e-06, |
|
"loss": 0.1684, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.5083848190644307, |
|
"grad_norm": 0.7262565816433312, |
|
"learning_rate": 9.98860553184476e-06, |
|
"loss": 0.1688, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.5119152691968226, |
|
"grad_norm": 0.9922761294466914, |
|
"learning_rate": 9.988216158430033e-06, |
|
"loss": 0.1846, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.5154457193292145, |
|
"grad_norm": 1.0610599926539903, |
|
"learning_rate": 9.987820251299121e-06, |
|
"loss": 0.1965, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.5189761694616064, |
|
"grad_norm": 0.8658381475355275, |
|
"learning_rate": 9.987417810970601e-06, |
|
"loss": 0.1923, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.5225066195939982, |
|
"grad_norm": 1.0687893341105537, |
|
"learning_rate": 9.987008837971595e-06, |
|
"loss": 0.2084, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.5260370697263901, |
|
"grad_norm": 1.0221589345087516, |
|
"learning_rate": 9.986593332837795e-06, |
|
"loss": 0.1825, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.529567519858782, |
|
"grad_norm": 0.9977245435165767, |
|
"learning_rate": 9.986171296113437e-06, |
|
"loss": 0.2014, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.5330979699911739, |
|
"grad_norm": 1.0012549684835264, |
|
"learning_rate": 9.985742728351317e-06, |
|
"loss": 0.1885, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.5366284201235657, |
|
"grad_norm": 1.2108459546620967, |
|
"learning_rate": 9.98530763011279e-06, |
|
"loss": 0.1888, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.5401588702559577, |
|
"grad_norm": 0.849586927833, |
|
"learning_rate": 9.98486600196776e-06, |
|
"loss": 0.1957, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.5436893203883495, |
|
"grad_norm": 0.929219836766702, |
|
"learning_rate": 9.984417844494681e-06, |
|
"loss": 0.2051, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.5472197705207414, |
|
"grad_norm": 0.7551838277587243, |
|
"learning_rate": 9.983963158280568e-06, |
|
"loss": 0.1687, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.5507502206531333, |
|
"grad_norm": 0.8965218043033845, |
|
"learning_rate": 9.983501943920978e-06, |
|
"loss": 0.2092, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.5542806707855251, |
|
"grad_norm": 0.9118641687914303, |
|
"learning_rate": 9.98303420202003e-06, |
|
"loss": 0.1857, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.5578111209179171, |
|
"grad_norm": 1.0557315457012684, |
|
"learning_rate": 9.98255993319038e-06, |
|
"loss": 0.1796, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.5613415710503089, |
|
"grad_norm": 0.9917786126115438, |
|
"learning_rate": 9.982079138053243e-06, |
|
"loss": 0.1797, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.5648720211827007, |
|
"grad_norm": 0.9264989403126399, |
|
"learning_rate": 9.981591817238379e-06, |
|
"loss": 0.1814, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.5684024713150927, |
|
"grad_norm": 1.1038265073069693, |
|
"learning_rate": 9.981097971384094e-06, |
|
"loss": 0.2039, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.5719329214474845, |
|
"grad_norm": 1.1450789849461764, |
|
"learning_rate": 9.980597601137244e-06, |
|
"loss": 0.2262, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.5754633715798765, |
|
"grad_norm": 1.0093808973803213, |
|
"learning_rate": 9.980090707153226e-06, |
|
"loss": 0.1767, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.5789938217122683, |
|
"grad_norm": 0.8305835473538516, |
|
"learning_rate": 9.979577290095988e-06, |
|
"loss": 0.1854, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.5825242718446602, |
|
"grad_norm": 0.8814645905158526, |
|
"learning_rate": 9.97905735063802e-06, |
|
"loss": 0.207, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.5860547219770521, |
|
"grad_norm": 1.1283393561929782, |
|
"learning_rate": 9.978530889460351e-06, |
|
"loss": 0.1482, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.589585172109444, |
|
"grad_norm": 0.8882147790176694, |
|
"learning_rate": 9.977997907252557e-06, |
|
"loss": 0.1778, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.5931156222418358, |
|
"grad_norm": 0.8063636359843489, |
|
"learning_rate": 9.977458404712755e-06, |
|
"loss": 0.1922, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.5966460723742277, |
|
"grad_norm": 0.8023695735241588, |
|
"learning_rate": 9.976912382547603e-06, |
|
"loss": 0.1979, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.6001765225066196, |
|
"grad_norm": 0.8128057874275411, |
|
"learning_rate": 9.976359841472294e-06, |
|
"loss": 0.196, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.6037069726390115, |
|
"grad_norm": 0.8202708710729476, |
|
"learning_rate": 9.975800782210565e-06, |
|
"loss": 0.1829, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.6072374227714034, |
|
"grad_norm": 0.9514477549047732, |
|
"learning_rate": 9.975235205494689e-06, |
|
"loss": 0.218, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.6107678729037952, |
|
"grad_norm": 1.0222112927480378, |
|
"learning_rate": 9.974663112065474e-06, |
|
"loss": 0.1783, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.6142983230361871, |
|
"grad_norm": 0.891305185008567, |
|
"learning_rate": 9.974084502672266e-06, |
|
"loss": 0.1588, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.617828773168579, |
|
"grad_norm": 0.8521407342632297, |
|
"learning_rate": 9.973499378072947e-06, |
|
"loss": 0.1997, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.6213592233009708, |
|
"grad_norm": 0.8261272762704142, |
|
"learning_rate": 9.972907739033927e-06, |
|
"loss": 0.2064, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.6248896734333628, |
|
"grad_norm": 0.9844072163334512, |
|
"learning_rate": 9.972309586330156e-06, |
|
"loss": 0.1934, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.6284201235657546, |
|
"grad_norm": 0.9760362259758513, |
|
"learning_rate": 9.97170492074511e-06, |
|
"loss": 0.1711, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.6319505736981466, |
|
"grad_norm": 0.8171800425370801, |
|
"learning_rate": 9.9710937430708e-06, |
|
"loss": 0.1814, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.6354810238305384, |
|
"grad_norm": 0.7848321856443494, |
|
"learning_rate": 9.970476054107763e-06, |
|
"loss": 0.1688, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.6390114739629302, |
|
"grad_norm": 0.7304874742153669, |
|
"learning_rate": 9.969851854665072e-06, |
|
"loss": 0.1975, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.6425419240953222, |
|
"grad_norm": 0.8122652782447508, |
|
"learning_rate": 9.969221145560315e-06, |
|
"loss": 0.1856, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.646072374227714, |
|
"grad_norm": 0.5663405788505017, |
|
"learning_rate": 9.968583927619618e-06, |
|
"loss": 0.1823, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.649602824360106, |
|
"grad_norm": 0.8631102591025225, |
|
"learning_rate": 9.967940201677628e-06, |
|
"loss": 0.191, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.6531332744924978, |
|
"grad_norm": 0.9219061678650854, |
|
"learning_rate": 9.967289968577515e-06, |
|
"loss": 0.1867, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.6566637246248896, |
|
"grad_norm": 0.8104447482059561, |
|
"learning_rate": 9.966633229170975e-06, |
|
"loss": 0.2085, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.6601941747572816, |
|
"grad_norm": 0.696325768189324, |
|
"learning_rate": 9.965969984318227e-06, |
|
"loss": 0.1955, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.6637246248896734, |
|
"grad_norm": 0.957747594144563, |
|
"learning_rate": 9.965300234888007e-06, |
|
"loss": 0.1973, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.6672550750220653, |
|
"grad_norm": 0.8838107190805117, |
|
"learning_rate": 9.964623981757572e-06, |
|
"loss": 0.1862, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.6707855251544572, |
|
"grad_norm": 1.0086692709079665, |
|
"learning_rate": 9.9639412258127e-06, |
|
"loss": 0.1876, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.6743159752868491, |
|
"grad_norm": 0.7851582403291492, |
|
"learning_rate": 9.963251967947687e-06, |
|
"loss": 0.1925, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.677846425419241, |
|
"grad_norm": 0.9006177750516297, |
|
"learning_rate": 9.962556209065343e-06, |
|
"loss": 0.1651, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.6813768755516328, |
|
"grad_norm": 0.8917643734840932, |
|
"learning_rate": 9.961853950076992e-06, |
|
"loss": 0.164, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.6849073256840247, |
|
"grad_norm": 0.8833175487061256, |
|
"learning_rate": 9.961145191902475e-06, |
|
"loss": 0.1858, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.6884377758164166, |
|
"grad_norm": 0.7703074827448632, |
|
"learning_rate": 9.960429935470146e-06, |
|
"loss": 0.1719, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.6919682259488085, |
|
"grad_norm": 0.978543268699946, |
|
"learning_rate": 9.959708181716869e-06, |
|
"loss": 0.1843, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.6954986760812003, |
|
"grad_norm": 1.0182659329814947, |
|
"learning_rate": 9.958979931588018e-06, |
|
"loss": 0.1698, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.6990291262135923, |
|
"grad_norm": 1.0948049910788071, |
|
"learning_rate": 9.958245186037477e-06, |
|
"loss": 0.1892, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.7025595763459841, |
|
"grad_norm": 0.6933804652704536, |
|
"learning_rate": 9.957503946027638e-06, |
|
"loss": 0.1672, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.706090026478376, |
|
"grad_norm": 0.8077103768832906, |
|
"learning_rate": 9.956756212529402e-06, |
|
"loss": 0.1585, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.7096204766107679, |
|
"grad_norm": 0.822023166461492, |
|
"learning_rate": 9.95600198652217e-06, |
|
"loss": 0.1801, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.7131509267431597, |
|
"grad_norm": 1.1206342687945703, |
|
"learning_rate": 9.955241268993852e-06, |
|
"loss": 0.201, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.7166813768755517, |
|
"grad_norm": 0.974499750770202, |
|
"learning_rate": 9.954474060940856e-06, |
|
"loss": 0.2085, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.7202118270079435, |
|
"grad_norm": 0.6598260944598164, |
|
"learning_rate": 9.953700363368097e-06, |
|
"loss": 0.1859, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.7237422771403353, |
|
"grad_norm": 0.8633415747390508, |
|
"learning_rate": 9.952920177288985e-06, |
|
"loss": 0.1963, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.7272727272727273, |
|
"grad_norm": 0.5898944244131702, |
|
"learning_rate": 9.952133503725435e-06, |
|
"loss": 0.2049, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.7308031774051191, |
|
"grad_norm": 1.2714984543033625, |
|
"learning_rate": 9.951340343707852e-06, |
|
"loss": 0.1837, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.7343336275375111, |
|
"grad_norm": 0.9097980008759828, |
|
"learning_rate": 9.950540698275144e-06, |
|
"loss": 0.1849, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.7378640776699029, |
|
"grad_norm": 0.6941779662092539, |
|
"learning_rate": 9.949734568474708e-06, |
|
"loss": 0.176, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.7413945278022948, |
|
"grad_norm": 1.0789428661342089, |
|
"learning_rate": 9.948921955362438e-06, |
|
"loss": 0.1845, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.7449249779346867, |
|
"grad_norm": 1.002969780293264, |
|
"learning_rate": 9.94810286000272e-06, |
|
"loss": 0.1932, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.7484554280670785, |
|
"grad_norm": 0.7346857479212374, |
|
"learning_rate": 9.94727728346843e-06, |
|
"loss": 0.1895, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.7519858781994704, |
|
"grad_norm": 1.1214956368968594, |
|
"learning_rate": 9.946445226840932e-06, |
|
"loss": 0.1872, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.7555163283318623, |
|
"grad_norm": 0.8098364912886733, |
|
"learning_rate": 9.945606691210081e-06, |
|
"loss": 0.182, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.7590467784642542, |
|
"grad_norm": 0.9351430688337309, |
|
"learning_rate": 9.944761677674214e-06, |
|
"loss": 0.194, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.7625772285966461, |
|
"grad_norm": 1.1657299434860284, |
|
"learning_rate": 9.943910187340154e-06, |
|
"loss": 0.1875, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.766107678729038, |
|
"grad_norm": 1.0056936253752848, |
|
"learning_rate": 9.943052221323212e-06, |
|
"loss": 0.1993, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.7696381288614298, |
|
"grad_norm": 1.018936012562021, |
|
"learning_rate": 9.942187780747177e-06, |
|
"loss": 0.205, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.7731685789938217, |
|
"grad_norm": 0.9704710377674022, |
|
"learning_rate": 9.94131686674432e-06, |
|
"loss": 0.1803, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.7766990291262136, |
|
"grad_norm": 0.844440847315935, |
|
"learning_rate": 9.940439480455386e-06, |
|
"loss": 0.1903, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.7802294792586054, |
|
"grad_norm": 0.6911229516194293, |
|
"learning_rate": 9.939555623029608e-06, |
|
"loss": 0.1648, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.7837599293909974, |
|
"grad_norm": 0.977598516917381, |
|
"learning_rate": 9.938665295624686e-06, |
|
"loss": 0.1775, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.7872903795233892, |
|
"grad_norm": 0.7894216874582392, |
|
"learning_rate": 9.937768499406798e-06, |
|
"loss": 0.1734, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.7908208296557812, |
|
"grad_norm": 0.8720889047054564, |
|
"learning_rate": 9.936865235550595e-06, |
|
"loss": 0.1818, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.794351279788173, |
|
"grad_norm": 0.7210753676381008, |
|
"learning_rate": 9.9359555052392e-06, |
|
"loss": 0.2001, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.7978817299205648, |
|
"grad_norm": 0.6358344164302707, |
|
"learning_rate": 9.935039309664204e-06, |
|
"loss": 0.1703, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.8014121800529568, |
|
"grad_norm": 0.8138407178306513, |
|
"learning_rate": 9.93411665002567e-06, |
|
"loss": 0.172, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.8049426301853486, |
|
"grad_norm": 0.974591399579874, |
|
"learning_rate": 9.93318752753212e-06, |
|
"loss": 0.2067, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.8084730803177406, |
|
"grad_norm": 0.8238739378175247, |
|
"learning_rate": 9.932251943400554e-06, |
|
"loss": 0.1607, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.8120035304501324, |
|
"grad_norm": 1.0422244856011833, |
|
"learning_rate": 9.931309898856423e-06, |
|
"loss": 0.1693, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.8155339805825242, |
|
"grad_norm": 0.812884439844158, |
|
"learning_rate": 9.930361395133647e-06, |
|
"loss": 0.1683, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.8190644307149162, |
|
"grad_norm": 0.6222138282668747, |
|
"learning_rate": 9.929406433474606e-06, |
|
"loss": 0.1636, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.822594880847308, |
|
"grad_norm": 0.8762443957945536, |
|
"learning_rate": 9.928445015130136e-06, |
|
"loss": 0.191, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.8261253309796999, |
|
"grad_norm": 0.9859315198495622, |
|
"learning_rate": 9.927477141359533e-06, |
|
"loss": 0.1895, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.8296557811120918, |
|
"grad_norm": 0.9589349463946001, |
|
"learning_rate": 9.926502813430545e-06, |
|
"loss": 0.197, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.8331862312444837, |
|
"grad_norm": 0.8718214863000249, |
|
"learning_rate": 9.92552203261938e-06, |
|
"loss": 0.1799, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.8367166813768756, |
|
"grad_norm": 1.1335307020440895, |
|
"learning_rate": 9.92453480021069e-06, |
|
"loss": 0.198, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.8402471315092674, |
|
"grad_norm": 0.8430408249858925, |
|
"learning_rate": 9.923541117497586e-06, |
|
"loss": 0.1478, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.8437775816416593, |
|
"grad_norm": 0.6769109520694074, |
|
"learning_rate": 9.922540985781621e-06, |
|
"loss": 0.1923, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.8473080317740512, |
|
"grad_norm": 0.6377271343571659, |
|
"learning_rate": 9.921534406372797e-06, |
|
"loss": 0.1879, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.8508384819064431, |
|
"grad_norm": 0.7271721148140101, |
|
"learning_rate": 9.920521380589566e-06, |
|
"loss": 0.1938, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.8543689320388349, |
|
"grad_norm": 0.9160321910693624, |
|
"learning_rate": 9.919501909758815e-06, |
|
"loss": 0.1689, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.8578993821712269, |
|
"grad_norm": 0.8451022682499405, |
|
"learning_rate": 9.918475995215881e-06, |
|
"loss": 0.1567, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.8614298323036187, |
|
"grad_norm": 0.7383618763118188, |
|
"learning_rate": 9.917443638304536e-06, |
|
"loss": 0.1998, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.8649602824360106, |
|
"grad_norm": 0.9977901978127347, |
|
"learning_rate": 9.916404840376993e-06, |
|
"loss": 0.2184, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.8684907325684025, |
|
"grad_norm": 0.905766448884252, |
|
"learning_rate": 9.915359602793902e-06, |
|
"loss": 0.1822, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.8720211827007943, |
|
"grad_norm": 0.668257498288602, |
|
"learning_rate": 9.914307926924344e-06, |
|
"loss": 0.1755, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.8755516328331863, |
|
"grad_norm": 1.2898306231343724, |
|
"learning_rate": 9.913249814145839e-06, |
|
"loss": 0.1943, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.8790820829655781, |
|
"grad_norm": 0.7801616088781465, |
|
"learning_rate": 9.912185265844333e-06, |
|
"loss": 0.1961, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.8826125330979699, |
|
"grad_norm": 0.8499032108492505, |
|
"learning_rate": 9.911114283414204e-06, |
|
"loss": 0.2022, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.8861429832303619, |
|
"grad_norm": 0.9986716197227599, |
|
"learning_rate": 9.910036868258256e-06, |
|
"loss": 0.1935, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.8896734333627537, |
|
"grad_norm": 0.8164649483937467, |
|
"learning_rate": 9.908953021787725e-06, |
|
"loss": 0.1772, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.8932038834951457, |
|
"grad_norm": 0.6585245832704895, |
|
"learning_rate": 9.907862745422258e-06, |
|
"loss": 0.1762, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.8967343336275375, |
|
"grad_norm": 0.7669263548180228, |
|
"learning_rate": 9.906766040589937e-06, |
|
"loss": 0.1673, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.9002647837599294, |
|
"grad_norm": 0.8311965531630117, |
|
"learning_rate": 9.905662908727258e-06, |
|
"loss": 0.1657, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.9037952338923213, |
|
"grad_norm": 0.6844670189568429, |
|
"learning_rate": 9.904553351279139e-06, |
|
"loss": 0.1594, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.9073256840247131, |
|
"grad_norm": 0.8500217928271374, |
|
"learning_rate": 9.903437369698906e-06, |
|
"loss": 0.2076, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.910856134157105, |
|
"grad_norm": 0.8553603507750409, |
|
"learning_rate": 9.902314965448309e-06, |
|
"loss": 0.1838, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.9143865842894969, |
|
"grad_norm": 0.8770431320611423, |
|
"learning_rate": 9.901186139997507e-06, |
|
"loss": 0.1935, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.9179170344218888, |
|
"grad_norm": 0.7281161731708585, |
|
"learning_rate": 9.90005089482507e-06, |
|
"loss": 0.1888, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.9214474845542807, |
|
"grad_norm": 0.6456235823671523, |
|
"learning_rate": 9.898909231417976e-06, |
|
"loss": 0.1764, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.9249779346866726, |
|
"grad_norm": 0.6768287107038422, |
|
"learning_rate": 9.89776115127161e-06, |
|
"loss": 0.163, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.9285083848190644, |
|
"grad_norm": 0.834905164466666, |
|
"learning_rate": 9.89660665588976e-06, |
|
"loss": 0.1842, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.9320388349514563, |
|
"grad_norm": 0.6460482257383012, |
|
"learning_rate": 9.895445746784623e-06, |
|
"loss": 0.165, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.9355692850838482, |
|
"grad_norm": 0.9772687205581816, |
|
"learning_rate": 9.89427842547679e-06, |
|
"loss": 0.2003, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.93909973521624, |
|
"grad_norm": 0.7686896072265816, |
|
"learning_rate": 9.893104693495255e-06, |
|
"loss": 0.1753, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.942630185348632, |
|
"grad_norm": 0.7667904026391998, |
|
"learning_rate": 9.891924552377405e-06, |
|
"loss": 0.1802, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.9461606354810238, |
|
"grad_norm": 0.6993230964449527, |
|
"learning_rate": 9.890738003669029e-06, |
|
"loss": 0.1701, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.9496910856134158, |
|
"grad_norm": 0.7632215639559887, |
|
"learning_rate": 9.8895450489243e-06, |
|
"loss": 0.1607, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.9532215357458076, |
|
"grad_norm": 0.8307898449698382, |
|
"learning_rate": 9.88834568970579e-06, |
|
"loss": 0.1631, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.9567519858781994, |
|
"grad_norm": 0.7176869974666727, |
|
"learning_rate": 9.887139927584453e-06, |
|
"loss": 0.1665, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.9602824360105914, |
|
"grad_norm": 0.924669948462229, |
|
"learning_rate": 9.885927764139633e-06, |
|
"loss": 0.2065, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.9638128861429832, |
|
"grad_norm": 0.7237412872405012, |
|
"learning_rate": 9.88470920095906e-06, |
|
"loss": 0.1802, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.9673433362753752, |
|
"grad_norm": 0.6998624857232224, |
|
"learning_rate": 9.883484239638842e-06, |
|
"loss": 0.1642, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.970873786407767, |
|
"grad_norm": 0.8569858798701728, |
|
"learning_rate": 9.882252881783472e-06, |
|
"loss": 0.1956, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.9744042365401588, |
|
"grad_norm": 0.7621961932465209, |
|
"learning_rate": 9.881015129005822e-06, |
|
"loss": 0.1841, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.9779346866725508, |
|
"grad_norm": 0.7592469083950414, |
|
"learning_rate": 9.879770982927135e-06, |
|
"loss": 0.1937, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.9814651368049426, |
|
"grad_norm": 0.6865266207305956, |
|
"learning_rate": 9.87852044517703e-06, |
|
"loss": 0.1957, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.9849955869373345, |
|
"grad_norm": 0.7228023197310592, |
|
"learning_rate": 9.877263517393503e-06, |
|
"loss": 0.1786, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.9885260370697264, |
|
"grad_norm": 0.9465468675591545, |
|
"learning_rate": 9.876000201222912e-06, |
|
"loss": 0.1661, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.9920564872021183, |
|
"grad_norm": 0.8521304202598817, |
|
"learning_rate": 9.87473049831999e-06, |
|
"loss": 0.154, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.9955869373345102, |
|
"grad_norm": 0.669879136764297, |
|
"learning_rate": 9.87345441034783e-06, |
|
"loss": 0.1624, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.999117387466902, |
|
"grad_norm": 0.9684619377618936, |
|
"learning_rate": 9.872171938977895e-06, |
|
"loss": 0.1881, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 1.002647837599294, |
|
"grad_norm": 0.85439046023442, |
|
"learning_rate": 9.870883085889997e-06, |
|
"loss": 0.1865, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 1.0061782877316858, |
|
"grad_norm": 0.8191806695595788, |
|
"learning_rate": 9.86958785277232e-06, |
|
"loss": 0.163, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 1.0097087378640777, |
|
"grad_norm": 0.9382251565209051, |
|
"learning_rate": 9.8682862413214e-06, |
|
"loss": 0.1758, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 1.0132391879964695, |
|
"grad_norm": 0.5193721161505495, |
|
"learning_rate": 9.866978253242123e-06, |
|
"loss": 0.1751, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 1.0167696381288613, |
|
"grad_norm": 0.8198070617026992, |
|
"learning_rate": 9.865663890247735e-06, |
|
"loss": 0.1535, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 1.0203000882612534, |
|
"grad_norm": 0.9683137087655609, |
|
"learning_rate": 9.864343154059825e-06, |
|
"loss": 0.193, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 1.0238305383936452, |
|
"grad_norm": 1.0396575123509755, |
|
"learning_rate": 9.863016046408336e-06, |
|
"loss": 0.1584, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.027360988526037, |
|
"grad_norm": 0.8171505980553876, |
|
"learning_rate": 9.861682569031553e-06, |
|
"loss": 0.1681, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 1.030891438658429, |
|
"grad_norm": 0.6927072039886698, |
|
"learning_rate": 9.860342723676105e-06, |
|
"loss": 0.1841, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 1.0344218887908208, |
|
"grad_norm": 0.9193349315868659, |
|
"learning_rate": 9.85899651209696e-06, |
|
"loss": 0.1786, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 1.0379523389232128, |
|
"grad_norm": 0.6395800368416922, |
|
"learning_rate": 9.857643936057431e-06, |
|
"loss": 0.1714, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 1.0414827890556047, |
|
"grad_norm": 0.7076027666270328, |
|
"learning_rate": 9.856284997329158e-06, |
|
"loss": 0.1628, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 1.0450132391879965, |
|
"grad_norm": 0.5571080103911318, |
|
"learning_rate": 9.854919697692123e-06, |
|
"loss": 0.1765, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 1.0485436893203883, |
|
"grad_norm": 0.8245809567555668, |
|
"learning_rate": 9.853548038934635e-06, |
|
"loss": 0.178, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 1.0520741394527802, |
|
"grad_norm": 1.3356935667806615, |
|
"learning_rate": 9.852170022853335e-06, |
|
"loss": 0.1629, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 1.055604589585172, |
|
"grad_norm": 0.5648163532404421, |
|
"learning_rate": 9.85078565125319e-06, |
|
"loss": 0.1638, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 1.059135039717564, |
|
"grad_norm": 0.8602300175351031, |
|
"learning_rate": 9.849394925947493e-06, |
|
"loss": 0.1801, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.062665489849956, |
|
"grad_norm": 0.8714458340169661, |
|
"learning_rate": 9.847997848757855e-06, |
|
"loss": 0.1494, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 1.0661959399823477, |
|
"grad_norm": 0.7793786634487098, |
|
"learning_rate": 9.846594421514213e-06, |
|
"loss": 0.1627, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 1.0697263901147396, |
|
"grad_norm": 0.7080125846941803, |
|
"learning_rate": 9.845184646054817e-06, |
|
"loss": 0.1776, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 1.0732568402471314, |
|
"grad_norm": 0.6880805275018903, |
|
"learning_rate": 9.843768524226233e-06, |
|
"loss": 0.1597, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 1.0767872903795235, |
|
"grad_norm": 0.7703863699473644, |
|
"learning_rate": 9.842346057883341e-06, |
|
"loss": 0.1603, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 1.0803177405119153, |
|
"grad_norm": 0.833743875181872, |
|
"learning_rate": 9.84091724888933e-06, |
|
"loss": 0.1613, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 1.0838481906443072, |
|
"grad_norm": 0.7125550254627817, |
|
"learning_rate": 9.839482099115696e-06, |
|
"loss": 0.1511, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 1.087378640776699, |
|
"grad_norm": 0.6773769437101177, |
|
"learning_rate": 9.838040610442243e-06, |
|
"loss": 0.1578, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 1.0909090909090908, |
|
"grad_norm": 0.7853675801397739, |
|
"learning_rate": 9.836592784757077e-06, |
|
"loss": 0.1548, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 1.0944395410414829, |
|
"grad_norm": 0.9629392348111224, |
|
"learning_rate": 9.835138623956603e-06, |
|
"loss": 0.1787, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.0979699911738747, |
|
"grad_norm": 1.0276282328498774, |
|
"learning_rate": 9.833678129945524e-06, |
|
"loss": 0.1798, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 1.1015004413062666, |
|
"grad_norm": 0.9134089726849168, |
|
"learning_rate": 9.83221130463684e-06, |
|
"loss": 0.1888, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 1.1050308914386584, |
|
"grad_norm": 0.9917300375319624, |
|
"learning_rate": 9.830738149951843e-06, |
|
"loss": 0.1884, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 1.1085613415710502, |
|
"grad_norm": 0.7089950679554146, |
|
"learning_rate": 9.829258667820114e-06, |
|
"loss": 0.1686, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 1.1120917917034423, |
|
"grad_norm": 0.578692441690528, |
|
"learning_rate": 9.827772860179528e-06, |
|
"loss": 0.1804, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 1.1156222418358341, |
|
"grad_norm": 0.739188688334005, |
|
"learning_rate": 9.826280728976234e-06, |
|
"loss": 0.1699, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 1.119152691968226, |
|
"grad_norm": 1.0203778463290891, |
|
"learning_rate": 9.824782276164678e-06, |
|
"loss": 0.1679, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 1.1226831421006178, |
|
"grad_norm": 0.9103291412358969, |
|
"learning_rate": 9.823277503707571e-06, |
|
"loss": 0.1885, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 1.1262135922330097, |
|
"grad_norm": 1.1240057634668683, |
|
"learning_rate": 9.821766413575915e-06, |
|
"loss": 0.1467, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 1.1297440423654015, |
|
"grad_norm": 0.7591033795680634, |
|
"learning_rate": 9.820249007748978e-06, |
|
"loss": 0.1708, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.1332744924977936, |
|
"grad_norm": 0.6461323484499221, |
|
"learning_rate": 9.818725288214309e-06, |
|
"loss": 0.1576, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 1.1368049426301854, |
|
"grad_norm": 0.8717085794666202, |
|
"learning_rate": 9.817195256967715e-06, |
|
"loss": 0.1846, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 1.1403353927625772, |
|
"grad_norm": 0.5766027007180119, |
|
"learning_rate": 9.815658916013283e-06, |
|
"loss": 0.1502, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 1.143865842894969, |
|
"grad_norm": 0.8409708523916547, |
|
"learning_rate": 9.814116267363355e-06, |
|
"loss": 0.1522, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 1.147396293027361, |
|
"grad_norm": 0.9388226084700922, |
|
"learning_rate": 9.812567313038542e-06, |
|
"loss": 0.1681, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.150926743159753, |
|
"grad_norm": 0.9369552573302047, |
|
"learning_rate": 9.811012055067708e-06, |
|
"loss": 0.1905, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 1.1544571932921448, |
|
"grad_norm": 1.0468936130746385, |
|
"learning_rate": 9.809450495487977e-06, |
|
"loss": 0.1838, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 1.1579876434245366, |
|
"grad_norm": 0.8466637747576687, |
|
"learning_rate": 9.80788263634473e-06, |
|
"loss": 0.1536, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 1.1615180935569285, |
|
"grad_norm": 0.8133260085099403, |
|
"learning_rate": 9.806308479691595e-06, |
|
"loss": 0.185, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 1.1650485436893203, |
|
"grad_norm": 0.634815173007273, |
|
"learning_rate": 9.80472802759045e-06, |
|
"loss": 0.1763, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.1685789938217122, |
|
"grad_norm": 0.7155087062007245, |
|
"learning_rate": 9.803141282111414e-06, |
|
"loss": 0.1689, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 1.1721094439541042, |
|
"grad_norm": 0.918184676277347, |
|
"learning_rate": 9.801548245332863e-06, |
|
"loss": 0.1771, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 1.175639894086496, |
|
"grad_norm": 0.847842933527483, |
|
"learning_rate": 9.799948919341398e-06, |
|
"loss": 0.1982, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 1.179170344218888, |
|
"grad_norm": 0.950371294714106, |
|
"learning_rate": 9.798343306231869e-06, |
|
"loss": 0.1715, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 1.1827007943512797, |
|
"grad_norm": 0.8452751079657315, |
|
"learning_rate": 9.796731408107349e-06, |
|
"loss": 0.1898, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 1.1862312444836718, |
|
"grad_norm": 1.0132828873057207, |
|
"learning_rate": 9.79511322707916e-06, |
|
"loss": 0.1832, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 1.1897616946160636, |
|
"grad_norm": 0.6796613513123204, |
|
"learning_rate": 9.793488765266838e-06, |
|
"loss": 0.1795, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 1.1932921447484555, |
|
"grad_norm": 0.7516255039006127, |
|
"learning_rate": 9.791858024798157e-06, |
|
"loss": 0.1682, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 1.1968225948808473, |
|
"grad_norm": 0.6882391031969943, |
|
"learning_rate": 9.790221007809106e-06, |
|
"loss": 0.158, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 1.2003530450132391, |
|
"grad_norm": 0.7235480772368882, |
|
"learning_rate": 9.788577716443903e-06, |
|
"loss": 0.1685, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.203883495145631, |
|
"grad_norm": 0.8054680221859308, |
|
"learning_rate": 9.786928152854977e-06, |
|
"loss": 0.1679, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 1.207413945278023, |
|
"grad_norm": 0.6537575157321099, |
|
"learning_rate": 9.785272319202978e-06, |
|
"loss": 0.1535, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 1.2109443954104149, |
|
"grad_norm": 0.8631226400216473, |
|
"learning_rate": 9.783610217656766e-06, |
|
"loss": 0.1686, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 1.2144748455428067, |
|
"grad_norm": 0.7074978790315094, |
|
"learning_rate": 9.781941850393412e-06, |
|
"loss": 0.1557, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 1.2180052956751986, |
|
"grad_norm": 0.7508923027783329, |
|
"learning_rate": 9.780267219598196e-06, |
|
"loss": 0.1734, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 1.2215357458075904, |
|
"grad_norm": 0.6041756898423745, |
|
"learning_rate": 9.7785863274646e-06, |
|
"loss": 0.1676, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 1.2250661959399824, |
|
"grad_norm": 0.8916407715059216, |
|
"learning_rate": 9.776899176194302e-06, |
|
"loss": 0.1612, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 1.2285966460723743, |
|
"grad_norm": 0.9080331160273555, |
|
"learning_rate": 9.775205767997188e-06, |
|
"loss": 0.1718, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 1.2321270962047661, |
|
"grad_norm": 0.844342085425393, |
|
"learning_rate": 9.773506105091336e-06, |
|
"loss": 0.1673, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 1.235657546337158, |
|
"grad_norm": 0.7564502855821623, |
|
"learning_rate": 9.771800189703014e-06, |
|
"loss": 0.1535, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.2391879964695498, |
|
"grad_norm": 0.6268844715263006, |
|
"learning_rate": 9.770088024066681e-06, |
|
"loss": 0.147, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 1.2427184466019416, |
|
"grad_norm": 0.7824396074378344, |
|
"learning_rate": 9.768369610424984e-06, |
|
"loss": 0.1865, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 1.2462488967343337, |
|
"grad_norm": 0.7832814173090706, |
|
"learning_rate": 9.766644951028755e-06, |
|
"loss": 0.1792, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 1.2497793468667255, |
|
"grad_norm": 0.9379854009237389, |
|
"learning_rate": 9.764914048137001e-06, |
|
"loss": 0.1657, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 1.2533097969991174, |
|
"grad_norm": 0.8058449874416329, |
|
"learning_rate": 9.763176904016914e-06, |
|
"loss": 0.1718, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 1.2568402471315092, |
|
"grad_norm": 0.8085192046750732, |
|
"learning_rate": 9.761433520943856e-06, |
|
"loss": 0.1776, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 1.2603706972639013, |
|
"grad_norm": 0.9591752385660486, |
|
"learning_rate": 9.759683901201364e-06, |
|
"loss": 0.1979, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 1.2639011473962931, |
|
"grad_norm": 1.0461750891605788, |
|
"learning_rate": 9.757928047081144e-06, |
|
"loss": 0.1926, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 1.267431597528685, |
|
"grad_norm": 0.7954582531988508, |
|
"learning_rate": 9.756165960883063e-06, |
|
"loss": 0.1648, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 1.2709620476610768, |
|
"grad_norm": 0.8863013446887811, |
|
"learning_rate": 9.75439764491516e-06, |
|
"loss": 0.1807, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.2744924977934686, |
|
"grad_norm": 0.8321220073097337, |
|
"learning_rate": 9.752623101493624e-06, |
|
"loss": 0.1771, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 1.2780229479258605, |
|
"grad_norm": 0.8204672244613037, |
|
"learning_rate": 9.750842332942805e-06, |
|
"loss": 0.1669, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 1.2815533980582523, |
|
"grad_norm": 0.8746356076847612, |
|
"learning_rate": 9.749055341595213e-06, |
|
"loss": 0.1511, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 1.2850838481906444, |
|
"grad_norm": 0.7387304926466468, |
|
"learning_rate": 9.747262129791497e-06, |
|
"loss": 0.1696, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 1.2886142983230362, |
|
"grad_norm": 0.6434577859365603, |
|
"learning_rate": 9.745462699880464e-06, |
|
"loss": 0.1617, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 1.292144748455428, |
|
"grad_norm": 0.8994503573558672, |
|
"learning_rate": 9.74365705421906e-06, |
|
"loss": 0.1771, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 1.2956751985878199, |
|
"grad_norm": 0.9104255097758082, |
|
"learning_rate": 9.741845195172376e-06, |
|
"loss": 0.1507, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 1.299205648720212, |
|
"grad_norm": 0.8237793644377707, |
|
"learning_rate": 9.740027125113638e-06, |
|
"loss": 0.1619, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 1.3027360988526038, |
|
"grad_norm": 0.7759651867155161, |
|
"learning_rate": 9.73820284642421e-06, |
|
"loss": 0.1671, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 1.3062665489849956, |
|
"grad_norm": 0.8529971231768904, |
|
"learning_rate": 9.736372361493584e-06, |
|
"loss": 0.1601, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.3097969991173875, |
|
"grad_norm": 0.6232528534187706, |
|
"learning_rate": 9.734535672719391e-06, |
|
"loss": 0.1662, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 1.3133274492497793, |
|
"grad_norm": 0.9946599681832587, |
|
"learning_rate": 9.732692782507376e-06, |
|
"loss": 0.1687, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 1.3168578993821711, |
|
"grad_norm": 0.7623188795800929, |
|
"learning_rate": 9.730843693271413e-06, |
|
"loss": 0.1683, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 1.3203883495145632, |
|
"grad_norm": 0.5696437347911626, |
|
"learning_rate": 9.728988407433497e-06, |
|
"loss": 0.1555, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 1.323918799646955, |
|
"grad_norm": 0.7695474373449546, |
|
"learning_rate": 9.727126927423736e-06, |
|
"loss": 0.1891, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.3274492497793469, |
|
"grad_norm": 0.7640349287167723, |
|
"learning_rate": 9.725259255680352e-06, |
|
"loss": 0.1822, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 1.3309796999117387, |
|
"grad_norm": 0.7909928993460938, |
|
"learning_rate": 9.72338539464968e-06, |
|
"loss": 0.1742, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 1.3345101500441308, |
|
"grad_norm": 1.1113868103845201, |
|
"learning_rate": 9.72150534678616e-06, |
|
"loss": 0.1647, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 1.3380406001765226, |
|
"grad_norm": 1.0428448743119225, |
|
"learning_rate": 9.719619114552332e-06, |
|
"loss": 0.1354, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 1.3415710503089144, |
|
"grad_norm": 0.8715636876334808, |
|
"learning_rate": 9.717726700418842e-06, |
|
"loss": 0.1772, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.3451015004413063, |
|
"grad_norm": 0.878842141186717, |
|
"learning_rate": 9.715828106864434e-06, |
|
"loss": 0.1657, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 1.3486319505736981, |
|
"grad_norm": 0.6697201435871655, |
|
"learning_rate": 9.713923336375936e-06, |
|
"loss": 0.1585, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 1.35216240070609, |
|
"grad_norm": 0.9696617071339634, |
|
"learning_rate": 9.712012391448283e-06, |
|
"loss": 0.1939, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 1.3556928508384818, |
|
"grad_norm": 0.8187281006125846, |
|
"learning_rate": 9.710095274584482e-06, |
|
"loss": 0.1653, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 1.3592233009708738, |
|
"grad_norm": 0.8974963672760499, |
|
"learning_rate": 9.70817198829563e-06, |
|
"loss": 0.1737, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 1.3627537511032657, |
|
"grad_norm": 0.7456778356448868, |
|
"learning_rate": 9.70624253510091e-06, |
|
"loss": 0.1635, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 1.3662842012356575, |
|
"grad_norm": 1.1629774820732057, |
|
"learning_rate": 9.704306917527575e-06, |
|
"loss": 0.2045, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 1.3698146513680494, |
|
"grad_norm": 0.8069603325247644, |
|
"learning_rate": 9.702365138110958e-06, |
|
"loss": 0.1603, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 1.3733451015004414, |
|
"grad_norm": 0.9350513325401503, |
|
"learning_rate": 9.700417199394457e-06, |
|
"loss": 0.1764, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 1.3768755516328333, |
|
"grad_norm": 0.5965930357229629, |
|
"learning_rate": 9.698463103929542e-06, |
|
"loss": 0.1688, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.380406001765225, |
|
"grad_norm": 0.9881802502016954, |
|
"learning_rate": 9.69650285427575e-06, |
|
"loss": 0.1996, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 1.383936451897617, |
|
"grad_norm": 0.6943128810137366, |
|
"learning_rate": 9.694536453000668e-06, |
|
"loss": 0.1685, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 1.3874669020300088, |
|
"grad_norm": 0.6982977323410529, |
|
"learning_rate": 9.692563902679954e-06, |
|
"loss": 0.1563, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 1.3909973521624006, |
|
"grad_norm": 0.7295520907046743, |
|
"learning_rate": 9.69058520589731e-06, |
|
"loss": 0.1692, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 1.3945278022947925, |
|
"grad_norm": 0.80781982226485, |
|
"learning_rate": 9.688600365244495e-06, |
|
"loss": 0.1762, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 1.3980582524271845, |
|
"grad_norm": 0.6182663519447118, |
|
"learning_rate": 9.686609383321316e-06, |
|
"loss": 0.1686, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 1.4015887025595763, |
|
"grad_norm": 0.5145110224464513, |
|
"learning_rate": 9.684612262735614e-06, |
|
"loss": 0.1626, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 1.4051191526919682, |
|
"grad_norm": 0.8760399209586848, |
|
"learning_rate": 9.682609006103283e-06, |
|
"loss": 0.1754, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 1.40864960282436, |
|
"grad_norm": 0.8161048235153104, |
|
"learning_rate": 9.680599616048247e-06, |
|
"loss": 0.164, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 1.412180052956752, |
|
"grad_norm": 0.7009270806175059, |
|
"learning_rate": 9.678584095202468e-06, |
|
"loss": 0.1613, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.415710503089144, |
|
"grad_norm": 0.6567037684930597, |
|
"learning_rate": 9.676562446205936e-06, |
|
"loss": 0.1498, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 1.4192409532215358, |
|
"grad_norm": 0.7885493080145554, |
|
"learning_rate": 9.674534671706666e-06, |
|
"loss": 0.1834, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 1.4227714033539276, |
|
"grad_norm": 0.561383249284057, |
|
"learning_rate": 9.672500774360698e-06, |
|
"loss": 0.1664, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 1.4263018534863194, |
|
"grad_norm": 0.8130649369686304, |
|
"learning_rate": 9.670460756832094e-06, |
|
"loss": 0.1375, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 1.4298323036187113, |
|
"grad_norm": 0.730228217232664, |
|
"learning_rate": 9.668414621792928e-06, |
|
"loss": 0.1602, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 1.4333627537511033, |
|
"grad_norm": 0.701767347722211, |
|
"learning_rate": 9.666362371923294e-06, |
|
"loss": 0.1619, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 1.4368932038834952, |
|
"grad_norm": 0.6142164548390184, |
|
"learning_rate": 9.664304009911286e-06, |
|
"loss": 0.1578, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 1.440423654015887, |
|
"grad_norm": 1.1635404922505252, |
|
"learning_rate": 9.66223953845301e-06, |
|
"loss": 0.1835, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 1.4439541041482788, |
|
"grad_norm": 0.8373651861820354, |
|
"learning_rate": 9.660168960252575e-06, |
|
"loss": 0.17, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 1.447484554280671, |
|
"grad_norm": 0.5626164848515763, |
|
"learning_rate": 9.658092278022088e-06, |
|
"loss": 0.1638, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.4510150044130627, |
|
"grad_norm": 0.9647932895601038, |
|
"learning_rate": 9.656009494481646e-06, |
|
"loss": 0.1737, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 1.4545454545454546, |
|
"grad_norm": 0.6769901378936639, |
|
"learning_rate": 9.653920612359344e-06, |
|
"loss": 0.1719, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 1.4580759046778464, |
|
"grad_norm": 0.5912616414391476, |
|
"learning_rate": 9.651825634391262e-06, |
|
"loss": 0.1547, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 1.4616063548102383, |
|
"grad_norm": 0.6130890022296159, |
|
"learning_rate": 9.649724563321465e-06, |
|
"loss": 0.1509, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 1.46513680494263, |
|
"grad_norm": 0.6981352025940074, |
|
"learning_rate": 9.647617401902003e-06, |
|
"loss": 0.1612, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.468667255075022, |
|
"grad_norm": 0.7832373213646434, |
|
"learning_rate": 9.645504152892895e-06, |
|
"loss": 0.1867, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 1.472197705207414, |
|
"grad_norm": 0.8544155478419707, |
|
"learning_rate": 9.64338481906214e-06, |
|
"loss": 0.1721, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 1.4757281553398058, |
|
"grad_norm": 0.9502151610480191, |
|
"learning_rate": 9.641259403185706e-06, |
|
"loss": 0.1743, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 1.4792586054721977, |
|
"grad_norm": 0.7733505393875248, |
|
"learning_rate": 9.639127908047527e-06, |
|
"loss": 0.1734, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 1.4827890556045895, |
|
"grad_norm": 0.8264946927060004, |
|
"learning_rate": 9.6369903364395e-06, |
|
"loss": 0.1911, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.4863195057369816, |
|
"grad_norm": 0.8237623458796101, |
|
"learning_rate": 9.63484669116148e-06, |
|
"loss": 0.169, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 1.4898499558693734, |
|
"grad_norm": 1.0384652928610334, |
|
"learning_rate": 9.632696975021278e-06, |
|
"loss": 0.1761, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 1.4933804060017652, |
|
"grad_norm": 0.9824278139544069, |
|
"learning_rate": 9.630541190834659e-06, |
|
"loss": 0.1854, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 1.496910856134157, |
|
"grad_norm": 0.7655842553137074, |
|
"learning_rate": 9.628379341425332e-06, |
|
"loss": 0.1762, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 1.500441306266549, |
|
"grad_norm": 0.7573467640480261, |
|
"learning_rate": 9.626211429624955e-06, |
|
"loss": 0.1719, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.5039717563989408, |
|
"grad_norm": 0.6342899695339446, |
|
"learning_rate": 9.624037458273121e-06, |
|
"loss": 0.1663, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 1.5075022065313326, |
|
"grad_norm": 0.7241886079251669, |
|
"learning_rate": 9.621857430217366e-06, |
|
"loss": 0.1612, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 1.5110326566637247, |
|
"grad_norm": 0.7713579132775925, |
|
"learning_rate": 9.619671348313156e-06, |
|
"loss": 0.1414, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 1.5145631067961165, |
|
"grad_norm": 0.6070241168065806, |
|
"learning_rate": 9.617479215423887e-06, |
|
"loss": 0.1752, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 1.5180935569285083, |
|
"grad_norm": 0.8462668684183862, |
|
"learning_rate": 9.615281034420882e-06, |
|
"loss": 0.1769, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.5216240070609004, |
|
"grad_norm": 0.7218104922456425, |
|
"learning_rate": 9.613076808183382e-06, |
|
"loss": 0.1531, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 1.5251544571932922, |
|
"grad_norm": 0.7261686662514826, |
|
"learning_rate": 9.610866539598551e-06, |
|
"loss": 0.1638, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 1.528684907325684, |
|
"grad_norm": 1.2087576640937627, |
|
"learning_rate": 9.608650231561466e-06, |
|
"loss": 0.1848, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 1.532215357458076, |
|
"grad_norm": 0.6381008529495783, |
|
"learning_rate": 9.606427886975113e-06, |
|
"loss": 0.1719, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 1.5357458075904677, |
|
"grad_norm": 0.9801974645943355, |
|
"learning_rate": 9.604199508750386e-06, |
|
"loss": 0.189, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.5392762577228596, |
|
"grad_norm": 0.7249244172345439, |
|
"learning_rate": 9.601965099806085e-06, |
|
"loss": 0.1729, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 1.5428067078552514, |
|
"grad_norm": 0.852227223208167, |
|
"learning_rate": 9.599724663068904e-06, |
|
"loss": 0.1522, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 1.5463371579876433, |
|
"grad_norm": 0.70519515152008, |
|
"learning_rate": 9.597478201473433e-06, |
|
"loss": 0.1682, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 1.5498676081200353, |
|
"grad_norm": 0.7777614025947788, |
|
"learning_rate": 9.595225717962157e-06, |
|
"loss": 0.16, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 1.5533980582524272, |
|
"grad_norm": 0.6144264510883923, |
|
"learning_rate": 9.592967215485449e-06, |
|
"loss": 0.1737, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.5569285083848192, |
|
"grad_norm": 0.6723553147570778, |
|
"learning_rate": 9.590702697001559e-06, |
|
"loss": 0.1864, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 1.560458958517211, |
|
"grad_norm": 0.786872255669317, |
|
"learning_rate": 9.588432165476626e-06, |
|
"loss": 0.1678, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 1.563989408649603, |
|
"grad_norm": 1.060953562194783, |
|
"learning_rate": 9.58615562388466e-06, |
|
"loss": 0.1743, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 1.5675198587819947, |
|
"grad_norm": 0.8590680802220704, |
|
"learning_rate": 9.583873075207541e-06, |
|
"loss": 0.1449, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 1.5710503089143866, |
|
"grad_norm": 0.8592072746191091, |
|
"learning_rate": 9.581584522435025e-06, |
|
"loss": 0.1689, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.5745807590467784, |
|
"grad_norm": 0.6431750629028619, |
|
"learning_rate": 9.579289968564723e-06, |
|
"loss": 0.1549, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 1.5781112091791702, |
|
"grad_norm": 0.7880523755608316, |
|
"learning_rate": 9.576989416602113e-06, |
|
"loss": 0.1569, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 1.581641659311562, |
|
"grad_norm": 0.7826920782979999, |
|
"learning_rate": 9.574682869560528e-06, |
|
"loss": 0.1655, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 1.5851721094439541, |
|
"grad_norm": 0.6507531502926956, |
|
"learning_rate": 9.57237033046115e-06, |
|
"loss": 0.1704, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 1.588702559576346, |
|
"grad_norm": 0.6131816635944837, |
|
"learning_rate": 9.570051802333017e-06, |
|
"loss": 0.1722, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.5922330097087378, |
|
"grad_norm": 0.8615528127088715, |
|
"learning_rate": 9.567727288213005e-06, |
|
"loss": 0.1671, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 1.5957634598411299, |
|
"grad_norm": 0.7175178771131497, |
|
"learning_rate": 9.565396791145833e-06, |
|
"loss": 0.1646, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 1.5992939099735217, |
|
"grad_norm": 0.6985637923301037, |
|
"learning_rate": 9.563060314184055e-06, |
|
"loss": 0.1919, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 1.6028243601059136, |
|
"grad_norm": 0.854067298117589, |
|
"learning_rate": 9.560717860388061e-06, |
|
"loss": 0.1819, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 1.6063548102383054, |
|
"grad_norm": 0.884963758284598, |
|
"learning_rate": 9.558369432826067e-06, |
|
"loss": 0.169, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 1.6098852603706972, |
|
"grad_norm": 0.7696374374267597, |
|
"learning_rate": 9.556015034574114e-06, |
|
"loss": 0.1538, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 1.613415710503089, |
|
"grad_norm": 0.6854959076334332, |
|
"learning_rate": 9.553654668716065e-06, |
|
"loss": 0.1743, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 1.616946160635481, |
|
"grad_norm": 0.7796697126004593, |
|
"learning_rate": 9.551288338343597e-06, |
|
"loss": 0.1841, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 1.6204766107678727, |
|
"grad_norm": 0.8659991614483475, |
|
"learning_rate": 9.548916046556202e-06, |
|
"loss": 0.1624, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 1.6240070609002648, |
|
"grad_norm": 0.9487399709116949, |
|
"learning_rate": 9.54653779646118e-06, |
|
"loss": 0.1855, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.6275375110326566, |
|
"grad_norm": 0.6655499212908461, |
|
"learning_rate": 9.544153591173632e-06, |
|
"loss": 0.1582, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 1.6310679611650487, |
|
"grad_norm": 0.6850321878653797, |
|
"learning_rate": 9.541763433816467e-06, |
|
"loss": 0.161, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 1.6345984112974405, |
|
"grad_norm": 0.6344259176675145, |
|
"learning_rate": 9.539367327520382e-06, |
|
"loss": 0.1899, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 1.6381288614298324, |
|
"grad_norm": 0.8057880340454765, |
|
"learning_rate": 9.536965275423869e-06, |
|
"loss": 0.1849, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 1.6416593115622242, |
|
"grad_norm": 0.6524278173237419, |
|
"learning_rate": 9.53455728067321e-06, |
|
"loss": 0.1616, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.645189761694616, |
|
"grad_norm": 0.721136853352441, |
|
"learning_rate": 9.532143346422466e-06, |
|
"loss": 0.1561, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 1.648720211827008, |
|
"grad_norm": 0.9490769657599735, |
|
"learning_rate": 9.529723475833485e-06, |
|
"loss": 0.1754, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 1.6522506619593997, |
|
"grad_norm": 0.5644282783067912, |
|
"learning_rate": 9.527297672075886e-06, |
|
"loss": 0.1627, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 1.6557811120917916, |
|
"grad_norm": 0.7598201706804275, |
|
"learning_rate": 9.524865938327058e-06, |
|
"loss": 0.1716, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 1.6593115622241836, |
|
"grad_norm": 0.7591422889029776, |
|
"learning_rate": 9.522428277772162e-06, |
|
"loss": 0.1451, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.6628420123565755, |
|
"grad_norm": 0.6982093938144709, |
|
"learning_rate": 9.519984693604116e-06, |
|
"loss": 0.1645, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 1.6663724624889673, |
|
"grad_norm": 0.5228701865370198, |
|
"learning_rate": 9.517535189023602e-06, |
|
"loss": 0.1563, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 1.6699029126213594, |
|
"grad_norm": 0.8204330233451095, |
|
"learning_rate": 9.515079767239056e-06, |
|
"loss": 0.1678, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 1.6734333627537512, |
|
"grad_norm": 0.9523128967867128, |
|
"learning_rate": 9.512618431466662e-06, |
|
"loss": 0.1664, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 1.676963812886143, |
|
"grad_norm": 0.7078287409806696, |
|
"learning_rate": 9.510151184930354e-06, |
|
"loss": 0.1658, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.6804942630185349, |
|
"grad_norm": 0.7520543206855101, |
|
"learning_rate": 9.507678030861805e-06, |
|
"loss": 0.1646, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 1.6840247131509267, |
|
"grad_norm": 0.9284654185473334, |
|
"learning_rate": 9.505198972500426e-06, |
|
"loss": 0.1684, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 1.6875551632833186, |
|
"grad_norm": 0.6224185496516447, |
|
"learning_rate": 9.502714013093364e-06, |
|
"loss": 0.1475, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 1.6910856134157104, |
|
"grad_norm": 0.7200200167776666, |
|
"learning_rate": 9.500223155895491e-06, |
|
"loss": 0.1739, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 1.6946160635481022, |
|
"grad_norm": 0.7893065612440167, |
|
"learning_rate": 9.497726404169411e-06, |
|
"loss": 0.1577, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.6981465136804943, |
|
"grad_norm": 0.806998234582965, |
|
"learning_rate": 9.495223761185443e-06, |
|
"loss": 0.1692, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 1.7016769638128861, |
|
"grad_norm": 0.8896910612567586, |
|
"learning_rate": 9.492715230221623e-06, |
|
"loss": 0.155, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 1.705207413945278, |
|
"grad_norm": 0.7929727685131104, |
|
"learning_rate": 9.490200814563704e-06, |
|
"loss": 0.1693, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 1.70873786407767, |
|
"grad_norm": 0.8281002779014005, |
|
"learning_rate": 9.48768051750514e-06, |
|
"loss": 0.162, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 1.7122683142100619, |
|
"grad_norm": 0.7574881698882824, |
|
"learning_rate": 9.485154342347093e-06, |
|
"loss": 0.1822, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.7157987643424537, |
|
"grad_norm": 0.7603690459222415, |
|
"learning_rate": 9.482622292398424e-06, |
|
"loss": 0.1705, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 1.7193292144748455, |
|
"grad_norm": 0.6922874253842599, |
|
"learning_rate": 9.480084370975691e-06, |
|
"loss": 0.1687, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 1.7228596646072374, |
|
"grad_norm": 0.80394336074322, |
|
"learning_rate": 9.477540581403137e-06, |
|
"loss": 0.1688, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 1.7263901147396292, |
|
"grad_norm": 0.8802309764956326, |
|
"learning_rate": 9.474990927012694e-06, |
|
"loss": 0.1304, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 1.729920564872021, |
|
"grad_norm": 0.6137981646644614, |
|
"learning_rate": 9.472435411143979e-06, |
|
"loss": 0.1664, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.733451015004413, |
|
"grad_norm": 0.9282185888184011, |
|
"learning_rate": 9.469874037144281e-06, |
|
"loss": 0.2012, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 1.736981465136805, |
|
"grad_norm": 0.6630691029948931, |
|
"learning_rate": 9.467306808368568e-06, |
|
"loss": 0.1502, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 1.7405119152691968, |
|
"grad_norm": 0.7577885818393465, |
|
"learning_rate": 9.46473372817947e-06, |
|
"loss": 0.1702, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 1.7440423654015889, |
|
"grad_norm": 0.5899998168235006, |
|
"learning_rate": 9.46215479994729e-06, |
|
"loss": 0.1602, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 1.7475728155339807, |
|
"grad_norm": 1.0306863382019156, |
|
"learning_rate": 9.459570027049983e-06, |
|
"loss": 0.1334, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.7511032656663725, |
|
"grad_norm": 0.865412362993785, |
|
"learning_rate": 9.456979412873163e-06, |
|
"loss": 0.1852, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 1.7546337157987644, |
|
"grad_norm": 0.8407963911360593, |
|
"learning_rate": 9.454382960810093e-06, |
|
"loss": 0.1691, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 1.7581641659311562, |
|
"grad_norm": 0.6988074166274713, |
|
"learning_rate": 9.451780674261687e-06, |
|
"loss": 0.16, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 1.761694616063548, |
|
"grad_norm": 0.6567772580817499, |
|
"learning_rate": 9.4491725566365e-06, |
|
"loss": 0.1678, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 1.7652250661959399, |
|
"grad_norm": 0.8826194122518776, |
|
"learning_rate": 9.446558611350718e-06, |
|
"loss": 0.177, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.7687555163283317, |
|
"grad_norm": 0.6900513296683881, |
|
"learning_rate": 9.44393884182817e-06, |
|
"loss": 0.1852, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 1.7722859664607238, |
|
"grad_norm": 0.6605181197941491, |
|
"learning_rate": 9.441313251500307e-06, |
|
"loss": 0.1557, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 1.7758164165931156, |
|
"grad_norm": 0.6146198835230601, |
|
"learning_rate": 9.438681843806207e-06, |
|
"loss": 0.1594, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 1.7793468667255075, |
|
"grad_norm": 0.5149059845127192, |
|
"learning_rate": 9.436044622192569e-06, |
|
"loss": 0.1665, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 1.7828773168578995, |
|
"grad_norm": 0.9135519362694595, |
|
"learning_rate": 9.433401590113702e-06, |
|
"loss": 0.1797, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 1.7864077669902914, |
|
"grad_norm": 0.8916844443544416, |
|
"learning_rate": 9.43075275103153e-06, |
|
"loss": 0.17, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 1.7899382171226832, |
|
"grad_norm": 0.8009121594911396, |
|
"learning_rate": 9.428098108415587e-06, |
|
"loss": 0.1599, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 1.793468667255075, |
|
"grad_norm": 0.8269496601871587, |
|
"learning_rate": 9.425437665742998e-06, |
|
"loss": 0.1633, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 1.7969991173874669, |
|
"grad_norm": 0.6798890374153103, |
|
"learning_rate": 9.422771426498493e-06, |
|
"loss": 0.1605, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 1.8005295675198587, |
|
"grad_norm": 0.6315062922758498, |
|
"learning_rate": 9.420099394174397e-06, |
|
"loss": 0.1795, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.8040600176522505, |
|
"grad_norm": 0.5746772810241558, |
|
"learning_rate": 9.417421572270612e-06, |
|
"loss": 0.1567, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 1.8075904677846424, |
|
"grad_norm": 0.8050650153052257, |
|
"learning_rate": 9.414737964294636e-06, |
|
"loss": 0.1729, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 1.8111209179170344, |
|
"grad_norm": 0.480452501514603, |
|
"learning_rate": 9.412048573761535e-06, |
|
"loss": 0.1766, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 1.8146513680494263, |
|
"grad_norm": 0.9104720825841831, |
|
"learning_rate": 9.409353404193958e-06, |
|
"loss": 0.1745, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 1.8181818181818183, |
|
"grad_norm": 0.8780636579991754, |
|
"learning_rate": 9.406652459122115e-06, |
|
"loss": 0.1762, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 1.8217122683142102, |
|
"grad_norm": 0.6984924204172243, |
|
"learning_rate": 9.403945742083788e-06, |
|
"loss": 0.1793, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 1.825242718446602, |
|
"grad_norm": 0.8238701991708997, |
|
"learning_rate": 9.401233256624318e-06, |
|
"loss": 0.1485, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 1.8287731685789939, |
|
"grad_norm": 0.8696582194623638, |
|
"learning_rate": 9.398515006296598e-06, |
|
"loss": 0.1891, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 1.8323036187113857, |
|
"grad_norm": 0.699771115532138, |
|
"learning_rate": 9.395790994661075e-06, |
|
"loss": 0.1553, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 1.8358340688437775, |
|
"grad_norm": 0.9125794295579036, |
|
"learning_rate": 9.393061225285743e-06, |
|
"loss": 0.163, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.8393645189761694, |
|
"grad_norm": 0.8950419906565266, |
|
"learning_rate": 9.390325701746135e-06, |
|
"loss": 0.167, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 1.8428949691085612, |
|
"grad_norm": 0.8710778282196566, |
|
"learning_rate": 9.387584427625323e-06, |
|
"loss": 0.1585, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 1.8464254192409533, |
|
"grad_norm": 0.8901232516687565, |
|
"learning_rate": 9.38483740651391e-06, |
|
"loss": 0.171, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 1.849955869373345, |
|
"grad_norm": 0.7244577661750393, |
|
"learning_rate": 9.382084642010026e-06, |
|
"loss": 0.1718, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 1.853486319505737, |
|
"grad_norm": 0.9044306478777916, |
|
"learning_rate": 9.379326137719329e-06, |
|
"loss": 0.1762, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.857016769638129, |
|
"grad_norm": 0.6730866435868901, |
|
"learning_rate": 9.376561897254987e-06, |
|
"loss": 0.1567, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 1.8605472197705208, |
|
"grad_norm": 0.816989704732525, |
|
"learning_rate": 9.37379192423769e-06, |
|
"loss": 0.1919, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 1.8640776699029127, |
|
"grad_norm": 0.7655388484499017, |
|
"learning_rate": 9.371016222295628e-06, |
|
"loss": 0.1736, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 1.8676081200353045, |
|
"grad_norm": 0.8699212931081676, |
|
"learning_rate": 9.368234795064502e-06, |
|
"loss": 0.1625, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 1.8711385701676964, |
|
"grad_norm": 1.0534597898924092, |
|
"learning_rate": 9.365447646187509e-06, |
|
"loss": 0.1381, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.8746690203000882, |
|
"grad_norm": 0.6614479535387474, |
|
"learning_rate": 9.36265477931534e-06, |
|
"loss": 0.1467, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 1.87819947043248, |
|
"grad_norm": 0.8393186062366995, |
|
"learning_rate": 9.359856198106176e-06, |
|
"loss": 0.1913, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 1.8817299205648719, |
|
"grad_norm": 0.5866280093193419, |
|
"learning_rate": 9.357051906225685e-06, |
|
"loss": 0.1566, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 1.885260370697264, |
|
"grad_norm": 0.9996266820247056, |
|
"learning_rate": 9.35424190734701e-06, |
|
"loss": 0.1919, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 1.8887908208296558, |
|
"grad_norm": 0.8127958941754003, |
|
"learning_rate": 9.351426205150778e-06, |
|
"loss": 0.1695, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 1.8923212709620476, |
|
"grad_norm": 0.6634743250735253, |
|
"learning_rate": 9.348604803325075e-06, |
|
"loss": 0.1555, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 1.8958517210944397, |
|
"grad_norm": 0.812340369225641, |
|
"learning_rate": 9.34577770556546e-06, |
|
"loss": 0.1645, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 1.8993821712268315, |
|
"grad_norm": 0.7497218737066156, |
|
"learning_rate": 9.342944915574952e-06, |
|
"loss": 0.1668, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 1.9029126213592233, |
|
"grad_norm": 0.5695613946823557, |
|
"learning_rate": 9.340106437064026e-06, |
|
"loss": 0.1625, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 1.9064430714916152, |
|
"grad_norm": 0.8336306538843815, |
|
"learning_rate": 9.337262273750604e-06, |
|
"loss": 0.1679, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.909973521624007, |
|
"grad_norm": 0.7985645323717766, |
|
"learning_rate": 9.334412429360058e-06, |
|
"loss": 0.1717, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 1.9135039717563989, |
|
"grad_norm": 0.803743435053368, |
|
"learning_rate": 9.3315569076252e-06, |
|
"loss": 0.1811, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 1.9170344218887907, |
|
"grad_norm": 0.7792918327227041, |
|
"learning_rate": 9.328695712286283e-06, |
|
"loss": 0.178, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 1.9205648720211828, |
|
"grad_norm": 1.0216399627321537, |
|
"learning_rate": 9.32582884709098e-06, |
|
"loss": 0.1702, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 1.9240953221535746, |
|
"grad_norm": 0.8524211551346756, |
|
"learning_rate": 9.322956315794403e-06, |
|
"loss": 0.1769, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 1.9276257722859664, |
|
"grad_norm": 1.0061876178961442, |
|
"learning_rate": 9.320078122159077e-06, |
|
"loss": 0.1725, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 1.9311562224183585, |
|
"grad_norm": 0.6449173961835437, |
|
"learning_rate": 9.317194269954948e-06, |
|
"loss": 0.1669, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 1.9346866725507503, |
|
"grad_norm": 0.6716320314286484, |
|
"learning_rate": 9.314304762959373e-06, |
|
"loss": 0.1837, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 1.9382171226831422, |
|
"grad_norm": 0.7847559077293133, |
|
"learning_rate": 9.311409604957116e-06, |
|
"loss": 0.1697, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.941747572815534, |
|
"grad_norm": 0.6479063264651344, |
|
"learning_rate": 9.30850879974034e-06, |
|
"loss": 0.1887, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.9452780229479258, |
|
"grad_norm": 0.5778944362755724, |
|
"learning_rate": 9.305602351108613e-06, |
|
"loss": 0.1732, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 1.9488084730803177, |
|
"grad_norm": 0.5993363081022134, |
|
"learning_rate": 9.302690262868882e-06, |
|
"loss": 0.1524, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 1.9523389232127095, |
|
"grad_norm": 0.7186259440033298, |
|
"learning_rate": 9.299772538835492e-06, |
|
"loss": 0.1562, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 1.9558693733451014, |
|
"grad_norm": 0.7547445542865213, |
|
"learning_rate": 9.296849182830167e-06, |
|
"loss": 0.159, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 1.9593998234774934, |
|
"grad_norm": 0.742454944410959, |
|
"learning_rate": 9.293920198682004e-06, |
|
"loss": 0.178, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.9629302736098853, |
|
"grad_norm": 0.731704919378445, |
|
"learning_rate": 9.290985590227478e-06, |
|
"loss": 0.178, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 1.966460723742277, |
|
"grad_norm": 0.6326777732721525, |
|
"learning_rate": 9.288045361310423e-06, |
|
"loss": 0.1572, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 1.9699911738746692, |
|
"grad_norm": 0.6096205323065964, |
|
"learning_rate": 9.285099515782045e-06, |
|
"loss": 0.1609, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 1.973521624007061, |
|
"grad_norm": 0.968981563236468, |
|
"learning_rate": 9.2821480575009e-06, |
|
"loss": 0.1554, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 1.9770520741394528, |
|
"grad_norm": 0.6731379335046332, |
|
"learning_rate": 9.279190990332896e-06, |
|
"loss": 0.1786, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.9805825242718447, |
|
"grad_norm": 0.5474632434518878, |
|
"learning_rate": 9.276228318151288e-06, |
|
"loss": 0.1825, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 1.9841129744042365, |
|
"grad_norm": 0.5300134793710058, |
|
"learning_rate": 9.273260044836675e-06, |
|
"loss": 0.1507, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 1.9876434245366283, |
|
"grad_norm": 0.6263089144886929, |
|
"learning_rate": 9.270286174276992e-06, |
|
"loss": 0.1753, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 1.9911738746690202, |
|
"grad_norm": 0.7895430535402537, |
|
"learning_rate": 9.267306710367503e-06, |
|
"loss": 0.1772, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 1.994704324801412, |
|
"grad_norm": 0.7740002132732854, |
|
"learning_rate": 9.2643216570108e-06, |
|
"loss": 0.1874, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.998234774933804, |
|
"grad_norm": 0.9874775418494401, |
|
"learning_rate": 9.261331018116795e-06, |
|
"loss": 0.173, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 2.001765225066196, |
|
"grad_norm": 0.6888142224041766, |
|
"learning_rate": 9.25833479760272e-06, |
|
"loss": 0.1811, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 2.005295675198588, |
|
"grad_norm": 0.6459025197429961, |
|
"learning_rate": 9.255332999393114e-06, |
|
"loss": 0.1507, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 2.00882612533098, |
|
"grad_norm": 0.6618092460843229, |
|
"learning_rate": 9.252325627419821e-06, |
|
"loss": 0.1302, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 2.0123565754633717, |
|
"grad_norm": 0.8386091353618218, |
|
"learning_rate": 9.249312685621989e-06, |
|
"loss": 0.1531, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.0158870255957635, |
|
"grad_norm": 0.7735686550828623, |
|
"learning_rate": 9.246294177946062e-06, |
|
"loss": 0.1554, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 2.0194174757281553, |
|
"grad_norm": 0.6350949023955103, |
|
"learning_rate": 9.24327010834577e-06, |
|
"loss": 0.1161, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 2.022947925860547, |
|
"grad_norm": 0.7901812966196844, |
|
"learning_rate": 9.24024048078213e-06, |
|
"loss": 0.1498, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 2.026478375992939, |
|
"grad_norm": 0.7745266020928578, |
|
"learning_rate": 9.237205299223444e-06, |
|
"loss": 0.1665, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 2.030008826125331, |
|
"grad_norm": 0.6972058102385509, |
|
"learning_rate": 9.234164567645278e-06, |
|
"loss": 0.1257, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 2.0335392762577227, |
|
"grad_norm": 0.6137168822739238, |
|
"learning_rate": 9.231118290030481e-06, |
|
"loss": 0.121, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 2.0370697263901145, |
|
"grad_norm": 0.7533855618361971, |
|
"learning_rate": 9.228066470369155e-06, |
|
"loss": 0.1352, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 2.040600176522507, |
|
"grad_norm": 0.792710448056873, |
|
"learning_rate": 9.225009112658667e-06, |
|
"loss": 0.144, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 2.0441306266548986, |
|
"grad_norm": 0.865387452735705, |
|
"learning_rate": 9.221946220903634e-06, |
|
"loss": 0.1521, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 2.0476610767872905, |
|
"grad_norm": 1.0235916124259214, |
|
"learning_rate": 9.218877799115929e-06, |
|
"loss": 0.1441, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.0511915269196823, |
|
"grad_norm": 1.0508276358903932, |
|
"learning_rate": 9.21580385131466e-06, |
|
"loss": 0.1685, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 2.054721977052074, |
|
"grad_norm": 0.8910905064383835, |
|
"learning_rate": 9.212724381526177e-06, |
|
"loss": 0.1736, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 2.058252427184466, |
|
"grad_norm": 1.0571601952147707, |
|
"learning_rate": 9.209639393784064e-06, |
|
"loss": 0.1565, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 2.061782877316858, |
|
"grad_norm": 0.8815436783906608, |
|
"learning_rate": 9.206548892129129e-06, |
|
"loss": 0.1371, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 2.0653133274492497, |
|
"grad_norm": 1.0517504956260713, |
|
"learning_rate": 9.203452880609407e-06, |
|
"loss": 0.1464, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 2.0688437775816415, |
|
"grad_norm": 0.8529237538392984, |
|
"learning_rate": 9.200351363280144e-06, |
|
"loss": 0.1266, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 2.0723742277140333, |
|
"grad_norm": 0.874098127040573, |
|
"learning_rate": 9.197244344203804e-06, |
|
"loss": 0.1349, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 2.0759046778464256, |
|
"grad_norm": 0.8109927532199982, |
|
"learning_rate": 9.194131827450054e-06, |
|
"loss": 0.1445, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 2.0794351279788175, |
|
"grad_norm": 0.6489560116761195, |
|
"learning_rate": 9.191013817095762e-06, |
|
"loss": 0.1294, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 2.0829655781112093, |
|
"grad_norm": 0.7508144606670216, |
|
"learning_rate": 9.18789031722499e-06, |
|
"loss": 0.1135, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.086496028243601, |
|
"grad_norm": 0.7295793522432309, |
|
"learning_rate": 9.184761331928997e-06, |
|
"loss": 0.1363, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 2.090026478375993, |
|
"grad_norm": 0.861095430673198, |
|
"learning_rate": 9.18162686530622e-06, |
|
"loss": 0.1279, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 2.093556928508385, |
|
"grad_norm": 0.6861216733576722, |
|
"learning_rate": 9.178486921462276e-06, |
|
"loss": 0.1224, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 2.0970873786407767, |
|
"grad_norm": 0.8019378401083357, |
|
"learning_rate": 9.175341504509961e-06, |
|
"loss": 0.1545, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 2.1006178287731685, |
|
"grad_norm": 1.526043856422149, |
|
"learning_rate": 9.172190618569236e-06, |
|
"loss": 0.1306, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 2.1041482789055603, |
|
"grad_norm": 0.8476895681346993, |
|
"learning_rate": 9.16903426776723e-06, |
|
"loss": 0.149, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 2.107678729037952, |
|
"grad_norm": 0.8161608102533534, |
|
"learning_rate": 9.16587245623822e-06, |
|
"loss": 0.1364, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 2.111209179170344, |
|
"grad_norm": 0.7223711233696612, |
|
"learning_rate": 9.162705188123647e-06, |
|
"loss": 0.142, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 2.1147396293027363, |
|
"grad_norm": 0.8507390490828752, |
|
"learning_rate": 9.159532467572094e-06, |
|
"loss": 0.1547, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 2.118270079435128, |
|
"grad_norm": 0.8316619115260057, |
|
"learning_rate": 9.156354298739285e-06, |
|
"loss": 0.1563, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.12180052956752, |
|
"grad_norm": 0.7544897788791859, |
|
"learning_rate": 9.153170685788081e-06, |
|
"loss": 0.1421, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 2.125330979699912, |
|
"grad_norm": 0.800819425552279, |
|
"learning_rate": 9.149981632888476e-06, |
|
"loss": 0.1579, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 2.1288614298323036, |
|
"grad_norm": 0.7838909562518437, |
|
"learning_rate": 9.146787144217588e-06, |
|
"loss": 0.1681, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 2.1323918799646955, |
|
"grad_norm": 0.9303054432448055, |
|
"learning_rate": 9.143587223959653e-06, |
|
"loss": 0.1397, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 2.1359223300970873, |
|
"grad_norm": 0.7625989738070533, |
|
"learning_rate": 9.140381876306026e-06, |
|
"loss": 0.1581, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 2.139452780229479, |
|
"grad_norm": 0.7857755303539646, |
|
"learning_rate": 9.137171105455165e-06, |
|
"loss": 0.1258, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 2.142983230361871, |
|
"grad_norm": 0.6441805120671155, |
|
"learning_rate": 9.133954915612635e-06, |
|
"loss": 0.1305, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 2.146513680494263, |
|
"grad_norm": 0.8467018145382312, |
|
"learning_rate": 9.130733310991102e-06, |
|
"loss": 0.1432, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 2.1500441306266547, |
|
"grad_norm": 0.7571331987070536, |
|
"learning_rate": 9.127506295810316e-06, |
|
"loss": 0.1536, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 2.153574580759047, |
|
"grad_norm": 1.0166648071075606, |
|
"learning_rate": 9.124273874297123e-06, |
|
"loss": 0.1641, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.157105030891439, |
|
"grad_norm": 1.0624683994204978, |
|
"learning_rate": 9.121036050685443e-06, |
|
"loss": 0.1568, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 2.1606354810238306, |
|
"grad_norm": 0.754254288642853, |
|
"learning_rate": 9.11779282921628e-06, |
|
"loss": 0.1456, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 2.1641659311562225, |
|
"grad_norm": 0.8711250730686613, |
|
"learning_rate": 9.114544214137699e-06, |
|
"loss": 0.1664, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 2.1676963812886143, |
|
"grad_norm": 0.8545203984797597, |
|
"learning_rate": 9.111290209704835e-06, |
|
"loss": 0.1494, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 2.171226831421006, |
|
"grad_norm": 0.8374462073082292, |
|
"learning_rate": 9.108030820179882e-06, |
|
"loss": 0.1541, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 2.174757281553398, |
|
"grad_norm": 0.7357997781637011, |
|
"learning_rate": 9.104766049832088e-06, |
|
"loss": 0.1491, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 2.17828773168579, |
|
"grad_norm": 0.7606936499199851, |
|
"learning_rate": 9.101495902937744e-06, |
|
"loss": 0.1577, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 2.1818181818181817, |
|
"grad_norm": 0.8031929036786638, |
|
"learning_rate": 9.098220383780191e-06, |
|
"loss": 0.1649, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 2.1853486319505735, |
|
"grad_norm": 1.064933842253509, |
|
"learning_rate": 9.094939496649803e-06, |
|
"loss": 0.1218, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 2.1888790820829658, |
|
"grad_norm": 0.7301032836107995, |
|
"learning_rate": 9.091653245843984e-06, |
|
"loss": 0.1376, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.1924095322153576, |
|
"grad_norm": 0.8431618205640518, |
|
"learning_rate": 9.088361635667162e-06, |
|
"loss": 0.148, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 2.1959399823477495, |
|
"grad_norm": 1.340479354749448, |
|
"learning_rate": 9.08506467043079e-06, |
|
"loss": 0.1616, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 2.1994704324801413, |
|
"grad_norm": 0.8064011593740745, |
|
"learning_rate": 9.081762354453333e-06, |
|
"loss": 0.1488, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 2.203000882612533, |
|
"grad_norm": 0.8004191241563091, |
|
"learning_rate": 9.078454692060264e-06, |
|
"loss": 0.1403, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 2.206531332744925, |
|
"grad_norm": 0.8885936871876676, |
|
"learning_rate": 9.075141687584056e-06, |
|
"loss": 0.1298, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 2.210061782877317, |
|
"grad_norm": 1.013124408250628, |
|
"learning_rate": 9.071823345364187e-06, |
|
"loss": 0.1461, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 2.2135922330097086, |
|
"grad_norm": 0.8812285753845427, |
|
"learning_rate": 9.068499669747122e-06, |
|
"loss": 0.1475, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 2.2171226831421005, |
|
"grad_norm": 0.8390832015738438, |
|
"learning_rate": 9.065170665086307e-06, |
|
"loss": 0.152, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 2.2206531332744923, |
|
"grad_norm": 0.6214987457344594, |
|
"learning_rate": 9.061836335742175e-06, |
|
"loss": 0.1418, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 2.2241835834068846, |
|
"grad_norm": 0.8210501747316103, |
|
"learning_rate": 9.058496686082132e-06, |
|
"loss": 0.1385, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.2277140335392764, |
|
"grad_norm": 0.8877640524225171, |
|
"learning_rate": 9.055151720480553e-06, |
|
"loss": 0.1636, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 2.2312444836716683, |
|
"grad_norm": 0.6993342223349599, |
|
"learning_rate": 9.051801443318772e-06, |
|
"loss": 0.1267, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 2.23477493380406, |
|
"grad_norm": 0.915383376214395, |
|
"learning_rate": 9.048445858985085e-06, |
|
"loss": 0.1346, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 2.238305383936452, |
|
"grad_norm": 0.8226201621022076, |
|
"learning_rate": 9.045084971874738e-06, |
|
"loss": 0.145, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 2.241835834068844, |
|
"grad_norm": 0.949771207964201, |
|
"learning_rate": 9.041718786389921e-06, |
|
"loss": 0.1729, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 2.2453662842012356, |
|
"grad_norm": 0.7669826168302643, |
|
"learning_rate": 9.038347306939768e-06, |
|
"loss": 0.1491, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 2.2488967343336275, |
|
"grad_norm": 0.8401151453900115, |
|
"learning_rate": 9.034970537940342e-06, |
|
"loss": 0.1587, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 2.2524271844660193, |
|
"grad_norm": 0.8362745634336628, |
|
"learning_rate": 9.031588483814639e-06, |
|
"loss": 0.152, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 2.255957634598411, |
|
"grad_norm": 0.828931692995688, |
|
"learning_rate": 9.028201148992572e-06, |
|
"loss": 0.127, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 2.259488084730803, |
|
"grad_norm": 0.8901277435562056, |
|
"learning_rate": 9.024808537910981e-06, |
|
"loss": 0.1472, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.2630185348631953, |
|
"grad_norm": 0.8458917100010721, |
|
"learning_rate": 9.021410655013604e-06, |
|
"loss": 0.1549, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 2.266548984995587, |
|
"grad_norm": 0.7511334646906924, |
|
"learning_rate": 9.018007504751094e-06, |
|
"loss": 0.1401, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 2.270079435127979, |
|
"grad_norm": 0.8998427764695459, |
|
"learning_rate": 9.014599091581e-06, |
|
"loss": 0.1665, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 2.2736098852603708, |
|
"grad_norm": 0.8241701592002284, |
|
"learning_rate": 9.011185419967761e-06, |
|
"loss": 0.136, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 2.2771403353927626, |
|
"grad_norm": 0.7830126296047867, |
|
"learning_rate": 9.007766494382713e-06, |
|
"loss": 0.151, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 2.2806707855251545, |
|
"grad_norm": 1.0742703925847628, |
|
"learning_rate": 9.004342319304063e-06, |
|
"loss": 0.1606, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 2.2842012356575463, |
|
"grad_norm": 0.9139889110017969, |
|
"learning_rate": 9.000912899216902e-06, |
|
"loss": 0.1556, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 2.287731685789938, |
|
"grad_norm": 0.9228198414722247, |
|
"learning_rate": 8.997478238613186e-06, |
|
"loss": 0.1635, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 2.29126213592233, |
|
"grad_norm": 0.6979642688439006, |
|
"learning_rate": 8.994038341991739e-06, |
|
"loss": 0.1443, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 2.294792586054722, |
|
"grad_norm": 0.7432603342893744, |
|
"learning_rate": 8.990593213858239e-06, |
|
"loss": 0.1497, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.2983230361871136, |
|
"grad_norm": 0.9788738069579437, |
|
"learning_rate": 8.987142858725221e-06, |
|
"loss": 0.1388, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 2.301853486319506, |
|
"grad_norm": 0.9451424967898011, |
|
"learning_rate": 8.983687281112066e-06, |
|
"loss": 0.1558, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 2.3053839364518978, |
|
"grad_norm": 0.796516553881763, |
|
"learning_rate": 8.98022648554499e-06, |
|
"loss": 0.1395, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 2.3089143865842896, |
|
"grad_norm": 0.6535029248700488, |
|
"learning_rate": 8.976760476557049e-06, |
|
"loss": 0.1466, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 2.3124448367166814, |
|
"grad_norm": 1.0145090677139883, |
|
"learning_rate": 8.973289258688125e-06, |
|
"loss": 0.1601, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 2.3159752868490733, |
|
"grad_norm": 0.9360528769068794, |
|
"learning_rate": 8.969812836484929e-06, |
|
"loss": 0.164, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 2.319505736981465, |
|
"grad_norm": 0.8047358252615086, |
|
"learning_rate": 8.966331214500978e-06, |
|
"loss": 0.1756, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 2.323036187113857, |
|
"grad_norm": 0.7975682754458384, |
|
"learning_rate": 8.962844397296608e-06, |
|
"loss": 0.1435, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 2.326566637246249, |
|
"grad_norm": 0.7277448447990951, |
|
"learning_rate": 8.959352389438959e-06, |
|
"loss": 0.1352, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 2.3300970873786406, |
|
"grad_norm": 0.8639595145485759, |
|
"learning_rate": 8.955855195501966e-06, |
|
"loss": 0.1497, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.3336275375110325, |
|
"grad_norm": 0.9741801125022028, |
|
"learning_rate": 8.952352820066359e-06, |
|
"loss": 0.1743, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 2.3371579876434243, |
|
"grad_norm": 0.7011036253403068, |
|
"learning_rate": 8.948845267719658e-06, |
|
"loss": 0.1483, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 2.3406884377758166, |
|
"grad_norm": 0.7915113319296646, |
|
"learning_rate": 8.945332543056158e-06, |
|
"loss": 0.1674, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 2.3442188879082084, |
|
"grad_norm": 0.7246375584288023, |
|
"learning_rate": 8.941814650676934e-06, |
|
"loss": 0.1429, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 2.3477493380406003, |
|
"grad_norm": 0.6572418012169463, |
|
"learning_rate": 8.938291595189826e-06, |
|
"loss": 0.1464, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 2.351279788172992, |
|
"grad_norm": 1.0088944178687222, |
|
"learning_rate": 8.93476338120944e-06, |
|
"loss": 0.1723, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 2.354810238305384, |
|
"grad_norm": 0.8873624823759165, |
|
"learning_rate": 8.931230013357136e-06, |
|
"loss": 0.1478, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 2.358340688437776, |
|
"grad_norm": 0.8232927235637216, |
|
"learning_rate": 8.927691496261025e-06, |
|
"loss": 0.1509, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 2.3618711385701676, |
|
"grad_norm": 0.7448944968086746, |
|
"learning_rate": 8.924147834555966e-06, |
|
"loss": 0.144, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 2.3654015887025595, |
|
"grad_norm": 0.6545431722863853, |
|
"learning_rate": 8.920599032883553e-06, |
|
"loss": 0.1442, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.3689320388349513, |
|
"grad_norm": 0.9857690033050072, |
|
"learning_rate": 8.917045095892115e-06, |
|
"loss": 0.1534, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 2.3724624889673436, |
|
"grad_norm": 0.8890672612580922, |
|
"learning_rate": 8.913486028236703e-06, |
|
"loss": 0.1414, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 2.375992939099735, |
|
"grad_norm": 0.7720214368079358, |
|
"learning_rate": 8.909921834579093e-06, |
|
"loss": 0.1379, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 2.3795233892321273, |
|
"grad_norm": 0.7968382535551254, |
|
"learning_rate": 8.906352519587774e-06, |
|
"loss": 0.1665, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 2.383053839364519, |
|
"grad_norm": 1.012361224051788, |
|
"learning_rate": 8.902778087937943e-06, |
|
"loss": 0.1506, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 2.386584289496911, |
|
"grad_norm": 1.6922659485205467, |
|
"learning_rate": 8.899198544311496e-06, |
|
"loss": 0.1621, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 2.3901147396293028, |
|
"grad_norm": 0.8810742656212996, |
|
"learning_rate": 8.89561389339703e-06, |
|
"loss": 0.1342, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 2.3936451897616946, |
|
"grad_norm": 0.8922813495137972, |
|
"learning_rate": 8.892024139889827e-06, |
|
"loss": 0.1348, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 2.3971756398940864, |
|
"grad_norm": 0.6428583850540706, |
|
"learning_rate": 8.888429288491857e-06, |
|
"loss": 0.1365, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 2.4007060900264783, |
|
"grad_norm": 0.959500169941628, |
|
"learning_rate": 8.884829343911763e-06, |
|
"loss": 0.1579, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.40423654015887, |
|
"grad_norm": 0.8928078544095518, |
|
"learning_rate": 8.88122431086486e-06, |
|
"loss": 0.158, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 2.407766990291262, |
|
"grad_norm": 0.7299606081515114, |
|
"learning_rate": 8.877614194073134e-06, |
|
"loss": 0.1332, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 2.4112974404236542, |
|
"grad_norm": 0.867643203493481, |
|
"learning_rate": 8.87399899826522e-06, |
|
"loss": 0.1567, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 2.414827890556046, |
|
"grad_norm": 0.8544098798566452, |
|
"learning_rate": 8.870378728176412e-06, |
|
"loss": 0.1479, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 2.418358340688438, |
|
"grad_norm": 1.0065430757516332, |
|
"learning_rate": 8.86675338854865e-06, |
|
"loss": 0.1826, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 2.4218887908208298, |
|
"grad_norm": 0.9392726338359665, |
|
"learning_rate": 8.863122984130514e-06, |
|
"loss": 0.1573, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 2.4254192409532216, |
|
"grad_norm": 0.815039972544626, |
|
"learning_rate": 8.859487519677215e-06, |
|
"loss": 0.1296, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 2.4289496910856134, |
|
"grad_norm": 0.8603703909617912, |
|
"learning_rate": 8.855846999950595e-06, |
|
"loss": 0.1547, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 2.4324801412180053, |
|
"grad_norm": 1.0058587648363357, |
|
"learning_rate": 8.852201429719119e-06, |
|
"loss": 0.1394, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 2.436010591350397, |
|
"grad_norm": 0.895199851785628, |
|
"learning_rate": 8.848550813757861e-06, |
|
"loss": 0.1379, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.439541041482789, |
|
"grad_norm": 0.7927306345545818, |
|
"learning_rate": 8.844895156848511e-06, |
|
"loss": 0.1347, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 2.443071491615181, |
|
"grad_norm": 0.7501822257778045, |
|
"learning_rate": 8.841234463779359e-06, |
|
"loss": 0.1472, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 2.4466019417475726, |
|
"grad_norm": 0.8155634914790267, |
|
"learning_rate": 8.83756873934529e-06, |
|
"loss": 0.1637, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 2.450132391879965, |
|
"grad_norm": 0.9416808234395326, |
|
"learning_rate": 8.83389798834778e-06, |
|
"loss": 0.1619, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 2.4536628420123567, |
|
"grad_norm": 0.7762931486236544, |
|
"learning_rate": 8.83022221559489e-06, |
|
"loss": 0.1607, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 2.4571932921447486, |
|
"grad_norm": 0.6647840629115083, |
|
"learning_rate": 8.82654142590126e-06, |
|
"loss": 0.1474, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 2.4607237422771404, |
|
"grad_norm": 0.809281030127039, |
|
"learning_rate": 8.822855624088099e-06, |
|
"loss": 0.1329, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 2.4642541924095323, |
|
"grad_norm": 0.7105112267949759, |
|
"learning_rate": 8.819164814983179e-06, |
|
"loss": 0.1513, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 2.467784642541924, |
|
"grad_norm": 0.6643065745395029, |
|
"learning_rate": 8.815469003420835e-06, |
|
"loss": 0.1355, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 2.471315092674316, |
|
"grad_norm": 0.6786894914054595, |
|
"learning_rate": 8.811768194241951e-06, |
|
"loss": 0.1491, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.4748455428067078, |
|
"grad_norm": 0.8985899448983452, |
|
"learning_rate": 8.808062392293964e-06, |
|
"loss": 0.1683, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 2.4783759929390996, |
|
"grad_norm": 0.8051939382089317, |
|
"learning_rate": 8.804351602430838e-06, |
|
"loss": 0.1316, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 2.4819064430714914, |
|
"grad_norm": 0.9657848784459468, |
|
"learning_rate": 8.800635829513083e-06, |
|
"loss": 0.1695, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 2.4854368932038833, |
|
"grad_norm": 0.8063728250819603, |
|
"learning_rate": 8.796915078407729e-06, |
|
"loss": 0.1524, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 2.4889673433362756, |
|
"grad_norm": 0.6461209785079439, |
|
"learning_rate": 8.793189353988325e-06, |
|
"loss": 0.1487, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 2.4924977934686674, |
|
"grad_norm": 0.6869278039173332, |
|
"learning_rate": 8.789458661134943e-06, |
|
"loss": 0.1442, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 2.4960282436010592, |
|
"grad_norm": 1.1034616460480864, |
|
"learning_rate": 8.785723004734153e-06, |
|
"loss": 0.1714, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 2.499558693733451, |
|
"grad_norm": 0.6104590542245811, |
|
"learning_rate": 8.781982389679033e-06, |
|
"loss": 0.1339, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 2.503089143865843, |
|
"grad_norm": 0.8481696722493157, |
|
"learning_rate": 8.778236820869152e-06, |
|
"loss": 0.1536, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 2.5066195939982348, |
|
"grad_norm": 0.8111930888716574, |
|
"learning_rate": 8.77448630321057e-06, |
|
"loss": 0.1422, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.5101500441306266, |
|
"grad_norm": 0.7036188732668607, |
|
"learning_rate": 8.770730841615827e-06, |
|
"loss": 0.1445, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 2.5136804942630184, |
|
"grad_norm": 0.9697606827791236, |
|
"learning_rate": 8.766970441003942e-06, |
|
"loss": 0.167, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 2.5172109443954103, |
|
"grad_norm": 0.8387174374399691, |
|
"learning_rate": 8.7632051063004e-06, |
|
"loss": 0.1585, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 2.5207413945278025, |
|
"grad_norm": 0.6631550381542344, |
|
"learning_rate": 8.75943484243715e-06, |
|
"loss": 0.1428, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 2.524271844660194, |
|
"grad_norm": 0.7695360425616815, |
|
"learning_rate": 8.755659654352599e-06, |
|
"loss": 0.1463, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 2.5278022947925862, |
|
"grad_norm": 0.7011902192003712, |
|
"learning_rate": 8.751879546991601e-06, |
|
"loss": 0.1406, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 2.531332744924978, |
|
"grad_norm": 0.7991566673347344, |
|
"learning_rate": 8.748094525305457e-06, |
|
"loss": 0.1545, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 2.53486319505737, |
|
"grad_norm": 0.8486318435641594, |
|
"learning_rate": 8.744304594251898e-06, |
|
"loss": 0.1565, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 2.5383936451897617, |
|
"grad_norm": 0.6882492324618286, |
|
"learning_rate": 8.740509758795094e-06, |
|
"loss": 0.141, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 2.5419240953221536, |
|
"grad_norm": 0.7803012919767831, |
|
"learning_rate": 8.736710023905634e-06, |
|
"loss": 0.1671, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.5454545454545454, |
|
"grad_norm": 1.0460659891563937, |
|
"learning_rate": 8.732905394560526e-06, |
|
"loss": 0.1662, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 2.5489849955869373, |
|
"grad_norm": 0.8104918877045203, |
|
"learning_rate": 8.729095875743187e-06, |
|
"loss": 0.1707, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 2.552515445719329, |
|
"grad_norm": 0.822076128070119, |
|
"learning_rate": 8.725281472443442e-06, |
|
"loss": 0.1434, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 2.556045895851721, |
|
"grad_norm": 0.8428018638045196, |
|
"learning_rate": 8.72146218965751e-06, |
|
"loss": 0.1452, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 2.559576345984113, |
|
"grad_norm": 0.6860938138056261, |
|
"learning_rate": 8.717638032388002e-06, |
|
"loss": 0.1452, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 2.5631067961165046, |
|
"grad_norm": 0.6679034911744651, |
|
"learning_rate": 8.713809005643919e-06, |
|
"loss": 0.1415, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 2.566637246248897, |
|
"grad_norm": 0.8156715771369578, |
|
"learning_rate": 8.709975114440632e-06, |
|
"loss": 0.1664, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 2.5701676963812887, |
|
"grad_norm": 0.950713499966215, |
|
"learning_rate": 8.706136363799888e-06, |
|
"loss": 0.1579, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 2.5736981465136806, |
|
"grad_norm": 0.6736037466099428, |
|
"learning_rate": 8.7022927587498e-06, |
|
"loss": 0.1376, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 2.5772285966460724, |
|
"grad_norm": 0.8007987125298717, |
|
"learning_rate": 8.698444304324837e-06, |
|
"loss": 0.1604, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.5807590467784642, |
|
"grad_norm": 0.7909053805653098, |
|
"learning_rate": 8.694591005565818e-06, |
|
"loss": 0.1525, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 2.584289496910856, |
|
"grad_norm": 0.7510519443516507, |
|
"learning_rate": 8.690732867519915e-06, |
|
"loss": 0.152, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 2.587819947043248, |
|
"grad_norm": 0.7070326571398531, |
|
"learning_rate": 8.686869895240631e-06, |
|
"loss": 0.1478, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 2.5913503971756398, |
|
"grad_norm": 0.7843881178771545, |
|
"learning_rate": 8.683002093787805e-06, |
|
"loss": 0.155, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 2.5948808473080316, |
|
"grad_norm": 0.74442690372609, |
|
"learning_rate": 8.679129468227601e-06, |
|
"loss": 0.1742, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 2.598411297440424, |
|
"grad_norm": 0.8873243162061539, |
|
"learning_rate": 8.675252023632497e-06, |
|
"loss": 0.1433, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 2.6019417475728153, |
|
"grad_norm": 1.0558814263186358, |
|
"learning_rate": 8.671369765081291e-06, |
|
"loss": 0.1383, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 2.6054721977052075, |
|
"grad_norm": 0.9661052979159337, |
|
"learning_rate": 8.66748269765908e-06, |
|
"loss": 0.1519, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 2.6090026478375994, |
|
"grad_norm": 0.7940230362582856, |
|
"learning_rate": 8.663590826457264e-06, |
|
"loss": 0.1449, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 2.6125330979699912, |
|
"grad_norm": 0.7712249737573764, |
|
"learning_rate": 8.659694156573533e-06, |
|
"loss": 0.1567, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.616063548102383, |
|
"grad_norm": 0.7354585280144065, |
|
"learning_rate": 8.655792693111864e-06, |
|
"loss": 0.147, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 2.619593998234775, |
|
"grad_norm": 0.7686478452231044, |
|
"learning_rate": 8.651886441182509e-06, |
|
"loss": 0.1462, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 2.6231244483671667, |
|
"grad_norm": 0.9615765356144627, |
|
"learning_rate": 8.647975405901996e-06, |
|
"loss": 0.1598, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 2.6266548984995586, |
|
"grad_norm": 0.7899472619604829, |
|
"learning_rate": 8.644059592393117e-06, |
|
"loss": 0.1473, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 2.6301853486319504, |
|
"grad_norm": 0.8716879941644811, |
|
"learning_rate": 8.640139005784924e-06, |
|
"loss": 0.1512, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 2.6337157987643423, |
|
"grad_norm": 0.9497477417594384, |
|
"learning_rate": 8.636213651212719e-06, |
|
"loss": 0.1501, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 2.6372462488967345, |
|
"grad_norm": 0.8255927742841409, |
|
"learning_rate": 8.632283533818048e-06, |
|
"loss": 0.147, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 2.6407766990291264, |
|
"grad_norm": 0.6589014957846149, |
|
"learning_rate": 8.628348658748702e-06, |
|
"loss": 0.146, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 2.644307149161518, |
|
"grad_norm": 0.7383869029734058, |
|
"learning_rate": 8.624409031158696e-06, |
|
"loss": 0.1444, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 2.64783759929391, |
|
"grad_norm": 0.8754969443335403, |
|
"learning_rate": 8.620464656208272e-06, |
|
"loss": 0.1533, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.651368049426302, |
|
"grad_norm": 0.8093544382322876, |
|
"learning_rate": 8.616515539063894e-06, |
|
"loss": 0.1466, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 2.6548984995586937, |
|
"grad_norm": 0.8094232555314647, |
|
"learning_rate": 8.612561684898235e-06, |
|
"loss": 0.1414, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 2.6584289496910856, |
|
"grad_norm": 0.9276446354153174, |
|
"learning_rate": 8.60860309889017e-06, |
|
"loss": 0.1608, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 2.6619593998234774, |
|
"grad_norm": 0.8670503330866877, |
|
"learning_rate": 8.604639786224778e-06, |
|
"loss": 0.148, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 2.6654898499558692, |
|
"grad_norm": 0.9236407101629923, |
|
"learning_rate": 8.600671752093323e-06, |
|
"loss": 0.1465, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 2.6690203000882615, |
|
"grad_norm": 0.7947533360687092, |
|
"learning_rate": 8.596699001693257e-06, |
|
"loss": 0.1382, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 2.672550750220653, |
|
"grad_norm": 0.9401531663981636, |
|
"learning_rate": 8.592721540228207e-06, |
|
"loss": 0.1545, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 2.676081200353045, |
|
"grad_norm": 0.6849210881232853, |
|
"learning_rate": 8.588739372907974e-06, |
|
"loss": 0.1422, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 2.679611650485437, |
|
"grad_norm": 0.9020222283025484, |
|
"learning_rate": 8.584752504948522e-06, |
|
"loss": 0.148, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 2.683142100617829, |
|
"grad_norm": 0.763357957155723, |
|
"learning_rate": 8.580760941571968e-06, |
|
"loss": 0.1387, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.6866725507502207, |
|
"grad_norm": 0.8677207545705664, |
|
"learning_rate": 8.576764688006583e-06, |
|
"loss": 0.1694, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 2.6902030008826125, |
|
"grad_norm": 0.7387471803373706, |
|
"learning_rate": 8.572763749486784e-06, |
|
"loss": 0.1507, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 2.6937334510150044, |
|
"grad_norm": 0.6986240692420573, |
|
"learning_rate": 8.568758131253115e-06, |
|
"loss": 0.1632, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 2.6972639011473962, |
|
"grad_norm": 0.9057220759646399, |
|
"learning_rate": 8.564747838552262e-06, |
|
"loss": 0.1607, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 2.700794351279788, |
|
"grad_norm": 0.8635819293383579, |
|
"learning_rate": 8.560732876637023e-06, |
|
"loss": 0.1446, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 2.70432480141218, |
|
"grad_norm": 0.773759396974051, |
|
"learning_rate": 8.55671325076632e-06, |
|
"loss": 0.152, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 2.707855251544572, |
|
"grad_norm": 0.7726341748367627, |
|
"learning_rate": 8.552688966205176e-06, |
|
"loss": 0.1507, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 2.7113857016769636, |
|
"grad_norm": 0.8172707078015327, |
|
"learning_rate": 8.548660028224724e-06, |
|
"loss": 0.1624, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 2.714916151809356, |
|
"grad_norm": 0.7096834505741385, |
|
"learning_rate": 8.544626442102188e-06, |
|
"loss": 0.1451, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 2.7184466019417477, |
|
"grad_norm": 0.7394117544451063, |
|
"learning_rate": 8.54058821312088e-06, |
|
"loss": 0.163, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.7219770520741395, |
|
"grad_norm": 0.9101784701617734, |
|
"learning_rate": 8.536545346570195e-06, |
|
"loss": 0.1597, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 2.7255075022065314, |
|
"grad_norm": 0.8528637952954311, |
|
"learning_rate": 8.532497847745602e-06, |
|
"loss": 0.1455, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 2.729037952338923, |
|
"grad_norm": 0.7128839210498725, |
|
"learning_rate": 8.528445721948637e-06, |
|
"loss": 0.1562, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 2.732568402471315, |
|
"grad_norm": 0.7514251001247261, |
|
"learning_rate": 8.524388974486896e-06, |
|
"loss": 0.1787, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 2.736098852603707, |
|
"grad_norm": 0.8717097957173104, |
|
"learning_rate": 8.520327610674029e-06, |
|
"loss": 0.1686, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 2.7396293027360987, |
|
"grad_norm": 0.8134061335902161, |
|
"learning_rate": 8.516261635829734e-06, |
|
"loss": 0.1444, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 2.7431597528684906, |
|
"grad_norm": 0.8130838438482634, |
|
"learning_rate": 8.512191055279749e-06, |
|
"loss": 0.1621, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 2.746690203000883, |
|
"grad_norm": 0.7609974982707016, |
|
"learning_rate": 8.50811587435584e-06, |
|
"loss": 0.1639, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 2.7502206531332742, |
|
"grad_norm": 0.8656038452478626, |
|
"learning_rate": 8.504036098395805e-06, |
|
"loss": 0.1584, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 2.7537511032656665, |
|
"grad_norm": 0.7488569534666748, |
|
"learning_rate": 8.499951732743457e-06, |
|
"loss": 0.148, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.7572815533980584, |
|
"grad_norm": 0.77448853187229, |
|
"learning_rate": 8.495862782748619e-06, |
|
"loss": 0.1406, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 2.76081200353045, |
|
"grad_norm": 0.6757102359665434, |
|
"learning_rate": 8.491769253767126e-06, |
|
"loss": 0.1301, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 2.764342453662842, |
|
"grad_norm": 0.7775212534985085, |
|
"learning_rate": 8.487671151160801e-06, |
|
"loss": 0.1595, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 2.767872903795234, |
|
"grad_norm": 0.7117920982524355, |
|
"learning_rate": 8.483568480297467e-06, |
|
"loss": 0.1378, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 2.7714033539276257, |
|
"grad_norm": 0.7937162920192259, |
|
"learning_rate": 8.479461246550922e-06, |
|
"loss": 0.152, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 2.7749338040600176, |
|
"grad_norm": 0.7164290770402717, |
|
"learning_rate": 8.475349455300947e-06, |
|
"loss": 0.1249, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 2.7784642541924094, |
|
"grad_norm": 0.7347153157580293, |
|
"learning_rate": 8.471233111933291e-06, |
|
"loss": 0.169, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 2.7819947043248012, |
|
"grad_norm": 0.6345954182380398, |
|
"learning_rate": 8.467112221839664e-06, |
|
"loss": 0.1476, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 2.7855251544571935, |
|
"grad_norm": 0.7710274326642901, |
|
"learning_rate": 8.462986790417733e-06, |
|
"loss": 0.1266, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 2.789055604589585, |
|
"grad_norm": 0.7730009077854115, |
|
"learning_rate": 8.458856823071111e-06, |
|
"loss": 0.1347, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.792586054721977, |
|
"grad_norm": 0.9320951665857266, |
|
"learning_rate": 8.454722325209354e-06, |
|
"loss": 0.1544, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 2.796116504854369, |
|
"grad_norm": 0.854024287144252, |
|
"learning_rate": 8.450583302247954e-06, |
|
"loss": 0.1776, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 2.799646954986761, |
|
"grad_norm": 0.8267045018857162, |
|
"learning_rate": 8.44643975960833e-06, |
|
"loss": 0.1805, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 2.8031774051191527, |
|
"grad_norm": 0.8689865555638648, |
|
"learning_rate": 8.442291702717814e-06, |
|
"loss": 0.1643, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 2.8067078552515445, |
|
"grad_norm": 0.78855645463223, |
|
"learning_rate": 8.43813913700966e-06, |
|
"loss": 0.1663, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 2.8102383053839364, |
|
"grad_norm": 0.8423324894257369, |
|
"learning_rate": 8.433982067923021e-06, |
|
"loss": 0.1539, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 2.813768755516328, |
|
"grad_norm": 0.6841465341626565, |
|
"learning_rate": 8.429820500902957e-06, |
|
"loss": 0.1707, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 2.81729920564872, |
|
"grad_norm": 0.637410884505517, |
|
"learning_rate": 8.425654441400408e-06, |
|
"loss": 0.1631, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 2.820829655781112, |
|
"grad_norm": 0.7435035501128142, |
|
"learning_rate": 8.421483894872208e-06, |
|
"loss": 0.1518, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 2.824360105913504, |
|
"grad_norm": 0.7390502021511453, |
|
"learning_rate": 8.417308866781065e-06, |
|
"loss": 0.1513, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.827890556045896, |
|
"grad_norm": 0.7074112942115228, |
|
"learning_rate": 8.413129362595555e-06, |
|
"loss": 0.1626, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 2.831421006178288, |
|
"grad_norm": 0.8297265070014697, |
|
"learning_rate": 8.408945387790121e-06, |
|
"loss": 0.1447, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 2.8349514563106797, |
|
"grad_norm": 0.6691649963937963, |
|
"learning_rate": 8.404756947845056e-06, |
|
"loss": 0.1453, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 2.8384819064430715, |
|
"grad_norm": 0.7275626975495146, |
|
"learning_rate": 8.40056404824651e-06, |
|
"loss": 0.162, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 2.8420123565754634, |
|
"grad_norm": 0.9804621444439445, |
|
"learning_rate": 8.396366694486466e-06, |
|
"loss": 0.1711, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 2.845542806707855, |
|
"grad_norm": 0.7841205320098567, |
|
"learning_rate": 8.392164892062746e-06, |
|
"loss": 0.1682, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 2.849073256840247, |
|
"grad_norm": 0.7259305687219134, |
|
"learning_rate": 8.387958646478998e-06, |
|
"loss": 0.1349, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 2.852603706972639, |
|
"grad_norm": 0.6285666615213146, |
|
"learning_rate": 8.383747963244688e-06, |
|
"loss": 0.1622, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 2.856134157105031, |
|
"grad_norm": 0.8191314538526839, |
|
"learning_rate": 8.379532847875099e-06, |
|
"loss": 0.1765, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 2.8596646072374226, |
|
"grad_norm": 0.7451704716031862, |
|
"learning_rate": 8.375313305891312e-06, |
|
"loss": 0.1451, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.863195057369815, |
|
"grad_norm": 0.6744145077220661, |
|
"learning_rate": 8.371089342820216e-06, |
|
"loss": 0.1643, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 2.8667255075022067, |
|
"grad_norm": 0.6887811045404645, |
|
"learning_rate": 8.366860964194479e-06, |
|
"loss": 0.1522, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 2.8702559576345985, |
|
"grad_norm": 0.63573464613807, |
|
"learning_rate": 8.362628175552566e-06, |
|
"loss": 0.1511, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 2.8737864077669903, |
|
"grad_norm": 0.6897759046360326, |
|
"learning_rate": 8.358390982438706e-06, |
|
"loss": 0.1434, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 2.877316857899382, |
|
"grad_norm": 0.7279048382426626, |
|
"learning_rate": 8.354149390402904e-06, |
|
"loss": 0.1351, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 2.880847308031774, |
|
"grad_norm": 0.7458563780785276, |
|
"learning_rate": 8.349903405000926e-06, |
|
"loss": 0.1517, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 2.884377758164166, |
|
"grad_norm": 0.7101890658117982, |
|
"learning_rate": 8.345653031794292e-06, |
|
"loss": 0.1462, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 2.8879082082965577, |
|
"grad_norm": 0.7742297467061107, |
|
"learning_rate": 8.341398276350268e-06, |
|
"loss": 0.1555, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 2.8914386584289495, |
|
"grad_norm": 0.7438278623504806, |
|
"learning_rate": 8.337139144241864e-06, |
|
"loss": 0.1396, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 2.894969108561342, |
|
"grad_norm": 0.724140366387057, |
|
"learning_rate": 8.332875641047817e-06, |
|
"loss": 0.1361, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.898499558693733, |
|
"grad_norm": 0.7091482359305543, |
|
"learning_rate": 8.328607772352593e-06, |
|
"loss": 0.1477, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 2.9020300088261255, |
|
"grad_norm": 0.714577592926291, |
|
"learning_rate": 8.324335543746377e-06, |
|
"loss": 0.155, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 2.9055604589585173, |
|
"grad_norm": 0.6821542598622823, |
|
"learning_rate": 8.32005896082506e-06, |
|
"loss": 0.1449, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 2.909090909090909, |
|
"grad_norm": 0.7674294440664721, |
|
"learning_rate": 8.31577802919024e-06, |
|
"loss": 0.162, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 2.912621359223301, |
|
"grad_norm": 0.6635918469316766, |
|
"learning_rate": 8.311492754449213e-06, |
|
"loss": 0.1445, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 2.916151809355693, |
|
"grad_norm": 0.7039005991376192, |
|
"learning_rate": 8.307203142214959e-06, |
|
"loss": 0.1473, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 2.9196822594880847, |
|
"grad_norm": 0.7210854795121673, |
|
"learning_rate": 8.302909198106144e-06, |
|
"loss": 0.1431, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 2.9232127096204765, |
|
"grad_norm": 0.8804727621267001, |
|
"learning_rate": 8.298610927747104e-06, |
|
"loss": 0.1514, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 2.9267431597528684, |
|
"grad_norm": 0.6567640224299616, |
|
"learning_rate": 8.294308336767843e-06, |
|
"loss": 0.1356, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 2.93027360988526, |
|
"grad_norm": 0.7266522819100054, |
|
"learning_rate": 8.290001430804026e-06, |
|
"loss": 0.1515, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.9338040600176525, |
|
"grad_norm": 0.7006333424868092, |
|
"learning_rate": 8.285690215496966e-06, |
|
"loss": 0.1518, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 2.937334510150044, |
|
"grad_norm": 1.0457753137928258, |
|
"learning_rate": 8.281374696493628e-06, |
|
"loss": 0.1623, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 2.940864960282436, |
|
"grad_norm": 0.7835519475042352, |
|
"learning_rate": 8.277054879446603e-06, |
|
"loss": 0.1512, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 2.944395410414828, |
|
"grad_norm": 0.694852963337788, |
|
"learning_rate": 8.27273077001412e-06, |
|
"loss": 0.1408, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 2.94792586054722, |
|
"grad_norm": 0.6914754193933184, |
|
"learning_rate": 8.26840237386003e-06, |
|
"loss": 0.1553, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 2.9514563106796117, |
|
"grad_norm": 0.7640107634127185, |
|
"learning_rate": 8.264069696653795e-06, |
|
"loss": 0.1563, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 2.9549867608120035, |
|
"grad_norm": 0.7031909948366248, |
|
"learning_rate": 8.25973274407049e-06, |
|
"loss": 0.1513, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 2.9585172109443953, |
|
"grad_norm": 0.7212903191170139, |
|
"learning_rate": 8.255391521790782e-06, |
|
"loss": 0.1452, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 2.962047661076787, |
|
"grad_norm": 0.9288624940870822, |
|
"learning_rate": 8.251046035500937e-06, |
|
"loss": 0.1611, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 2.965578111209179, |
|
"grad_norm": 0.8307453353673667, |
|
"learning_rate": 8.246696290892804e-06, |
|
"loss": 0.1511, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.969108561341571, |
|
"grad_norm": 0.7346059651517513, |
|
"learning_rate": 8.24234229366381e-06, |
|
"loss": 0.1645, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 2.972639011473963, |
|
"grad_norm": 0.8211385940924918, |
|
"learning_rate": 8.237984049516952e-06, |
|
"loss": 0.1576, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 2.9761694616063545, |
|
"grad_norm": 0.6365132434025104, |
|
"learning_rate": 8.23362156416079e-06, |
|
"loss": 0.1228, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 2.979699911738747, |
|
"grad_norm": 0.8824809696079167, |
|
"learning_rate": 8.229254843309438e-06, |
|
"loss": 0.1509, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 2.9832303618711387, |
|
"grad_norm": 0.8601855800826128, |
|
"learning_rate": 8.22488389268256e-06, |
|
"loss": 0.1615, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 2.9867608120035305, |
|
"grad_norm": 0.731123229093688, |
|
"learning_rate": 8.220508718005359e-06, |
|
"loss": 0.1378, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 2.9902912621359223, |
|
"grad_norm": 0.6227120430460796, |
|
"learning_rate": 8.21612932500857e-06, |
|
"loss": 0.1399, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 2.993821712268314, |
|
"grad_norm": 0.7545771750137871, |
|
"learning_rate": 8.211745719428456e-06, |
|
"loss": 0.1701, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 2.997352162400706, |
|
"grad_norm": 0.8046466791224096, |
|
"learning_rate": 8.207357907006794e-06, |
|
"loss": 0.1549, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 3.000882612533098, |
|
"grad_norm": 0.7642023797697144, |
|
"learning_rate": 8.202965893490877e-06, |
|
"loss": 0.1368, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 3.0044130626654897, |
|
"grad_norm": 0.6716017754172908, |
|
"learning_rate": 8.198569684633492e-06, |
|
"loss": 0.1162, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 3.0079435127978815, |
|
"grad_norm": 0.6431034665161904, |
|
"learning_rate": 8.194169286192928e-06, |
|
"loss": 0.1195, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 3.011473962930274, |
|
"grad_norm": 0.7143235899902817, |
|
"learning_rate": 8.18976470393296e-06, |
|
"loss": 0.112, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 3.0150044130626656, |
|
"grad_norm": 0.698231889557549, |
|
"learning_rate": 8.185355943622845e-06, |
|
"loss": 0.1234, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 3.0185348631950575, |
|
"grad_norm": 0.9000098025363069, |
|
"learning_rate": 8.180943011037306e-06, |
|
"loss": 0.1249, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 3.0220653133274493, |
|
"grad_norm": 0.8015546761444414, |
|
"learning_rate": 8.176525911956539e-06, |
|
"loss": 0.1245, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 3.025595763459841, |
|
"grad_norm": 0.8094451013988107, |
|
"learning_rate": 8.172104652166194e-06, |
|
"loss": 0.1018, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 3.029126213592233, |
|
"grad_norm": 0.720799250577229, |
|
"learning_rate": 8.167679237457368e-06, |
|
"loss": 0.0992, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 3.032656663724625, |
|
"grad_norm": 0.8000553496291071, |
|
"learning_rate": 8.163249673626603e-06, |
|
"loss": 0.1194, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 3.0361871138570167, |
|
"grad_norm": 0.8669635173424244, |
|
"learning_rate": 8.158815966475877e-06, |
|
"loss": 0.0816, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 3.0397175639894085, |
|
"grad_norm": 0.9025742229651821, |
|
"learning_rate": 8.154378121812593e-06, |
|
"loss": 0.1018, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 3.0432480141218003, |
|
"grad_norm": 0.9845870522804518, |
|
"learning_rate": 8.149936145449575e-06, |
|
"loss": 0.1251, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 3.046778464254192, |
|
"grad_norm": 1.0256068011823871, |
|
"learning_rate": 8.145490043205056e-06, |
|
"loss": 0.1239, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 3.0503089143865845, |
|
"grad_norm": 1.148156571965961, |
|
"learning_rate": 8.141039820902676e-06, |
|
"loss": 0.1104, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 3.0538393645189763, |
|
"grad_norm": 1.2193696303353918, |
|
"learning_rate": 8.136585484371471e-06, |
|
"loss": 0.1152, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 3.057369814651368, |
|
"grad_norm": 1.2915877393106403, |
|
"learning_rate": 8.132127039445864e-06, |
|
"loss": 0.118, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 3.06090026478376, |
|
"grad_norm": 1.2149067405554672, |
|
"learning_rate": 8.127664491965662e-06, |
|
"loss": 0.1142, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 3.064430714916152, |
|
"grad_norm": 0.990566681159559, |
|
"learning_rate": 8.123197847776043e-06, |
|
"loss": 0.1078, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 3.0679611650485437, |
|
"grad_norm": 0.9565804354564945, |
|
"learning_rate": 8.118727112727556e-06, |
|
"loss": 0.1135, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 3.0714916151809355, |
|
"grad_norm": 1.0200728098865612, |
|
"learning_rate": 8.1142522926761e-06, |
|
"loss": 0.1195, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 3.0750220653133273, |
|
"grad_norm": 1.0243439506618488, |
|
"learning_rate": 8.109773393482934e-06, |
|
"loss": 0.1024, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 3.078552515445719, |
|
"grad_norm": 1.0049145977279004, |
|
"learning_rate": 8.105290421014653e-06, |
|
"loss": 0.1076, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 3.082082965578111, |
|
"grad_norm": 0.8112529372591591, |
|
"learning_rate": 8.10080338114319e-06, |
|
"loss": 0.0982, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 3.0856134157105033, |
|
"grad_norm": 1.0002477688608336, |
|
"learning_rate": 8.096312279745804e-06, |
|
"loss": 0.13, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 3.089143865842895, |
|
"grad_norm": 1.0623544903602675, |
|
"learning_rate": 8.09181712270508e-06, |
|
"loss": 0.1025, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 3.092674315975287, |
|
"grad_norm": 1.0410125624967994, |
|
"learning_rate": 8.087317915908904e-06, |
|
"loss": 0.1063, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 3.096204766107679, |
|
"grad_norm": 1.0118590449093119, |
|
"learning_rate": 8.082814665250476e-06, |
|
"loss": 0.0954, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 3.0997352162400706, |
|
"grad_norm": 0.9546611695399589, |
|
"learning_rate": 8.078307376628292e-06, |
|
"loss": 0.1245, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 3.1032656663724625, |
|
"grad_norm": 0.9269391315446083, |
|
"learning_rate": 8.073796055946133e-06, |
|
"loss": 0.1174, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 3.1067961165048543, |
|
"grad_norm": 1.0305117514176123, |
|
"learning_rate": 8.06928070911306e-06, |
|
"loss": 0.1026, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 3.110326566637246, |
|
"grad_norm": 1.208839666085655, |
|
"learning_rate": 8.064761342043415e-06, |
|
"loss": 0.1248, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 3.113857016769638, |
|
"grad_norm": 0.952513604276628, |
|
"learning_rate": 8.060237960656799e-06, |
|
"loss": 0.1085, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 3.11738746690203, |
|
"grad_norm": 0.8730540650291831, |
|
"learning_rate": 8.055710570878076e-06, |
|
"loss": 0.1134, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 3.120917917034422, |
|
"grad_norm": 1.006482604780559, |
|
"learning_rate": 8.051179178637356e-06, |
|
"loss": 0.1049, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 3.124448367166814, |
|
"grad_norm": 0.9112291303575337, |
|
"learning_rate": 8.046643789869991e-06, |
|
"loss": 0.108, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 3.127978817299206, |
|
"grad_norm": 1.0693444544873563, |
|
"learning_rate": 8.042104410516576e-06, |
|
"loss": 0.1313, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 3.1315092674315976, |
|
"grad_norm": 1.0053001607105967, |
|
"learning_rate": 8.037561046522923e-06, |
|
"loss": 0.1142, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 3.1350397175639895, |
|
"grad_norm": 0.9933955602184302, |
|
"learning_rate": 8.033013703840067e-06, |
|
"loss": 0.1278, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 3.1385701676963813, |
|
"grad_norm": 1.0821161416075462, |
|
"learning_rate": 8.028462388424258e-06, |
|
"loss": 0.1213, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 3.142100617828773, |
|
"grad_norm": 1.0448661581558534, |
|
"learning_rate": 8.023907106236944e-06, |
|
"loss": 0.1016, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 3.145631067961165, |
|
"grad_norm": 1.1354463844063831, |
|
"learning_rate": 8.019347863244773e-06, |
|
"loss": 0.1198, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 3.149161518093557, |
|
"grad_norm": 1.008070141194459, |
|
"learning_rate": 8.014784665419577e-06, |
|
"loss": 0.1252, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 3.1526919682259487, |
|
"grad_norm": 1.06009230334142, |
|
"learning_rate": 8.01021751873837e-06, |
|
"loss": 0.1286, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 3.1562224183583405, |
|
"grad_norm": 1.138468713844925, |
|
"learning_rate": 8.005646429183342e-06, |
|
"loss": 0.1129, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 3.159752868490733, |
|
"grad_norm": 1.2328861278363776, |
|
"learning_rate": 8.001071402741843e-06, |
|
"loss": 0.1235, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 3.1632833186231246, |
|
"grad_norm": 1.0853132184834922, |
|
"learning_rate": 7.996492445406377e-06, |
|
"loss": 0.1282, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 3.1668137687555165, |
|
"grad_norm": 0.933433925231955, |
|
"learning_rate": 7.991909563174605e-06, |
|
"loss": 0.1325, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 3.1703442188879083, |
|
"grad_norm": 0.8319438313846836, |
|
"learning_rate": 7.987322762049323e-06, |
|
"loss": 0.0987, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 3.1738746690203, |
|
"grad_norm": 0.9872815349839481, |
|
"learning_rate": 7.982732048038462e-06, |
|
"loss": 0.1195, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 3.177405119152692, |
|
"grad_norm": 0.9612402010886701, |
|
"learning_rate": 7.978137427155078e-06, |
|
"loss": 0.1258, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 3.180935569285084, |
|
"grad_norm": 0.8927263036280142, |
|
"learning_rate": 7.973538905417344e-06, |
|
"loss": 0.1072, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 3.1844660194174756, |
|
"grad_norm": 0.8841286752626785, |
|
"learning_rate": 7.968936488848541e-06, |
|
"loss": 0.1047, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 3.1879964695498675, |
|
"grad_norm": 0.912445757407591, |
|
"learning_rate": 7.964330183477057e-06, |
|
"loss": 0.1276, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 3.1915269196822593, |
|
"grad_norm": 1.059803669034267, |
|
"learning_rate": 7.959719995336364e-06, |
|
"loss": 0.1032, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 3.195057369814651, |
|
"grad_norm": 0.9513868377440023, |
|
"learning_rate": 7.955105930465032e-06, |
|
"loss": 0.0911, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 3.1985878199470434, |
|
"grad_norm": 0.8797745669418451, |
|
"learning_rate": 7.9504879949067e-06, |
|
"loss": 0.1143, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 3.2021182700794353, |
|
"grad_norm": 1.0029598114248315, |
|
"learning_rate": 7.945866194710079e-06, |
|
"loss": 0.1161, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 3.205648720211827, |
|
"grad_norm": 1.1319418617063524, |
|
"learning_rate": 7.941240535928942e-06, |
|
"loss": 0.1268, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 3.209179170344219, |
|
"grad_norm": 0.9958247294998045, |
|
"learning_rate": 7.936611024622117e-06, |
|
"loss": 0.1074, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 3.212709620476611, |
|
"grad_norm": 1.0922819941075517, |
|
"learning_rate": 7.93197766685348e-06, |
|
"loss": 0.1179, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 3.2162400706090026, |
|
"grad_norm": 1.136941227912527, |
|
"learning_rate": 7.92734046869194e-06, |
|
"loss": 0.1331, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 3.2197705207413945, |
|
"grad_norm": 1.0047937727286098, |
|
"learning_rate": 7.922699436211438e-06, |
|
"loss": 0.1104, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 3.2233009708737863, |
|
"grad_norm": 0.9964788630911344, |
|
"learning_rate": 7.918054575490943e-06, |
|
"loss": 0.1252, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 3.226831421006178, |
|
"grad_norm": 1.0926246384683764, |
|
"learning_rate": 7.913405892614433e-06, |
|
"loss": 0.1044, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 3.23036187113857, |
|
"grad_norm": 0.9688331888143029, |
|
"learning_rate": 7.908753393670891e-06, |
|
"loss": 0.0945, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 3.233892321270962, |
|
"grad_norm": 1.111353673933071, |
|
"learning_rate": 7.904097084754301e-06, |
|
"loss": 0.1148, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 3.237422771403354, |
|
"grad_norm": 1.092326321596579, |
|
"learning_rate": 7.899436971963637e-06, |
|
"loss": 0.1276, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 3.240953221535746, |
|
"grad_norm": 0.7874168063997805, |
|
"learning_rate": 7.894773061402855e-06, |
|
"loss": 0.0932, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 3.244483671668138, |
|
"grad_norm": 0.9597684028837579, |
|
"learning_rate": 7.89010535918089e-06, |
|
"loss": 0.1158, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 3.2480141218005296, |
|
"grad_norm": 1.0816606744001607, |
|
"learning_rate": 7.885433871411634e-06, |
|
"loss": 0.1204, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 3.2515445719329215, |
|
"grad_norm": 1.035099273475842, |
|
"learning_rate": 7.880758604213945e-06, |
|
"loss": 0.125, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 3.2550750220653133, |
|
"grad_norm": 0.8982660262483314, |
|
"learning_rate": 7.876079563711631e-06, |
|
"loss": 0.104, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 3.258605472197705, |
|
"grad_norm": 1.0276050394216014, |
|
"learning_rate": 7.871396756033438e-06, |
|
"loss": 0.119, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 3.262135922330097, |
|
"grad_norm": 1.0128946832960732, |
|
"learning_rate": 7.866710187313048e-06, |
|
"loss": 0.118, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 3.265666372462489, |
|
"grad_norm": 0.8704316103506846, |
|
"learning_rate": 7.862019863689075e-06, |
|
"loss": 0.1139, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 3.269196822594881, |
|
"grad_norm": 0.9671125371629974, |
|
"learning_rate": 7.857325791305042e-06, |
|
"loss": 0.137, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 3.2727272727272725, |
|
"grad_norm": 0.868940825788469, |
|
"learning_rate": 7.852627976309388e-06, |
|
"loss": 0.1124, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 3.2762577228596648, |
|
"grad_norm": 0.9630053987036548, |
|
"learning_rate": 7.847926424855454e-06, |
|
"loss": 0.1253, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 3.2797881729920566, |
|
"grad_norm": 1.0029586796082006, |
|
"learning_rate": 7.843221143101474e-06, |
|
"loss": 0.1137, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 3.2833186231244484, |
|
"grad_norm": 0.8820085770728214, |
|
"learning_rate": 7.838512137210565e-06, |
|
"loss": 0.1255, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 3.2868490732568403, |
|
"grad_norm": 0.9828634895838252, |
|
"learning_rate": 7.833799413350732e-06, |
|
"loss": 0.1035, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 3.290379523389232, |
|
"grad_norm": 1.319511988574606, |
|
"learning_rate": 7.829082977694835e-06, |
|
"loss": 0.1325, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 3.293909973521624, |
|
"grad_norm": 1.2144177327586667, |
|
"learning_rate": 7.824362836420609e-06, |
|
"loss": 0.1403, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 3.297440423654016, |
|
"grad_norm": 1.0719443808897844, |
|
"learning_rate": 7.819638995710636e-06, |
|
"loss": 0.1185, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 3.3009708737864076, |
|
"grad_norm": 1.0518286795857918, |
|
"learning_rate": 7.814911461752346e-06, |
|
"loss": 0.1065, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 3.3045013239187995, |
|
"grad_norm": 1.024493911176814, |
|
"learning_rate": 7.810180240738004e-06, |
|
"loss": 0.1151, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 3.3080317740511918, |
|
"grad_norm": 1.127575182515073, |
|
"learning_rate": 7.805445338864709e-06, |
|
"loss": 0.1308, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 3.3115622241835836, |
|
"grad_norm": 0.9456157508829596, |
|
"learning_rate": 7.800706762334375e-06, |
|
"loss": 0.107, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 3.3150926743159754, |
|
"grad_norm": 0.9074545977721913, |
|
"learning_rate": 7.795964517353734e-06, |
|
"loss": 0.1138, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 3.3186231244483673, |
|
"grad_norm": 1.094922914297947, |
|
"learning_rate": 7.791218610134324e-06, |
|
"loss": 0.1349, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 3.322153574580759, |
|
"grad_norm": 1.133601456805787, |
|
"learning_rate": 7.786469046892475e-06, |
|
"loss": 0.1069, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 3.325684024713151, |
|
"grad_norm": 1.0182500924518352, |
|
"learning_rate": 7.781715833849311e-06, |
|
"loss": 0.1154, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 3.329214474845543, |
|
"grad_norm": 1.046201515291234, |
|
"learning_rate": 7.776958977230731e-06, |
|
"loss": 0.1251, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 3.3327449249779346, |
|
"grad_norm": 0.957386976461524, |
|
"learning_rate": 7.77219848326741e-06, |
|
"loss": 0.1101, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 3.3362753751103265, |
|
"grad_norm": 1.1740948292667477, |
|
"learning_rate": 7.76743435819479e-06, |
|
"loss": 0.1443, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 3.3398058252427183, |
|
"grad_norm": 1.850891607235183, |
|
"learning_rate": 7.76266660825306e-06, |
|
"loss": 0.1018, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 3.34333627537511, |
|
"grad_norm": 0.9097207671724479, |
|
"learning_rate": 7.757895239687168e-06, |
|
"loss": 0.1041, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 3.3468667255075024, |
|
"grad_norm": 0.9016475882306597, |
|
"learning_rate": 7.753120258746793e-06, |
|
"loss": 0.1151, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 3.3503971756398943, |
|
"grad_norm": 1.399326889021019, |
|
"learning_rate": 7.748341671686355e-06, |
|
"loss": 0.1361, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 3.353927625772286, |
|
"grad_norm": 0.8958929274360947, |
|
"learning_rate": 7.743559484764987e-06, |
|
"loss": 0.109, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 3.357458075904678, |
|
"grad_norm": 0.9661628430088748, |
|
"learning_rate": 7.73877370424654e-06, |
|
"loss": 0.1184, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 3.3609885260370698, |
|
"grad_norm": 1.0856106807644708, |
|
"learning_rate": 7.733984336399578e-06, |
|
"loss": 0.1258, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 3.3645189761694616, |
|
"grad_norm": 1.1339299007520192, |
|
"learning_rate": 7.729191387497357e-06, |
|
"loss": 0.1313, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 3.3680494263018534, |
|
"grad_norm": 0.9462346995224554, |
|
"learning_rate": 7.72439486381783e-06, |
|
"loss": 0.1299, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 3.3715798764342453, |
|
"grad_norm": 0.9773900040503245, |
|
"learning_rate": 7.719594771643623e-06, |
|
"loss": 0.115, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 3.375110326566637, |
|
"grad_norm": 0.8812615869761742, |
|
"learning_rate": 7.714791117262047e-06, |
|
"loss": 0.1178, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 3.378640776699029, |
|
"grad_norm": 0.872321741116879, |
|
"learning_rate": 7.709983906965071e-06, |
|
"loss": 0.1177, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 3.382171226831421, |
|
"grad_norm": 0.9590632048655139, |
|
"learning_rate": 7.705173147049326e-06, |
|
"loss": 0.1209, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 3.385701676963813, |
|
"grad_norm": 0.9774010811015693, |
|
"learning_rate": 7.700358843816087e-06, |
|
"loss": 0.0988, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 3.389232127096205, |
|
"grad_norm": 1.05084415330861, |
|
"learning_rate": 7.695541003571275e-06, |
|
"loss": 0.1281, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 3.3927625772285968, |
|
"grad_norm": 0.9815810438992207, |
|
"learning_rate": 7.690719632625447e-06, |
|
"loss": 0.1127, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 3.3962930273609886, |
|
"grad_norm": 0.9979507342808466, |
|
"learning_rate": 7.685894737293773e-06, |
|
"loss": 0.1224, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 3.3998234774933804, |
|
"grad_norm": 1.0098708954426872, |
|
"learning_rate": 7.681066323896053e-06, |
|
"loss": 0.1418, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 3.4033539276257723, |
|
"grad_norm": 0.8683560303898923, |
|
"learning_rate": 7.676234398756688e-06, |
|
"loss": 0.11, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 3.406884377758164, |
|
"grad_norm": 0.9354248879782321, |
|
"learning_rate": 7.671398968204675e-06, |
|
"loss": 0.1125, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 3.410414827890556, |
|
"grad_norm": 0.9688499014186751, |
|
"learning_rate": 7.66656003857361e-06, |
|
"loss": 0.1233, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 3.413945278022948, |
|
"grad_norm": 1.0624016948474533, |
|
"learning_rate": 7.66171761620167e-06, |
|
"loss": 0.1051, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 3.4174757281553396, |
|
"grad_norm": 0.9463189432354194, |
|
"learning_rate": 7.656871707431604e-06, |
|
"loss": 0.1171, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 3.4210061782877315, |
|
"grad_norm": 0.9894483338890546, |
|
"learning_rate": 7.65202231861073e-06, |
|
"loss": 0.1108, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 3.4245366284201237, |
|
"grad_norm": 0.9687362431205614, |
|
"learning_rate": 7.647169456090925e-06, |
|
"loss": 0.1116, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 3.4280670785525156, |
|
"grad_norm": 0.9557897471038936, |
|
"learning_rate": 7.642313126228618e-06, |
|
"loss": 0.1153, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 3.4315975286849074, |
|
"grad_norm": 1.0869572132117586, |
|
"learning_rate": 7.63745333538477e-06, |
|
"loss": 0.1236, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 3.4351279788172993, |
|
"grad_norm": 0.9605589326914042, |
|
"learning_rate": 7.632590089924885e-06, |
|
"loss": 0.1065, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 3.438658428949691, |
|
"grad_norm": 1.0323514330728734, |
|
"learning_rate": 7.627723396218988e-06, |
|
"loss": 0.1175, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 3.442188879082083, |
|
"grad_norm": 1.0118187975842183, |
|
"learning_rate": 7.622853260641623e-06, |
|
"loss": 0.1186, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 3.4457193292144748, |
|
"grad_norm": 1.024656799596295, |
|
"learning_rate": 7.61797968957184e-06, |
|
"loss": 0.1271, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 3.4492497793468666, |
|
"grad_norm": 0.8727047804672294, |
|
"learning_rate": 7.613102689393188e-06, |
|
"loss": 0.1144, |
|
"step": 977 |
|
}, |
|
{ |
|
"epoch": 3.4527802294792584, |
|
"grad_norm": 0.8394841490325925, |
|
"learning_rate": 7.608222266493712e-06, |
|
"loss": 0.1123, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 3.4563106796116507, |
|
"grad_norm": 1.1351511857664562, |
|
"learning_rate": 7.603338427265935e-06, |
|
"loss": 0.1012, |
|
"step": 979 |
|
}, |
|
{ |
|
"epoch": 3.459841129744042, |
|
"grad_norm": 1.128335444374776, |
|
"learning_rate": 7.598451178106857e-06, |
|
"loss": 0.1206, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 3.4633715798764344, |
|
"grad_norm": 0.8576208704238955, |
|
"learning_rate": 7.593560525417947e-06, |
|
"loss": 0.103, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 3.4669020300088262, |
|
"grad_norm": 1.1074290695297295, |
|
"learning_rate": 7.588666475605127e-06, |
|
"loss": 0.1254, |
|
"step": 982 |
|
}, |
|
{ |
|
"epoch": 3.470432480141218, |
|
"grad_norm": 1.1040068892498724, |
|
"learning_rate": 7.583769035078772e-06, |
|
"loss": 0.1143, |
|
"step": 983 |
|
}, |
|
{ |
|
"epoch": 3.47396293027361, |
|
"grad_norm": 1.0664311386837333, |
|
"learning_rate": 7.578868210253698e-06, |
|
"loss": 0.1353, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 3.4774933804060018, |
|
"grad_norm": 1.059615015944644, |
|
"learning_rate": 7.5739640075491546e-06, |
|
"loss": 0.1386, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 3.4810238305383936, |
|
"grad_norm": 0.9899439186497315, |
|
"learning_rate": 7.569056433388813e-06, |
|
"loss": 0.1258, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 3.4845542806707854, |
|
"grad_norm": 0.9224864093833172, |
|
"learning_rate": 7.5641454942007606e-06, |
|
"loss": 0.1225, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 3.4880847308031773, |
|
"grad_norm": 0.887546649860455, |
|
"learning_rate": 7.559231196417495e-06, |
|
"loss": 0.1249, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 3.491615180935569, |
|
"grad_norm": 0.9660667346820182, |
|
"learning_rate": 7.554313546475915e-06, |
|
"loss": 0.1163, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 3.4951456310679614, |
|
"grad_norm": 0.9707962317309367, |
|
"learning_rate": 7.5493925508173025e-06, |
|
"loss": 0.1321, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 3.4986760812003532, |
|
"grad_norm": 1.1180387890566459, |
|
"learning_rate": 7.5444682158873305e-06, |
|
"loss": 0.1221, |
|
"step": 991 |
|
}, |
|
{ |
|
"epoch": 3.502206531332745, |
|
"grad_norm": 0.8799016210637428, |
|
"learning_rate": 7.539540548136037e-06, |
|
"loss": 0.1371, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 3.505736981465137, |
|
"grad_norm": 1.0482875409664072, |
|
"learning_rate": 7.534609554017833e-06, |
|
"loss": 0.1175, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 3.5092674315975287, |
|
"grad_norm": 0.8991365777316297, |
|
"learning_rate": 7.529675239991483e-06, |
|
"loss": 0.1293, |
|
"step": 994 |
|
}, |
|
{ |
|
"epoch": 3.5127978817299206, |
|
"grad_norm": 1.0386653386530387, |
|
"learning_rate": 7.524737612520101e-06, |
|
"loss": 0.1229, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 3.5163283318623124, |
|
"grad_norm": 1.0711541879063016, |
|
"learning_rate": 7.5197966780711385e-06, |
|
"loss": 0.1201, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 3.5198587819947043, |
|
"grad_norm": 1.0318173451390964, |
|
"learning_rate": 7.5148524431163825e-06, |
|
"loss": 0.1316, |
|
"step": 997 |
|
}, |
|
{ |
|
"epoch": 3.523389232127096, |
|
"grad_norm": 0.984403787758931, |
|
"learning_rate": 7.509904914131941e-06, |
|
"loss": 0.1197, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 3.526919682259488, |
|
"grad_norm": 0.8522121035569924, |
|
"learning_rate": 7.504954097598238e-06, |
|
"loss": 0.1061, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 3.5304501323918798, |
|
"grad_norm": 0.849756117125948, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 0.1067, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 3.533980582524272, |
|
"grad_norm": 0.7751208269103894, |
|
"learning_rate": 7.495042627826257e-06, |
|
"loss": 0.1091, |
|
"step": 1001 |
|
}, |
|
{ |
|
"epoch": 3.537511032656664, |
|
"grad_norm": 0.9495031542830646, |
|
"learning_rate": 7.490081987570321e-06, |
|
"loss": 0.121, |
|
"step": 1002 |
|
}, |
|
{ |
|
"epoch": 3.5410414827890557, |
|
"grad_norm": 1.0312096390159498, |
|
"learning_rate": 7.48511808572979e-06, |
|
"loss": 0.1304, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 3.5445719329214476, |
|
"grad_norm": 1.054899906830368, |
|
"learning_rate": 7.480150928806534e-06, |
|
"loss": 0.1361, |
|
"step": 1004 |
|
}, |
|
{ |
|
"epoch": 3.5481023830538394, |
|
"grad_norm": 0.979527615798883, |
|
"learning_rate": 7.475180523306684e-06, |
|
"loss": 0.1277, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 3.5516328331862312, |
|
"grad_norm": 1.0015458182337016, |
|
"learning_rate": 7.470206875740627e-06, |
|
"loss": 0.1149, |
|
"step": 1006 |
|
}, |
|
{ |
|
"epoch": 3.555163283318623, |
|
"grad_norm": 1.0755807694863864, |
|
"learning_rate": 7.465229992622996e-06, |
|
"loss": 0.1178, |
|
"step": 1007 |
|
}, |
|
{ |
|
"epoch": 3.558693733451015, |
|
"grad_norm": 1.0077743388921196, |
|
"learning_rate": 7.4602498804726646e-06, |
|
"loss": 0.1267, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 3.5622241835834068, |
|
"grad_norm": 0.9760852883342952, |
|
"learning_rate": 7.455266545812733e-06, |
|
"loss": 0.1144, |
|
"step": 1009 |
|
}, |
|
{ |
|
"epoch": 3.565754633715799, |
|
"grad_norm": 0.9599179886281429, |
|
"learning_rate": 7.450279995170524e-06, |
|
"loss": 0.1233, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 3.5692850838481904, |
|
"grad_norm": 1.0506465130972065, |
|
"learning_rate": 7.445290235077571e-06, |
|
"loss": 0.1223, |
|
"step": 1011 |
|
}, |
|
{ |
|
"epoch": 3.5728155339805827, |
|
"grad_norm": 0.9904879926238966, |
|
"learning_rate": 7.440297272069615e-06, |
|
"loss": 0.1103, |
|
"step": 1012 |
|
}, |
|
{ |
|
"epoch": 3.5763459841129746, |
|
"grad_norm": 0.9373307300830205, |
|
"learning_rate": 7.435301112686588e-06, |
|
"loss": 0.1089, |
|
"step": 1013 |
|
}, |
|
{ |
|
"epoch": 3.5798764342453664, |
|
"grad_norm": 0.9642807996392302, |
|
"learning_rate": 7.430301763472612e-06, |
|
"loss": 0.1397, |
|
"step": 1014 |
|
}, |
|
{ |
|
"epoch": 3.5834068843777582, |
|
"grad_norm": 1.0798478829987526, |
|
"learning_rate": 7.425299230975982e-06, |
|
"loss": 0.1271, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 3.58693733451015, |
|
"grad_norm": 0.9757898668217716, |
|
"learning_rate": 7.420293521749169e-06, |
|
"loss": 0.1254, |
|
"step": 1016 |
|
}, |
|
{ |
|
"epoch": 3.590467784642542, |
|
"grad_norm": 0.7957246885082836, |
|
"learning_rate": 7.415284642348801e-06, |
|
"loss": 0.1024, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 3.5939982347749337, |
|
"grad_norm": 1.0499954601099537, |
|
"learning_rate": 7.410272599335662e-06, |
|
"loss": 0.1104, |
|
"step": 1018 |
|
}, |
|
{ |
|
"epoch": 3.5975286849073256, |
|
"grad_norm": 1.017468888424547, |
|
"learning_rate": 7.4052573992746735e-06, |
|
"loss": 0.1329, |
|
"step": 1019 |
|
}, |
|
{ |
|
"epoch": 3.6010591350397174, |
|
"grad_norm": 0.9050306794358908, |
|
"learning_rate": 7.400239048734899e-06, |
|
"loss": 0.1001, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 3.6045895851721097, |
|
"grad_norm": 0.9892172267357415, |
|
"learning_rate": 7.395217554289524e-06, |
|
"loss": 0.1116, |
|
"step": 1021 |
|
}, |
|
{ |
|
"epoch": 3.608120035304501, |
|
"grad_norm": 0.825558862940724, |
|
"learning_rate": 7.390192922515855e-06, |
|
"loss": 0.1178, |
|
"step": 1022 |
|
}, |
|
{ |
|
"epoch": 3.6116504854368934, |
|
"grad_norm": 1.0870666189939848, |
|
"learning_rate": 7.385165159995306e-06, |
|
"loss": 0.1188, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 3.615180935569285, |
|
"grad_norm": 0.9461150823847166, |
|
"learning_rate": 7.38013427331339e-06, |
|
"loss": 0.1387, |
|
"step": 1024 |
|
}, |
|
{ |
|
"epoch": 3.618711385701677, |
|
"grad_norm": 0.8931705317958296, |
|
"learning_rate": 7.3751002690597175e-06, |
|
"loss": 0.1145, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 3.622241835834069, |
|
"grad_norm": 1.0044315340520118, |
|
"learning_rate": 7.370063153827978e-06, |
|
"loss": 0.1311, |
|
"step": 1026 |
|
}, |
|
{ |
|
"epoch": 3.6257722859664607, |
|
"grad_norm": 0.9366772554201618, |
|
"learning_rate": 7.365022934215935e-06, |
|
"loss": 0.1347, |
|
"step": 1027 |
|
}, |
|
{ |
|
"epoch": 3.6293027360988526, |
|
"grad_norm": 1.0287287282798234, |
|
"learning_rate": 7.3599796168254256e-06, |
|
"loss": 0.12, |
|
"step": 1028 |
|
}, |
|
{ |
|
"epoch": 3.6328331862312444, |
|
"grad_norm": 1.2655633592932094, |
|
"learning_rate": 7.3549332082623336e-06, |
|
"loss": 0.1225, |
|
"step": 1029 |
|
}, |
|
{ |
|
"epoch": 3.6363636363636362, |
|
"grad_norm": 0.962932765361898, |
|
"learning_rate": 7.349883715136601e-06, |
|
"loss": 0.1163, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 3.639894086496028, |
|
"grad_norm": 1.1898739705783477, |
|
"learning_rate": 7.344831144062204e-06, |
|
"loss": 0.1342, |
|
"step": 1031 |
|
}, |
|
{ |
|
"epoch": 3.6434245366284204, |
|
"grad_norm": 0.9775960634954581, |
|
"learning_rate": 7.339775501657156e-06, |
|
"loss": 0.1159, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 3.6469549867608118, |
|
"grad_norm": 0.9985532411205827, |
|
"learning_rate": 7.334716794543487e-06, |
|
"loss": 0.1215, |
|
"step": 1033 |
|
}, |
|
{ |
|
"epoch": 3.650485436893204, |
|
"grad_norm": 0.8371099521051029, |
|
"learning_rate": 7.329655029347246e-06, |
|
"loss": 0.133, |
|
"step": 1034 |
|
}, |
|
{ |
|
"epoch": 3.654015887025596, |
|
"grad_norm": 0.8282699034669408, |
|
"learning_rate": 7.324590212698486e-06, |
|
"loss": 0.1128, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 3.6575463371579877, |
|
"grad_norm": 0.8892127258122363, |
|
"learning_rate": 7.319522351231256e-06, |
|
"loss": 0.1193, |
|
"step": 1036 |
|
}, |
|
{ |
|
"epoch": 3.6610767872903796, |
|
"grad_norm": 1.0611463535886227, |
|
"learning_rate": 7.314451451583593e-06, |
|
"loss": 0.1371, |
|
"step": 1037 |
|
}, |
|
{ |
|
"epoch": 3.6646072374227714, |
|
"grad_norm": 0.834905111221197, |
|
"learning_rate": 7.309377520397517e-06, |
|
"loss": 0.1198, |
|
"step": 1038 |
|
}, |
|
{ |
|
"epoch": 3.6681376875551632, |
|
"grad_norm": 0.8795916947723678, |
|
"learning_rate": 7.304300564319013e-06, |
|
"loss": 0.1154, |
|
"step": 1039 |
|
}, |
|
{ |
|
"epoch": 3.671668137687555, |
|
"grad_norm": 0.932134034033176, |
|
"learning_rate": 7.299220589998037e-06, |
|
"loss": 0.1172, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 3.675198587819947, |
|
"grad_norm": 1.060320989726038, |
|
"learning_rate": 7.294137604088487e-06, |
|
"loss": 0.1272, |
|
"step": 1041 |
|
}, |
|
{ |
|
"epoch": 3.6787290379523387, |
|
"grad_norm": 1.0865767224751157, |
|
"learning_rate": 7.2890516132482145e-06, |
|
"loss": 0.1174, |
|
"step": 1042 |
|
}, |
|
{ |
|
"epoch": 3.682259488084731, |
|
"grad_norm": 0.9524420011955063, |
|
"learning_rate": 7.283962624139004e-06, |
|
"loss": 0.1025, |
|
"step": 1043 |
|
}, |
|
{ |
|
"epoch": 3.6857899382171224, |
|
"grad_norm": 1.0723266113080743, |
|
"learning_rate": 7.278870643426565e-06, |
|
"loss": 0.1188, |
|
"step": 1044 |
|
}, |
|
{ |
|
"epoch": 3.6893203883495147, |
|
"grad_norm": 0.9889202356506911, |
|
"learning_rate": 7.27377567778053e-06, |
|
"loss": 0.1171, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 3.6928508384819065, |
|
"grad_norm": 1.1387504413356822, |
|
"learning_rate": 7.268677733874438e-06, |
|
"loss": 0.1347, |
|
"step": 1046 |
|
}, |
|
{ |
|
"epoch": 3.6963812886142984, |
|
"grad_norm": 0.9054547213610565, |
|
"learning_rate": 7.263576818385728e-06, |
|
"loss": 0.1192, |
|
"step": 1047 |
|
}, |
|
{ |
|
"epoch": 3.69991173874669, |
|
"grad_norm": 0.9389240088652642, |
|
"learning_rate": 7.258472937995736e-06, |
|
"loss": 0.105, |
|
"step": 1048 |
|
}, |
|
{ |
|
"epoch": 3.703442188879082, |
|
"grad_norm": 1.0097808039069447, |
|
"learning_rate": 7.253366099389678e-06, |
|
"loss": 0.1283, |
|
"step": 1049 |
|
}, |
|
{ |
|
"epoch": 3.706972639011474, |
|
"grad_norm": 0.9312349763358698, |
|
"learning_rate": 7.2482563092566446e-06, |
|
"loss": 0.1177, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 3.7105030891438657, |
|
"grad_norm": 0.8462385685589844, |
|
"learning_rate": 7.243143574289595e-06, |
|
"loss": 0.1176, |
|
"step": 1051 |
|
}, |
|
{ |
|
"epoch": 3.7140335392762576, |
|
"grad_norm": 0.9999611373105524, |
|
"learning_rate": 7.23802790118534e-06, |
|
"loss": 0.12, |
|
"step": 1052 |
|
}, |
|
{ |
|
"epoch": 3.7175639894086494, |
|
"grad_norm": 1.029486778996518, |
|
"learning_rate": 7.232909296644548e-06, |
|
"loss": 0.1313, |
|
"step": 1053 |
|
}, |
|
{ |
|
"epoch": 3.7210944395410417, |
|
"grad_norm": 0.9598825118610063, |
|
"learning_rate": 7.227787767371717e-06, |
|
"loss": 0.1057, |
|
"step": 1054 |
|
}, |
|
{ |
|
"epoch": 3.7246248896734335, |
|
"grad_norm": 0.8914389737346524, |
|
"learning_rate": 7.222663320075185e-06, |
|
"loss": 0.1196, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 3.7281553398058254, |
|
"grad_norm": 0.9459232821476742, |
|
"learning_rate": 7.217535961467105e-06, |
|
"loss": 0.1254, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 3.731685789938217, |
|
"grad_norm": 1.121524174912343, |
|
"learning_rate": 7.212405698263446e-06, |
|
"loss": 0.1296, |
|
"step": 1057 |
|
}, |
|
{ |
|
"epoch": 3.735216240070609, |
|
"grad_norm": 1.0305189372866497, |
|
"learning_rate": 7.2072725371839826e-06, |
|
"loss": 0.1123, |
|
"step": 1058 |
|
}, |
|
{ |
|
"epoch": 3.738746690203001, |
|
"grad_norm": 1.0095800993562205, |
|
"learning_rate": 7.202136484952286e-06, |
|
"loss": 0.1109, |
|
"step": 1059 |
|
}, |
|
{ |
|
"epoch": 3.7422771403353927, |
|
"grad_norm": 1.091036374867166, |
|
"learning_rate": 7.1969975482957075e-06, |
|
"loss": 0.125, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 3.7458075904677846, |
|
"grad_norm": 1.1262861615878523, |
|
"learning_rate": 7.191855733945388e-06, |
|
"loss": 0.1304, |
|
"step": 1061 |
|
}, |
|
{ |
|
"epoch": 3.7493380406001764, |
|
"grad_norm": 1.0644115251365422, |
|
"learning_rate": 7.1867110486362255e-06, |
|
"loss": 0.1221, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 3.7528684907325687, |
|
"grad_norm": 1.0506019899313102, |
|
"learning_rate": 7.1815634991068895e-06, |
|
"loss": 0.1205, |
|
"step": 1063 |
|
}, |
|
{ |
|
"epoch": 3.75639894086496, |
|
"grad_norm": 1.0044867932848505, |
|
"learning_rate": 7.176413092099791e-06, |
|
"loss": 0.1234, |
|
"step": 1064 |
|
}, |
|
{ |
|
"epoch": 3.7599293909973523, |
|
"grad_norm": 0.9095279203965683, |
|
"learning_rate": 7.171259834361092e-06, |
|
"loss": 0.1205, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 3.763459841129744, |
|
"grad_norm": 1.1445775552595612, |
|
"learning_rate": 7.1661037326406825e-06, |
|
"loss": 0.1275, |
|
"step": 1066 |
|
}, |
|
{ |
|
"epoch": 3.766990291262136, |
|
"grad_norm": 0.946883327804174, |
|
"learning_rate": 7.1609447936921814e-06, |
|
"loss": 0.1398, |
|
"step": 1067 |
|
}, |
|
{ |
|
"epoch": 3.770520741394528, |
|
"grad_norm": 0.9041398561076188, |
|
"learning_rate": 7.155783024272925e-06, |
|
"loss": 0.13, |
|
"step": 1068 |
|
}, |
|
{ |
|
"epoch": 3.7740511915269197, |
|
"grad_norm": 0.8457469332005921, |
|
"learning_rate": 7.150618431143951e-06, |
|
"loss": 0.1143, |
|
"step": 1069 |
|
}, |
|
{ |
|
"epoch": 3.7775816416593115, |
|
"grad_norm": 0.8578352898530813, |
|
"learning_rate": 7.145451021070003e-06, |
|
"loss": 0.1247, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 3.7811120917917034, |
|
"grad_norm": 0.9352514012935873, |
|
"learning_rate": 7.140280800819507e-06, |
|
"loss": 0.1344, |
|
"step": 1071 |
|
}, |
|
{ |
|
"epoch": 3.784642541924095, |
|
"grad_norm": 0.9799389793882611, |
|
"learning_rate": 7.135107777164576e-06, |
|
"loss": 0.1223, |
|
"step": 1072 |
|
}, |
|
{ |
|
"epoch": 3.788172992056487, |
|
"grad_norm": 1.065041516403633, |
|
"learning_rate": 7.129931956880992e-06, |
|
"loss": 0.1375, |
|
"step": 1073 |
|
}, |
|
{ |
|
"epoch": 3.7917034421888793, |
|
"grad_norm": 1.0449759566953258, |
|
"learning_rate": 7.124753346748201e-06, |
|
"loss": 0.1171, |
|
"step": 1074 |
|
}, |
|
{ |
|
"epoch": 3.7952338923212707, |
|
"grad_norm": 1.1549382686805798, |
|
"learning_rate": 7.119571953549305e-06, |
|
"loss": 0.1281, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 3.798764342453663, |
|
"grad_norm": 1.007228257856292, |
|
"learning_rate": 7.114387784071044e-06, |
|
"loss": 0.1269, |
|
"step": 1076 |
|
}, |
|
{ |
|
"epoch": 3.802294792586055, |
|
"grad_norm": 1.0225753082093523, |
|
"learning_rate": 7.109200845103806e-06, |
|
"loss": 0.1199, |
|
"step": 1077 |
|
}, |
|
{ |
|
"epoch": 3.8058252427184467, |
|
"grad_norm": 0.9355931698169853, |
|
"learning_rate": 7.104011143441596e-06, |
|
"loss": 0.1169, |
|
"step": 1078 |
|
}, |
|
{ |
|
"epoch": 3.8093556928508385, |
|
"grad_norm": 0.963698358188344, |
|
"learning_rate": 7.098818685882044e-06, |
|
"loss": 0.1179, |
|
"step": 1079 |
|
}, |
|
{ |
|
"epoch": 3.8128861429832304, |
|
"grad_norm": 0.7777902932062336, |
|
"learning_rate": 7.093623479226386e-06, |
|
"loss": 0.1186, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 3.816416593115622, |
|
"grad_norm": 0.9518180229607743, |
|
"learning_rate": 7.088425530279461e-06, |
|
"loss": 0.1062, |
|
"step": 1081 |
|
}, |
|
{ |
|
"epoch": 3.819947043248014, |
|
"grad_norm": 1.034322676608895, |
|
"learning_rate": 7.083224845849701e-06, |
|
"loss": 0.1144, |
|
"step": 1082 |
|
}, |
|
{ |
|
"epoch": 3.823477493380406, |
|
"grad_norm": 1.0557347999282005, |
|
"learning_rate": 7.078021432749121e-06, |
|
"loss": 0.1243, |
|
"step": 1083 |
|
}, |
|
{ |
|
"epoch": 3.8270079435127977, |
|
"grad_norm": 0.8967864020526236, |
|
"learning_rate": 7.072815297793303e-06, |
|
"loss": 0.1189, |
|
"step": 1084 |
|
}, |
|
{ |
|
"epoch": 3.83053839364519, |
|
"grad_norm": 0.9394124246894394, |
|
"learning_rate": 7.067606447801405e-06, |
|
"loss": 0.1116, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 3.8340688437775814, |
|
"grad_norm": 0.9390953780920334, |
|
"learning_rate": 7.062394889596133e-06, |
|
"loss": 0.1258, |
|
"step": 1086 |
|
}, |
|
{ |
|
"epoch": 3.8375992939099737, |
|
"grad_norm": 0.9777220254987733, |
|
"learning_rate": 7.057180630003746e-06, |
|
"loss": 0.1241, |
|
"step": 1087 |
|
}, |
|
{ |
|
"epoch": 3.8411297440423655, |
|
"grad_norm": 0.8779726966694773, |
|
"learning_rate": 7.051963675854035e-06, |
|
"loss": 0.1267, |
|
"step": 1088 |
|
}, |
|
{ |
|
"epoch": 3.8446601941747574, |
|
"grad_norm": 1.0359165893132334, |
|
"learning_rate": 7.046744033980329e-06, |
|
"loss": 0.1168, |
|
"step": 1089 |
|
}, |
|
{ |
|
"epoch": 3.848190644307149, |
|
"grad_norm": 0.9976911554856348, |
|
"learning_rate": 7.041521711219468e-06, |
|
"loss": 0.1317, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 3.851721094439541, |
|
"grad_norm": 1.0169743441045915, |
|
"learning_rate": 7.03629671441181e-06, |
|
"loss": 0.1398, |
|
"step": 1091 |
|
}, |
|
{ |
|
"epoch": 3.855251544571933, |
|
"grad_norm": 0.9568634693122333, |
|
"learning_rate": 7.031069050401215e-06, |
|
"loss": 0.0919, |
|
"step": 1092 |
|
}, |
|
{ |
|
"epoch": 3.8587819947043247, |
|
"grad_norm": 0.9083188838175256, |
|
"learning_rate": 7.025838726035032e-06, |
|
"loss": 0.1098, |
|
"step": 1093 |
|
}, |
|
{ |
|
"epoch": 3.8623124448367165, |
|
"grad_norm": 0.9312177288214185, |
|
"learning_rate": 7.020605748164101e-06, |
|
"loss": 0.1275, |
|
"step": 1094 |
|
}, |
|
{ |
|
"epoch": 3.8658428949691084, |
|
"grad_norm": 0.8178614978701555, |
|
"learning_rate": 7.015370123642733e-06, |
|
"loss": 0.11, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 3.8693733451015007, |
|
"grad_norm": 0.9943777665520181, |
|
"learning_rate": 7.010131859328706e-06, |
|
"loss": 0.1141, |
|
"step": 1096 |
|
}, |
|
{ |
|
"epoch": 3.872903795233892, |
|
"grad_norm": 0.7533041541573118, |
|
"learning_rate": 7.00489096208326e-06, |
|
"loss": 0.0988, |
|
"step": 1097 |
|
}, |
|
{ |
|
"epoch": 3.8764342453662843, |
|
"grad_norm": 0.9539898985521694, |
|
"learning_rate": 6.999647438771076e-06, |
|
"loss": 0.1158, |
|
"step": 1098 |
|
}, |
|
{ |
|
"epoch": 3.879964695498676, |
|
"grad_norm": 0.9694635032226555, |
|
"learning_rate": 6.994401296260285e-06, |
|
"loss": 0.1476, |
|
"step": 1099 |
|
}, |
|
{ |
|
"epoch": 3.883495145631068, |
|
"grad_norm": 0.9466515792078597, |
|
"learning_rate": 6.989152541422438e-06, |
|
"loss": 0.1207, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 3.88702559576346, |
|
"grad_norm": 0.9232735267824136, |
|
"learning_rate": 6.983901181132514e-06, |
|
"loss": 0.1219, |
|
"step": 1101 |
|
}, |
|
{ |
|
"epoch": 3.8905560458958517, |
|
"grad_norm": 0.9035726923781363, |
|
"learning_rate": 6.978647222268904e-06, |
|
"loss": 0.1139, |
|
"step": 1102 |
|
}, |
|
{ |
|
"epoch": 3.8940864960282435, |
|
"grad_norm": 0.9447988957374035, |
|
"learning_rate": 6.973390671713401e-06, |
|
"loss": 0.1106, |
|
"step": 1103 |
|
}, |
|
{ |
|
"epoch": 3.8976169461606354, |
|
"grad_norm": 0.9993165292258056, |
|
"learning_rate": 6.9681315363511946e-06, |
|
"loss": 0.1126, |
|
"step": 1104 |
|
}, |
|
{ |
|
"epoch": 3.901147396293027, |
|
"grad_norm": 0.9567127281227304, |
|
"learning_rate": 6.9628698230708604e-06, |
|
"loss": 0.126, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 3.904677846425419, |
|
"grad_norm": 0.9704513493237618, |
|
"learning_rate": 6.957605538764347e-06, |
|
"loss": 0.1135, |
|
"step": 1106 |
|
}, |
|
{ |
|
"epoch": 3.9082082965578113, |
|
"grad_norm": 0.9049083128378328, |
|
"learning_rate": 6.9523386903269715e-06, |
|
"loss": 0.118, |
|
"step": 1107 |
|
}, |
|
{ |
|
"epoch": 3.911738746690203, |
|
"grad_norm": 1.1410548052057965, |
|
"learning_rate": 6.947069284657416e-06, |
|
"loss": 0.1232, |
|
"step": 1108 |
|
}, |
|
{ |
|
"epoch": 3.915269196822595, |
|
"grad_norm": 0.9628185789378251, |
|
"learning_rate": 6.941797328657702e-06, |
|
"loss": 0.1012, |
|
"step": 1109 |
|
}, |
|
{ |
|
"epoch": 3.918799646954987, |
|
"grad_norm": 0.986500497408887, |
|
"learning_rate": 6.936522829233202e-06, |
|
"loss": 0.1267, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 3.9223300970873787, |
|
"grad_norm": 0.9404574261439161, |
|
"learning_rate": 6.93124579329261e-06, |
|
"loss": 0.1177, |
|
"step": 1111 |
|
}, |
|
{ |
|
"epoch": 3.9258605472197705, |
|
"grad_norm": 0.9652812442302577, |
|
"learning_rate": 6.92596622774795e-06, |
|
"loss": 0.1099, |
|
"step": 1112 |
|
}, |
|
{ |
|
"epoch": 3.9293909973521624, |
|
"grad_norm": 0.9255614343816538, |
|
"learning_rate": 6.920684139514555e-06, |
|
"loss": 0.1157, |
|
"step": 1113 |
|
}, |
|
{ |
|
"epoch": 3.932921447484554, |
|
"grad_norm": 1.1133925624513497, |
|
"learning_rate": 6.9153995355110645e-06, |
|
"loss": 0.13, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 3.936451897616946, |
|
"grad_norm": 1.0496747828703927, |
|
"learning_rate": 6.9101124226594106e-06, |
|
"loss": 0.1123, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 3.9399823477493383, |
|
"grad_norm": 1.0246048820560874, |
|
"learning_rate": 6.904822807884816e-06, |
|
"loss": 0.1158, |
|
"step": 1116 |
|
}, |
|
{ |
|
"epoch": 3.9435127978817297, |
|
"grad_norm": 1.0333168676182993, |
|
"learning_rate": 6.899530698115776e-06, |
|
"loss": 0.1315, |
|
"step": 1117 |
|
}, |
|
{ |
|
"epoch": 3.947043248014122, |
|
"grad_norm": 1.0064120983759648, |
|
"learning_rate": 6.894236100284058e-06, |
|
"loss": 0.1282, |
|
"step": 1118 |
|
}, |
|
{ |
|
"epoch": 3.950573698146514, |
|
"grad_norm": 0.963684336758527, |
|
"learning_rate": 6.888939021324681e-06, |
|
"loss": 0.1217, |
|
"step": 1119 |
|
}, |
|
{ |
|
"epoch": 3.9541041482789057, |
|
"grad_norm": 0.9892019560831151, |
|
"learning_rate": 6.883639468175926e-06, |
|
"loss": 0.1239, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 3.9576345984112975, |
|
"grad_norm": 1.0120690523201414, |
|
"learning_rate": 6.878337447779304e-06, |
|
"loss": 0.1335, |
|
"step": 1121 |
|
}, |
|
{ |
|
"epoch": 3.9611650485436893, |
|
"grad_norm": 0.8218662981254, |
|
"learning_rate": 6.873032967079562e-06, |
|
"loss": 0.1185, |
|
"step": 1122 |
|
}, |
|
{ |
|
"epoch": 3.964695498676081, |
|
"grad_norm": 0.9915669809027783, |
|
"learning_rate": 6.867726033024667e-06, |
|
"loss": 0.1181, |
|
"step": 1123 |
|
}, |
|
{ |
|
"epoch": 3.968225948808473, |
|
"grad_norm": 0.7674002943828144, |
|
"learning_rate": 6.862416652565804e-06, |
|
"loss": 0.11, |
|
"step": 1124 |
|
}, |
|
{ |
|
"epoch": 3.971756398940865, |
|
"grad_norm": 0.9775474573427773, |
|
"learning_rate": 6.85710483265736e-06, |
|
"loss": 0.1357, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 3.9752868490732567, |
|
"grad_norm": 0.9046890209780649, |
|
"learning_rate": 6.8517905802569164e-06, |
|
"loss": 0.1145, |
|
"step": 1126 |
|
}, |
|
{ |
|
"epoch": 3.978817299205649, |
|
"grad_norm": 0.921557821840316, |
|
"learning_rate": 6.846473902325243e-06, |
|
"loss": 0.1392, |
|
"step": 1127 |
|
}, |
|
{ |
|
"epoch": 3.9823477493380404, |
|
"grad_norm": 0.9794967931322031, |
|
"learning_rate": 6.8411548058262824e-06, |
|
"loss": 0.1088, |
|
"step": 1128 |
|
}, |
|
{ |
|
"epoch": 3.9858781994704326, |
|
"grad_norm": 0.8740159825381886, |
|
"learning_rate": 6.835833297727148e-06, |
|
"loss": 0.1139, |
|
"step": 1129 |
|
}, |
|
{ |
|
"epoch": 3.9894086496028245, |
|
"grad_norm": 0.8628586385174893, |
|
"learning_rate": 6.830509384998114e-06, |
|
"loss": 0.1082, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 3.9929390997352163, |
|
"grad_norm": 0.9488799231960956, |
|
"learning_rate": 6.8251830746126e-06, |
|
"loss": 0.1316, |
|
"step": 1131 |
|
}, |
|
{ |
|
"epoch": 3.996469549867608, |
|
"grad_norm": 1.0703456877526685, |
|
"learning_rate": 6.81985437354717e-06, |
|
"loss": 0.1279, |
|
"step": 1132 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 0.967853650509713, |
|
"learning_rate": 6.814523288781516e-06, |
|
"loss": 0.1305, |
|
"step": 1133 |
|
}, |
|
{ |
|
"epoch": 4.003530450132392, |
|
"grad_norm": 0.7981320722327793, |
|
"learning_rate": 6.809189827298457e-06, |
|
"loss": 0.0885, |
|
"step": 1134 |
|
}, |
|
{ |
|
"epoch": 4.007060900264784, |
|
"grad_norm": 0.5854902750468941, |
|
"learning_rate": 6.803853996083918e-06, |
|
"loss": 0.0793, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 4.010591350397176, |
|
"grad_norm": 0.7049273298585778, |
|
"learning_rate": 6.7985158021269335e-06, |
|
"loss": 0.0758, |
|
"step": 1136 |
|
}, |
|
{ |
|
"epoch": 4.014121800529567, |
|
"grad_norm": 0.6869340400131053, |
|
"learning_rate": 6.79317525241963e-06, |
|
"loss": 0.0933, |
|
"step": 1137 |
|
}, |
|
{ |
|
"epoch": 4.01765225066196, |
|
"grad_norm": 0.5900057189092714, |
|
"learning_rate": 6.787832353957225e-06, |
|
"loss": 0.0706, |
|
"step": 1138 |
|
}, |
|
{ |
|
"epoch": 4.021182700794351, |
|
"grad_norm": 0.7601438078339049, |
|
"learning_rate": 6.782487113738002e-06, |
|
"loss": 0.0769, |
|
"step": 1139 |
|
}, |
|
{ |
|
"epoch": 4.024713150926743, |
|
"grad_norm": 0.851783451355363, |
|
"learning_rate": 6.777139538763324e-06, |
|
"loss": 0.0711, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 4.028243601059135, |
|
"grad_norm": 0.82880112354727, |
|
"learning_rate": 6.7717896360376025e-06, |
|
"loss": 0.0717, |
|
"step": 1141 |
|
}, |
|
{ |
|
"epoch": 4.031774051191527, |
|
"grad_norm": 1.1143799167216393, |
|
"learning_rate": 6.766437412568302e-06, |
|
"loss": 0.0784, |
|
"step": 1142 |
|
}, |
|
{ |
|
"epoch": 4.035304501323918, |
|
"grad_norm": 0.8210313928830595, |
|
"learning_rate": 6.761082875365929e-06, |
|
"loss": 0.0626, |
|
"step": 1143 |
|
}, |
|
{ |
|
"epoch": 4.038834951456311, |
|
"grad_norm": 0.7779155627289819, |
|
"learning_rate": 6.755726031444014e-06, |
|
"loss": 0.0649, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 4.042365401588703, |
|
"grad_norm": 1.1156761361527734, |
|
"learning_rate": 6.750366887819116e-06, |
|
"loss": 0.0787, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 4.045895851721094, |
|
"grad_norm": 1.11782008947278, |
|
"learning_rate": 6.745005451510804e-06, |
|
"loss": 0.0742, |
|
"step": 1146 |
|
}, |
|
{ |
|
"epoch": 4.049426301853487, |
|
"grad_norm": 1.0344771965558572, |
|
"learning_rate": 6.739641729541645e-06, |
|
"loss": 0.0838, |
|
"step": 1147 |
|
}, |
|
{ |
|
"epoch": 4.052956751985878, |
|
"grad_norm": 2.0154736605675763, |
|
"learning_rate": 6.734275728937208e-06, |
|
"loss": 0.1034, |
|
"step": 1148 |
|
}, |
|
{ |
|
"epoch": 4.05648720211827, |
|
"grad_norm": 1.2294533756515769, |
|
"learning_rate": 6.728907456726042e-06, |
|
"loss": 0.0619, |
|
"step": 1149 |
|
}, |
|
{ |
|
"epoch": 4.060017652250662, |
|
"grad_norm": 1.4290104269014912, |
|
"learning_rate": 6.723536919939669e-06, |
|
"loss": 0.0686, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 4.063548102383054, |
|
"grad_norm": 1.5020515871139606, |
|
"learning_rate": 6.718164125612585e-06, |
|
"loss": 0.0761, |
|
"step": 1151 |
|
}, |
|
{ |
|
"epoch": 4.067078552515445, |
|
"grad_norm": 1.4908120498463364, |
|
"learning_rate": 6.712789080782235e-06, |
|
"loss": 0.0828, |
|
"step": 1152 |
|
}, |
|
{ |
|
"epoch": 4.070609002647838, |
|
"grad_norm": 1.2267050202060312, |
|
"learning_rate": 6.707411792489016e-06, |
|
"loss": 0.0643, |
|
"step": 1153 |
|
}, |
|
{ |
|
"epoch": 4.074139452780229, |
|
"grad_norm": 1.3569581964127149, |
|
"learning_rate": 6.7020322677762614e-06, |
|
"loss": 0.082, |
|
"step": 1154 |
|
}, |
|
{ |
|
"epoch": 4.077669902912621, |
|
"grad_norm": 1.1848448933110132, |
|
"learning_rate": 6.696650513690238e-06, |
|
"loss": 0.0639, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 4.081200353045014, |
|
"grad_norm": 1.181765635994709, |
|
"learning_rate": 6.691266537280128e-06, |
|
"loss": 0.057, |
|
"step": 1156 |
|
}, |
|
{ |
|
"epoch": 4.084730803177405, |
|
"grad_norm": 1.016848718981542, |
|
"learning_rate": 6.685880345598027e-06, |
|
"loss": 0.0537, |
|
"step": 1157 |
|
}, |
|
{ |
|
"epoch": 4.088261253309797, |
|
"grad_norm": 1.3224355302362136, |
|
"learning_rate": 6.680491945698929e-06, |
|
"loss": 0.064, |
|
"step": 1158 |
|
}, |
|
{ |
|
"epoch": 4.091791703442189, |
|
"grad_norm": 1.4746292032964976, |
|
"learning_rate": 6.675101344640727e-06, |
|
"loss": 0.0892, |
|
"step": 1159 |
|
}, |
|
{ |
|
"epoch": 4.095322153574581, |
|
"grad_norm": 1.5383513145198393, |
|
"learning_rate": 6.66970854948419e-06, |
|
"loss": 0.0766, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 4.098852603706972, |
|
"grad_norm": 1.207790515633538, |
|
"learning_rate": 6.664313567292966e-06, |
|
"loss": 0.0693, |
|
"step": 1161 |
|
}, |
|
{ |
|
"epoch": 4.102383053839365, |
|
"grad_norm": 1.288949461541623, |
|
"learning_rate": 6.658916405133562e-06, |
|
"loss": 0.0826, |
|
"step": 1162 |
|
}, |
|
{ |
|
"epoch": 4.105913503971756, |
|
"grad_norm": 1.0936618806184264, |
|
"learning_rate": 6.6535170700753506e-06, |
|
"loss": 0.0789, |
|
"step": 1163 |
|
}, |
|
{ |
|
"epoch": 4.109443954104148, |
|
"grad_norm": 1.087774369178708, |
|
"learning_rate": 6.6481155691905375e-06, |
|
"loss": 0.0571, |
|
"step": 1164 |
|
}, |
|
{ |
|
"epoch": 4.112974404236541, |
|
"grad_norm": 1.3919202257673968, |
|
"learning_rate": 6.6427119095541745e-06, |
|
"loss": 0.0655, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 4.116504854368932, |
|
"grad_norm": 1.0692129775554304, |
|
"learning_rate": 6.637306098244138e-06, |
|
"loss": 0.0758, |
|
"step": 1166 |
|
}, |
|
{ |
|
"epoch": 4.120035304501324, |
|
"grad_norm": 1.1487653554819723, |
|
"learning_rate": 6.631898142341122e-06, |
|
"loss": 0.0619, |
|
"step": 1167 |
|
}, |
|
{ |
|
"epoch": 4.123565754633716, |
|
"grad_norm": 1.036223404954323, |
|
"learning_rate": 6.626488048928632e-06, |
|
"loss": 0.0593, |
|
"step": 1168 |
|
}, |
|
{ |
|
"epoch": 4.127096204766108, |
|
"grad_norm": 1.0121114014928942, |
|
"learning_rate": 6.621075825092972e-06, |
|
"loss": 0.0667, |
|
"step": 1169 |
|
}, |
|
{ |
|
"epoch": 4.130626654898499, |
|
"grad_norm": 1.1972992188904066, |
|
"learning_rate": 6.6156614779232345e-06, |
|
"loss": 0.0733, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 4.134157105030892, |
|
"grad_norm": 1.197590502375495, |
|
"learning_rate": 6.610245014511296e-06, |
|
"loss": 0.0611, |
|
"step": 1171 |
|
}, |
|
{ |
|
"epoch": 4.137687555163283, |
|
"grad_norm": 1.0881571550345615, |
|
"learning_rate": 6.604826441951803e-06, |
|
"loss": 0.061, |
|
"step": 1172 |
|
}, |
|
{ |
|
"epoch": 4.141218005295675, |
|
"grad_norm": 0.9686194494515865, |
|
"learning_rate": 6.599405767342167e-06, |
|
"loss": 0.0559, |
|
"step": 1173 |
|
}, |
|
{ |
|
"epoch": 4.144748455428067, |
|
"grad_norm": 1.3831994016631128, |
|
"learning_rate": 6.593982997782549e-06, |
|
"loss": 0.0642, |
|
"step": 1174 |
|
}, |
|
{ |
|
"epoch": 4.148278905560459, |
|
"grad_norm": 1.241798068520593, |
|
"learning_rate": 6.588558140375858e-06, |
|
"loss": 0.0707, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 4.151809355692851, |
|
"grad_norm": 1.2592043421219463, |
|
"learning_rate": 6.583131202227736e-06, |
|
"loss": 0.0851, |
|
"step": 1176 |
|
}, |
|
{ |
|
"epoch": 4.155339805825243, |
|
"grad_norm": 1.1890876424522738, |
|
"learning_rate": 6.577702190446552e-06, |
|
"loss": 0.0764, |
|
"step": 1177 |
|
}, |
|
{ |
|
"epoch": 4.158870255957635, |
|
"grad_norm": 1.1892762885244956, |
|
"learning_rate": 6.572271112143386e-06, |
|
"loss": 0.0823, |
|
"step": 1178 |
|
}, |
|
{ |
|
"epoch": 4.162400706090026, |
|
"grad_norm": 1.098687094404278, |
|
"learning_rate": 6.566837974432032e-06, |
|
"loss": 0.0582, |
|
"step": 1179 |
|
}, |
|
{ |
|
"epoch": 4.165931156222419, |
|
"grad_norm": 1.1795811578347426, |
|
"learning_rate": 6.561402784428974e-06, |
|
"loss": 0.0787, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 4.16946160635481, |
|
"grad_norm": 1.0850929765127277, |
|
"learning_rate": 6.5559655492533915e-06, |
|
"loss": 0.0576, |
|
"step": 1181 |
|
}, |
|
{ |
|
"epoch": 4.172992056487202, |
|
"grad_norm": 1.2039461760737675, |
|
"learning_rate": 6.550526276027138e-06, |
|
"loss": 0.075, |
|
"step": 1182 |
|
}, |
|
{ |
|
"epoch": 4.176522506619594, |
|
"grad_norm": 1.1924260467355354, |
|
"learning_rate": 6.545084971874738e-06, |
|
"loss": 0.0759, |
|
"step": 1183 |
|
}, |
|
{ |
|
"epoch": 4.180052956751986, |
|
"grad_norm": 1.349601668880594, |
|
"learning_rate": 6.539641643923376e-06, |
|
"loss": 0.0762, |
|
"step": 1184 |
|
}, |
|
{ |
|
"epoch": 4.183583406884377, |
|
"grad_norm": 1.5302048823422634, |
|
"learning_rate": 6.534196299302887e-06, |
|
"loss": 0.0721, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 4.18711385701677, |
|
"grad_norm": 1.1161238197658445, |
|
"learning_rate": 6.528748945145749e-06, |
|
"loss": 0.0752, |
|
"step": 1186 |
|
}, |
|
{ |
|
"epoch": 4.190644307149162, |
|
"grad_norm": 1.0876160222524982, |
|
"learning_rate": 6.523299588587069e-06, |
|
"loss": 0.0672, |
|
"step": 1187 |
|
}, |
|
{ |
|
"epoch": 4.194174757281553, |
|
"grad_norm": 1.1607986364903538, |
|
"learning_rate": 6.51784823676458e-06, |
|
"loss": 0.0743, |
|
"step": 1188 |
|
}, |
|
{ |
|
"epoch": 4.197705207413946, |
|
"grad_norm": 1.0772412583188433, |
|
"learning_rate": 6.5123948968186275e-06, |
|
"loss": 0.0673, |
|
"step": 1189 |
|
}, |
|
{ |
|
"epoch": 4.201235657546337, |
|
"grad_norm": 1.2617922130605492, |
|
"learning_rate": 6.506939575892161e-06, |
|
"loss": 0.0619, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 4.204766107678729, |
|
"grad_norm": 1.2483711396156587, |
|
"learning_rate": 6.501482281130724e-06, |
|
"loss": 0.0871, |
|
"step": 1191 |
|
}, |
|
{ |
|
"epoch": 4.208296557811121, |
|
"grad_norm": 1.1989951304725375, |
|
"learning_rate": 6.496023019682447e-06, |
|
"loss": 0.068, |
|
"step": 1192 |
|
}, |
|
{ |
|
"epoch": 4.211827007943513, |
|
"grad_norm": 1.038016185342428, |
|
"learning_rate": 6.490561798698033e-06, |
|
"loss": 0.0672, |
|
"step": 1193 |
|
}, |
|
{ |
|
"epoch": 4.215357458075904, |
|
"grad_norm": 1.2879497667669522, |
|
"learning_rate": 6.4850986253307566e-06, |
|
"loss": 0.0813, |
|
"step": 1194 |
|
}, |
|
{ |
|
"epoch": 4.218887908208297, |
|
"grad_norm": 1.138562370359682, |
|
"learning_rate": 6.479633506736447e-06, |
|
"loss": 0.0798, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 4.222418358340688, |
|
"grad_norm": 1.2141342731122955, |
|
"learning_rate": 6.474166450073482e-06, |
|
"loss": 0.0753, |
|
"step": 1196 |
|
}, |
|
{ |
|
"epoch": 4.22594880847308, |
|
"grad_norm": 1.1524180729165578, |
|
"learning_rate": 6.468697462502776e-06, |
|
"loss": 0.0669, |
|
"step": 1197 |
|
}, |
|
{ |
|
"epoch": 4.229479258605473, |
|
"grad_norm": 1.2503618917005421, |
|
"learning_rate": 6.463226551187776e-06, |
|
"loss": 0.0802, |
|
"step": 1198 |
|
}, |
|
{ |
|
"epoch": 4.233009708737864, |
|
"grad_norm": 1.0897556077719, |
|
"learning_rate": 6.457753723294443e-06, |
|
"loss": 0.0687, |
|
"step": 1199 |
|
}, |
|
{ |
|
"epoch": 4.236540158870256, |
|
"grad_norm": 1.2215195146222155, |
|
"learning_rate": 6.452278985991257e-06, |
|
"loss": 0.0859, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 4.240070609002648, |
|
"grad_norm": 1.123982384909271, |
|
"learning_rate": 6.4468023464491906e-06, |
|
"loss": 0.0693, |
|
"step": 1201 |
|
}, |
|
{ |
|
"epoch": 4.24360105913504, |
|
"grad_norm": 1.161677655741415, |
|
"learning_rate": 6.441323811841715e-06, |
|
"loss": 0.0749, |
|
"step": 1202 |
|
}, |
|
{ |
|
"epoch": 4.247131509267431, |
|
"grad_norm": 1.1029803998203793, |
|
"learning_rate": 6.435843389344778e-06, |
|
"loss": 0.0666, |
|
"step": 1203 |
|
}, |
|
{ |
|
"epoch": 4.250661959399824, |
|
"grad_norm": 1.3744931507694842, |
|
"learning_rate": 6.4303610861368026e-06, |
|
"loss": 0.0844, |
|
"step": 1204 |
|
}, |
|
{ |
|
"epoch": 4.254192409532215, |
|
"grad_norm": 1.0659314315032296, |
|
"learning_rate": 6.4248769093986765e-06, |
|
"loss": 0.0723, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 4.257722859664607, |
|
"grad_norm": 1.2659662434004557, |
|
"learning_rate": 6.419390866313741e-06, |
|
"loss": 0.0786, |
|
"step": 1206 |
|
}, |
|
{ |
|
"epoch": 4.261253309796999, |
|
"grad_norm": 1.2058156571448462, |
|
"learning_rate": 6.413902964067776e-06, |
|
"loss": 0.0708, |
|
"step": 1207 |
|
}, |
|
{ |
|
"epoch": 4.264783759929391, |
|
"grad_norm": 1.4552116581087131, |
|
"learning_rate": 6.408413209849007e-06, |
|
"loss": 0.0802, |
|
"step": 1208 |
|
}, |
|
{ |
|
"epoch": 4.268314210061783, |
|
"grad_norm": 1.3140338475558748, |
|
"learning_rate": 6.402921610848076e-06, |
|
"loss": 0.0701, |
|
"step": 1209 |
|
}, |
|
{ |
|
"epoch": 4.271844660194175, |
|
"grad_norm": 1.4948147235906752, |
|
"learning_rate": 6.397428174258048e-06, |
|
"loss": 0.0765, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 4.275375110326567, |
|
"grad_norm": 1.4432674805143244, |
|
"learning_rate": 6.391932907274389e-06, |
|
"loss": 0.0736, |
|
"step": 1211 |
|
}, |
|
{ |
|
"epoch": 4.278905560458958, |
|
"grad_norm": 1.1269206127304876, |
|
"learning_rate": 6.386435817094971e-06, |
|
"loss": 0.0762, |
|
"step": 1212 |
|
}, |
|
{ |
|
"epoch": 4.282436010591351, |
|
"grad_norm": 1.4426602486276199, |
|
"learning_rate": 6.380936910920046e-06, |
|
"loss": 0.0834, |
|
"step": 1213 |
|
}, |
|
{ |
|
"epoch": 4.285966460723742, |
|
"grad_norm": 0.9454852560163308, |
|
"learning_rate": 6.375436195952245e-06, |
|
"loss": 0.0677, |
|
"step": 1214 |
|
}, |
|
{ |
|
"epoch": 4.289496910856134, |
|
"grad_norm": 1.1295627967839161, |
|
"learning_rate": 6.369933679396571e-06, |
|
"loss": 0.057, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 4.293027360988526, |
|
"grad_norm": 1.19896192193708, |
|
"learning_rate": 6.364429368460391e-06, |
|
"loss": 0.0786, |
|
"step": 1216 |
|
}, |
|
{ |
|
"epoch": 4.296557811120918, |
|
"grad_norm": 1.2164688961559822, |
|
"learning_rate": 6.358923270353412e-06, |
|
"loss": 0.0743, |
|
"step": 1217 |
|
}, |
|
{ |
|
"epoch": 4.300088261253309, |
|
"grad_norm": 1.4316527320361974, |
|
"learning_rate": 6.353415392287692e-06, |
|
"loss": 0.0735, |
|
"step": 1218 |
|
}, |
|
{ |
|
"epoch": 4.303618711385702, |
|
"grad_norm": 1.1280198948959894, |
|
"learning_rate": 6.347905741477613e-06, |
|
"loss": 0.0638, |
|
"step": 1219 |
|
}, |
|
{ |
|
"epoch": 4.307149161518094, |
|
"grad_norm": 1.3537055389950547, |
|
"learning_rate": 6.342394325139886e-06, |
|
"loss": 0.0844, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 4.310679611650485, |
|
"grad_norm": 1.175521881817613, |
|
"learning_rate": 6.336881150493525e-06, |
|
"loss": 0.0697, |
|
"step": 1221 |
|
}, |
|
{ |
|
"epoch": 4.314210061782878, |
|
"grad_norm": 1.3032526032205891, |
|
"learning_rate": 6.3313662247598575e-06, |
|
"loss": 0.0681, |
|
"step": 1222 |
|
}, |
|
{ |
|
"epoch": 4.317740511915269, |
|
"grad_norm": 1.2291592653948193, |
|
"learning_rate": 6.325849555162496e-06, |
|
"loss": 0.078, |
|
"step": 1223 |
|
}, |
|
{ |
|
"epoch": 4.321270962047661, |
|
"grad_norm": 1.2292408189938648, |
|
"learning_rate": 6.320331148927344e-06, |
|
"loss": 0.0605, |
|
"step": 1224 |
|
}, |
|
{ |
|
"epoch": 4.324801412180053, |
|
"grad_norm": 1.317695842582305, |
|
"learning_rate": 6.314811013282574e-06, |
|
"loss": 0.074, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 4.328331862312445, |
|
"grad_norm": 1.3300537195408657, |
|
"learning_rate": 6.309289155458629e-06, |
|
"loss": 0.0781, |
|
"step": 1226 |
|
}, |
|
{ |
|
"epoch": 4.331862312444836, |
|
"grad_norm": 1.2795961802893723, |
|
"learning_rate": 6.303765582688204e-06, |
|
"loss": 0.0758, |
|
"step": 1227 |
|
}, |
|
{ |
|
"epoch": 4.335392762577229, |
|
"grad_norm": 1.089546205227726, |
|
"learning_rate": 6.298240302206242e-06, |
|
"loss": 0.0716, |
|
"step": 1228 |
|
}, |
|
{ |
|
"epoch": 4.338923212709621, |
|
"grad_norm": 1.4436985362559518, |
|
"learning_rate": 6.29271332124992e-06, |
|
"loss": 0.0733, |
|
"step": 1229 |
|
}, |
|
{ |
|
"epoch": 4.342453662842012, |
|
"grad_norm": 1.2123854120814914, |
|
"learning_rate": 6.287184647058648e-06, |
|
"loss": 0.063, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 4.345984112974405, |
|
"grad_norm": 1.1599856785429075, |
|
"learning_rate": 6.281654286874047e-06, |
|
"loss": 0.0703, |
|
"step": 1231 |
|
}, |
|
{ |
|
"epoch": 4.349514563106796, |
|
"grad_norm": 1.2863269775769035, |
|
"learning_rate": 6.276122247939953e-06, |
|
"loss": 0.08, |
|
"step": 1232 |
|
}, |
|
{ |
|
"epoch": 4.353045013239188, |
|
"grad_norm": 0.9802224693174152, |
|
"learning_rate": 6.2705885375023955e-06, |
|
"loss": 0.0669, |
|
"step": 1233 |
|
}, |
|
{ |
|
"epoch": 4.35657546337158, |
|
"grad_norm": 1.3003686599476758, |
|
"learning_rate": 6.265053162809597e-06, |
|
"loss": 0.0758, |
|
"step": 1234 |
|
}, |
|
{ |
|
"epoch": 4.360105913503972, |
|
"grad_norm": 1.2476813765188703, |
|
"learning_rate": 6.259516131111957e-06, |
|
"loss": 0.0719, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 4.363636363636363, |
|
"grad_norm": 1.0504270751123217, |
|
"learning_rate": 6.253977449662047e-06, |
|
"loss": 0.0614, |
|
"step": 1236 |
|
}, |
|
{ |
|
"epoch": 4.367166813768756, |
|
"grad_norm": 1.1435259908690352, |
|
"learning_rate": 6.2484371257146e-06, |
|
"loss": 0.0768, |
|
"step": 1237 |
|
}, |
|
{ |
|
"epoch": 4.370697263901147, |
|
"grad_norm": 1.3840251434761308, |
|
"learning_rate": 6.242895166526502e-06, |
|
"loss": 0.0742, |
|
"step": 1238 |
|
}, |
|
{ |
|
"epoch": 4.374227714033539, |
|
"grad_norm": 1.2422183469141699, |
|
"learning_rate": 6.2373515793567715e-06, |
|
"loss": 0.08, |
|
"step": 1239 |
|
}, |
|
{ |
|
"epoch": 4.3777581641659316, |
|
"grad_norm": 1.1315889355948883, |
|
"learning_rate": 6.231806371466574e-06, |
|
"loss": 0.0697, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 4.381288614298323, |
|
"grad_norm": 1.081317701161799, |
|
"learning_rate": 6.226259550119187e-06, |
|
"loss": 0.0672, |
|
"step": 1241 |
|
}, |
|
{ |
|
"epoch": 4.384819064430715, |
|
"grad_norm": 1.139178949237585, |
|
"learning_rate": 6.220711122580004e-06, |
|
"loss": 0.0762, |
|
"step": 1242 |
|
}, |
|
{ |
|
"epoch": 4.388349514563107, |
|
"grad_norm": 1.2934745666925676, |
|
"learning_rate": 6.215161096116524e-06, |
|
"loss": 0.081, |
|
"step": 1243 |
|
}, |
|
{ |
|
"epoch": 4.391879964695499, |
|
"grad_norm": 1.4026385643923822, |
|
"learning_rate": 6.209609477998339e-06, |
|
"loss": 0.0759, |
|
"step": 1244 |
|
}, |
|
{ |
|
"epoch": 4.39541041482789, |
|
"grad_norm": 1.2849606159130602, |
|
"learning_rate": 6.2040562754971265e-06, |
|
"loss": 0.0713, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 4.398940864960283, |
|
"grad_norm": 1.0919624470047933, |
|
"learning_rate": 6.1985014958866386e-06, |
|
"loss": 0.0731, |
|
"step": 1246 |
|
}, |
|
{ |
|
"epoch": 4.402471315092674, |
|
"grad_norm": 1.2552988710153292, |
|
"learning_rate": 6.192945146442693e-06, |
|
"loss": 0.0802, |
|
"step": 1247 |
|
}, |
|
{ |
|
"epoch": 4.406001765225066, |
|
"grad_norm": 1.2944459168853082, |
|
"learning_rate": 6.187387234443164e-06, |
|
"loss": 0.084, |
|
"step": 1248 |
|
}, |
|
{ |
|
"epoch": 4.4095322153574585, |
|
"grad_norm": 1.1228808317446295, |
|
"learning_rate": 6.181827767167974e-06, |
|
"loss": 0.0716, |
|
"step": 1249 |
|
}, |
|
{ |
|
"epoch": 4.41306266548985, |
|
"grad_norm": 1.1881280477511678, |
|
"learning_rate": 6.176266751899078e-06, |
|
"loss": 0.067, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 4.416593115622242, |
|
"grad_norm": 1.1569504030576783, |
|
"learning_rate": 6.170704195920465e-06, |
|
"loss": 0.0702, |
|
"step": 1251 |
|
}, |
|
{ |
|
"epoch": 4.420123565754634, |
|
"grad_norm": 1.189002472009121, |
|
"learning_rate": 6.165140106518137e-06, |
|
"loss": 0.0694, |
|
"step": 1252 |
|
}, |
|
{ |
|
"epoch": 4.423654015887026, |
|
"grad_norm": 1.232505313641185, |
|
"learning_rate": 6.1595744909801085e-06, |
|
"loss": 0.0754, |
|
"step": 1253 |
|
}, |
|
{ |
|
"epoch": 4.427184466019417, |
|
"grad_norm": 1.2147460452464223, |
|
"learning_rate": 6.154007356596387e-06, |
|
"loss": 0.0652, |
|
"step": 1254 |
|
}, |
|
{ |
|
"epoch": 4.43071491615181, |
|
"grad_norm": 1.1480977785419901, |
|
"learning_rate": 6.148438710658979e-06, |
|
"loss": 0.0625, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 4.434245366284201, |
|
"grad_norm": 1.218331723091377, |
|
"learning_rate": 6.142868560461859e-06, |
|
"loss": 0.0821, |
|
"step": 1256 |
|
}, |
|
{ |
|
"epoch": 4.437775816416593, |
|
"grad_norm": 1.1738427410555219, |
|
"learning_rate": 6.137296913300982e-06, |
|
"loss": 0.074, |
|
"step": 1257 |
|
}, |
|
{ |
|
"epoch": 4.441306266548985, |
|
"grad_norm": 1.19855168208864, |
|
"learning_rate": 6.131723776474258e-06, |
|
"loss": 0.0738, |
|
"step": 1258 |
|
}, |
|
{ |
|
"epoch": 4.444836716681377, |
|
"grad_norm": 1.3103544915169332, |
|
"learning_rate": 6.12614915728155e-06, |
|
"loss": 0.0781, |
|
"step": 1259 |
|
}, |
|
{ |
|
"epoch": 4.448367166813769, |
|
"grad_norm": 1.0115359906207875, |
|
"learning_rate": 6.120573063024663e-06, |
|
"loss": 0.0715, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 4.451897616946161, |
|
"grad_norm": 0.9684920252271274, |
|
"learning_rate": 6.114995501007334e-06, |
|
"loss": 0.0626, |
|
"step": 1261 |
|
}, |
|
{ |
|
"epoch": 4.455428067078553, |
|
"grad_norm": 1.0836045218047463, |
|
"learning_rate": 6.109416478535222e-06, |
|
"loss": 0.0707, |
|
"step": 1262 |
|
}, |
|
{ |
|
"epoch": 4.458958517210944, |
|
"grad_norm": 1.1090014070135534, |
|
"learning_rate": 6.103836002915898e-06, |
|
"loss": 0.0638, |
|
"step": 1263 |
|
}, |
|
{ |
|
"epoch": 4.4624889673433366, |
|
"grad_norm": 1.300961015257033, |
|
"learning_rate": 6.098254081458839e-06, |
|
"loss": 0.0727, |
|
"step": 1264 |
|
}, |
|
{ |
|
"epoch": 4.466019417475728, |
|
"grad_norm": 1.2337280261124597, |
|
"learning_rate": 6.092670721475414e-06, |
|
"loss": 0.0748, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 4.46954986760812, |
|
"grad_norm": 1.1278101885787395, |
|
"learning_rate": 6.0870859302788756e-06, |
|
"loss": 0.0703, |
|
"step": 1266 |
|
}, |
|
{ |
|
"epoch": 4.473080317740512, |
|
"grad_norm": 1.0631423710066719, |
|
"learning_rate": 6.081499715184353e-06, |
|
"loss": 0.0706, |
|
"step": 1267 |
|
}, |
|
{ |
|
"epoch": 4.476610767872904, |
|
"grad_norm": 1.143193764341101, |
|
"learning_rate": 6.075912083508839e-06, |
|
"loss": 0.0694, |
|
"step": 1268 |
|
}, |
|
{ |
|
"epoch": 4.480141218005295, |
|
"grad_norm": 1.1879992175907763, |
|
"learning_rate": 6.070323042571184e-06, |
|
"loss": 0.0683, |
|
"step": 1269 |
|
}, |
|
{ |
|
"epoch": 4.483671668137688, |
|
"grad_norm": 1.4154753236711644, |
|
"learning_rate": 6.064732599692079e-06, |
|
"loss": 0.0841, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 4.48720211827008, |
|
"grad_norm": 1.0845224887804554, |
|
"learning_rate": 6.059140762194059e-06, |
|
"loss": 0.0753, |
|
"step": 1271 |
|
}, |
|
{ |
|
"epoch": 4.490732568402471, |
|
"grad_norm": 1.1903474371803968, |
|
"learning_rate": 6.053547537401477e-06, |
|
"loss": 0.0706, |
|
"step": 1272 |
|
}, |
|
{ |
|
"epoch": 4.4942630185348635, |
|
"grad_norm": 1.616858067887646, |
|
"learning_rate": 6.047952932640513e-06, |
|
"loss": 0.0912, |
|
"step": 1273 |
|
}, |
|
{ |
|
"epoch": 4.497793468667255, |
|
"grad_norm": 1.0913358558919948, |
|
"learning_rate": 6.042356955239145e-06, |
|
"loss": 0.0604, |
|
"step": 1274 |
|
}, |
|
{ |
|
"epoch": 4.501323918799647, |
|
"grad_norm": 1.1880905119942522, |
|
"learning_rate": 6.036759612527157e-06, |
|
"loss": 0.0759, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 4.504854368932039, |
|
"grad_norm": 1.1747169454730277, |
|
"learning_rate": 6.031160911836112e-06, |
|
"loss": 0.0679, |
|
"step": 1276 |
|
}, |
|
{ |
|
"epoch": 4.508384819064431, |
|
"grad_norm": 1.211427366915763, |
|
"learning_rate": 6.025560860499365e-06, |
|
"loss": 0.0689, |
|
"step": 1277 |
|
}, |
|
{ |
|
"epoch": 4.511915269196822, |
|
"grad_norm": 1.417931942561815, |
|
"learning_rate": 6.019959465852025e-06, |
|
"loss": 0.0868, |
|
"step": 1278 |
|
}, |
|
{ |
|
"epoch": 4.515445719329215, |
|
"grad_norm": 1.2209993004826432, |
|
"learning_rate": 6.014356735230973e-06, |
|
"loss": 0.0646, |
|
"step": 1279 |
|
}, |
|
{ |
|
"epoch": 4.518976169461606, |
|
"grad_norm": 1.1649085744269059, |
|
"learning_rate": 6.00875267597483e-06, |
|
"loss": 0.0639, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 4.522506619593998, |
|
"grad_norm": 1.011949914685396, |
|
"learning_rate": 6.003147295423967e-06, |
|
"loss": 0.0597, |
|
"step": 1281 |
|
}, |
|
{ |
|
"epoch": 4.5260370697263905, |
|
"grad_norm": 1.1309980576687069, |
|
"learning_rate": 5.997540600920479e-06, |
|
"loss": 0.0782, |
|
"step": 1282 |
|
}, |
|
{ |
|
"epoch": 4.529567519858782, |
|
"grad_norm": 1.0299510613826361, |
|
"learning_rate": 5.9919325998081825e-06, |
|
"loss": 0.066, |
|
"step": 1283 |
|
}, |
|
{ |
|
"epoch": 4.533097969991174, |
|
"grad_norm": 1.0789969131073938, |
|
"learning_rate": 5.986323299432608e-06, |
|
"loss": 0.0719, |
|
"step": 1284 |
|
}, |
|
{ |
|
"epoch": 4.536628420123566, |
|
"grad_norm": 1.0170912861541357, |
|
"learning_rate": 5.980712707140985e-06, |
|
"loss": 0.064, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 4.540158870255958, |
|
"grad_norm": 1.1901299269101344, |
|
"learning_rate": 5.97510083028224e-06, |
|
"loss": 0.0707, |
|
"step": 1286 |
|
}, |
|
{ |
|
"epoch": 4.543689320388349, |
|
"grad_norm": 1.236036942535051, |
|
"learning_rate": 5.969487676206975e-06, |
|
"loss": 0.0695, |
|
"step": 1287 |
|
}, |
|
{ |
|
"epoch": 4.5472197705207416, |
|
"grad_norm": 1.0739729905219972, |
|
"learning_rate": 5.9638732522674716e-06, |
|
"loss": 0.0683, |
|
"step": 1288 |
|
}, |
|
{ |
|
"epoch": 4.550750220653133, |
|
"grad_norm": 1.1896504864727457, |
|
"learning_rate": 5.95825756581767e-06, |
|
"loss": 0.0692, |
|
"step": 1289 |
|
}, |
|
{ |
|
"epoch": 4.554280670785525, |
|
"grad_norm": 1.0099956859061028, |
|
"learning_rate": 5.952640624213166e-06, |
|
"loss": 0.0717, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 4.557811120917917, |
|
"grad_norm": 1.1995738747246374, |
|
"learning_rate": 5.947022434811202e-06, |
|
"loss": 0.0611, |
|
"step": 1291 |
|
}, |
|
{ |
|
"epoch": 4.561341571050309, |
|
"grad_norm": 1.6264487607523304, |
|
"learning_rate": 5.94140300497065e-06, |
|
"loss": 0.0817, |
|
"step": 1292 |
|
}, |
|
{ |
|
"epoch": 4.564872021182701, |
|
"grad_norm": 1.2030247661685802, |
|
"learning_rate": 5.935782342052007e-06, |
|
"loss": 0.0605, |
|
"step": 1293 |
|
}, |
|
{ |
|
"epoch": 4.568402471315093, |
|
"grad_norm": 1.117896941640317, |
|
"learning_rate": 5.93016045341739e-06, |
|
"loss": 0.0521, |
|
"step": 1294 |
|
}, |
|
{ |
|
"epoch": 4.571932921447485, |
|
"grad_norm": 1.282540263365002, |
|
"learning_rate": 5.924537346430517e-06, |
|
"loss": 0.0719, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 4.575463371579876, |
|
"grad_norm": 1.4173574483743017, |
|
"learning_rate": 5.918913028456705e-06, |
|
"loss": 0.0665, |
|
"step": 1296 |
|
}, |
|
{ |
|
"epoch": 4.5789938217122685, |
|
"grad_norm": 1.7378141839355061, |
|
"learning_rate": 5.9132875068628555e-06, |
|
"loss": 0.082, |
|
"step": 1297 |
|
}, |
|
{ |
|
"epoch": 4.58252427184466, |
|
"grad_norm": 1.5474239856959229, |
|
"learning_rate": 5.907660789017446e-06, |
|
"loss": 0.0747, |
|
"step": 1298 |
|
}, |
|
{ |
|
"epoch": 4.586054721977052, |
|
"grad_norm": 1.2885896943234003, |
|
"learning_rate": 5.902032882290518e-06, |
|
"loss": 0.0744, |
|
"step": 1299 |
|
}, |
|
{ |
|
"epoch": 4.589585172109444, |
|
"grad_norm": 1.3460860331275364, |
|
"learning_rate": 5.896403794053679e-06, |
|
"loss": 0.0767, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 4.593115622241836, |
|
"grad_norm": 1.1096734280955396, |
|
"learning_rate": 5.890773531680076e-06, |
|
"loss": 0.0777, |
|
"step": 1301 |
|
}, |
|
{ |
|
"epoch": 4.596646072374227, |
|
"grad_norm": 1.1779234981799707, |
|
"learning_rate": 5.885142102544398e-06, |
|
"loss": 0.0636, |
|
"step": 1302 |
|
}, |
|
{ |
|
"epoch": 4.60017652250662, |
|
"grad_norm": 1.3624220771981437, |
|
"learning_rate": 5.8795095140228595e-06, |
|
"loss": 0.0894, |
|
"step": 1303 |
|
}, |
|
{ |
|
"epoch": 4.603706972639012, |
|
"grad_norm": 1.343188625919169, |
|
"learning_rate": 5.8738757734931975e-06, |
|
"loss": 0.0809, |
|
"step": 1304 |
|
}, |
|
{ |
|
"epoch": 4.607237422771403, |
|
"grad_norm": 1.4446432282791357, |
|
"learning_rate": 5.8682408883346535e-06, |
|
"loss": 0.0766, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 4.6107678729037955, |
|
"grad_norm": 1.3556894299906552, |
|
"learning_rate": 5.86260486592797e-06, |
|
"loss": 0.0797, |
|
"step": 1306 |
|
}, |
|
{ |
|
"epoch": 4.614298323036187, |
|
"grad_norm": 1.278747621381562, |
|
"learning_rate": 5.856967713655379e-06, |
|
"loss": 0.0833, |
|
"step": 1307 |
|
}, |
|
{ |
|
"epoch": 4.617828773168579, |
|
"grad_norm": 1.1918217353301246, |
|
"learning_rate": 5.851329438900595e-06, |
|
"loss": 0.0799, |
|
"step": 1308 |
|
}, |
|
{ |
|
"epoch": 4.621359223300971, |
|
"grad_norm": 1.1278725314966394, |
|
"learning_rate": 5.845690049048799e-06, |
|
"loss": 0.0643, |
|
"step": 1309 |
|
}, |
|
{ |
|
"epoch": 4.624889673433363, |
|
"grad_norm": 1.2557557372315369, |
|
"learning_rate": 5.8400495514866355e-06, |
|
"loss": 0.0879, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 4.628420123565754, |
|
"grad_norm": 1.2461622162231867, |
|
"learning_rate": 5.8344079536021955e-06, |
|
"loss": 0.0809, |
|
"step": 1311 |
|
}, |
|
{ |
|
"epoch": 4.631950573698147, |
|
"grad_norm": 1.2846191093338162, |
|
"learning_rate": 5.82876526278502e-06, |
|
"loss": 0.0795, |
|
"step": 1312 |
|
}, |
|
{ |
|
"epoch": 4.635481023830538, |
|
"grad_norm": 1.1240893202153388, |
|
"learning_rate": 5.823121486426069e-06, |
|
"loss": 0.073, |
|
"step": 1313 |
|
}, |
|
{ |
|
"epoch": 4.63901147396293, |
|
"grad_norm": 1.2575691959822162, |
|
"learning_rate": 5.817476631917736e-06, |
|
"loss": 0.0843, |
|
"step": 1314 |
|
}, |
|
{ |
|
"epoch": 4.6425419240953225, |
|
"grad_norm": 1.017222394290193, |
|
"learning_rate": 5.811830706653819e-06, |
|
"loss": 0.0673, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 4.646072374227714, |
|
"grad_norm": 1.0729116322930856, |
|
"learning_rate": 5.806183718029524e-06, |
|
"loss": 0.0732, |
|
"step": 1316 |
|
}, |
|
{ |
|
"epoch": 4.649602824360106, |
|
"grad_norm": 1.1579892771980353, |
|
"learning_rate": 5.800535673441446e-06, |
|
"loss": 0.0775, |
|
"step": 1317 |
|
}, |
|
{ |
|
"epoch": 4.653133274492498, |
|
"grad_norm": 1.0293530970686642, |
|
"learning_rate": 5.794886580287565e-06, |
|
"loss": 0.0636, |
|
"step": 1318 |
|
}, |
|
{ |
|
"epoch": 4.65666372462489, |
|
"grad_norm": 0.9376266948248516, |
|
"learning_rate": 5.789236445967233e-06, |
|
"loss": 0.0603, |
|
"step": 1319 |
|
}, |
|
{ |
|
"epoch": 4.660194174757281, |
|
"grad_norm": 0.9987883216154062, |
|
"learning_rate": 5.783585277881167e-06, |
|
"loss": 0.0762, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 4.6637246248896735, |
|
"grad_norm": 1.2329836983396925, |
|
"learning_rate": 5.777933083431437e-06, |
|
"loss": 0.0831, |
|
"step": 1321 |
|
}, |
|
{ |
|
"epoch": 4.667255075022065, |
|
"grad_norm": 1.1400090314981657, |
|
"learning_rate": 5.77227987002146e-06, |
|
"loss": 0.0715, |
|
"step": 1322 |
|
}, |
|
{ |
|
"epoch": 4.670785525154457, |
|
"grad_norm": 1.108926279713076, |
|
"learning_rate": 5.766625645055982e-06, |
|
"loss": 0.0727, |
|
"step": 1323 |
|
}, |
|
{ |
|
"epoch": 4.674315975286849, |
|
"grad_norm": 1.036285195186968, |
|
"learning_rate": 5.760970415941082e-06, |
|
"loss": 0.0619, |
|
"step": 1324 |
|
}, |
|
{ |
|
"epoch": 4.677846425419241, |
|
"grad_norm": 1.2377457117763253, |
|
"learning_rate": 5.755314190084146e-06, |
|
"loss": 0.0679, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 4.681376875551633, |
|
"grad_norm": 1.2816857024923805, |
|
"learning_rate": 5.749656974893873e-06, |
|
"loss": 0.0761, |
|
"step": 1326 |
|
}, |
|
{ |
|
"epoch": 4.684907325684025, |
|
"grad_norm": 1.4320175248552778, |
|
"learning_rate": 5.743998777780252e-06, |
|
"loss": 0.0735, |
|
"step": 1327 |
|
}, |
|
{ |
|
"epoch": 4.688437775816417, |
|
"grad_norm": 1.21805904446881, |
|
"learning_rate": 5.73833960615456e-06, |
|
"loss": 0.0654, |
|
"step": 1328 |
|
}, |
|
{ |
|
"epoch": 4.691968225948808, |
|
"grad_norm": 1.2720158896144405, |
|
"learning_rate": 5.732679467429353e-06, |
|
"loss": 0.0872, |
|
"step": 1329 |
|
}, |
|
{ |
|
"epoch": 4.6954986760812005, |
|
"grad_norm": 1.362220253742237, |
|
"learning_rate": 5.72701836901845e-06, |
|
"loss": 0.0763, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 4.699029126213592, |
|
"grad_norm": 1.3376103897361957, |
|
"learning_rate": 5.7213563183369295e-06, |
|
"loss": 0.0718, |
|
"step": 1331 |
|
}, |
|
{ |
|
"epoch": 4.702559576345984, |
|
"grad_norm": 1.2102663618571865, |
|
"learning_rate": 5.715693322801117e-06, |
|
"loss": 0.0738, |
|
"step": 1332 |
|
}, |
|
{ |
|
"epoch": 4.7060900264783765, |
|
"grad_norm": 1.2277058279665478, |
|
"learning_rate": 5.710029389828572e-06, |
|
"loss": 0.0719, |
|
"step": 1333 |
|
}, |
|
{ |
|
"epoch": 4.709620476610768, |
|
"grad_norm": 1.2773556567896192, |
|
"learning_rate": 5.70436452683809e-06, |
|
"loss": 0.0902, |
|
"step": 1334 |
|
}, |
|
{ |
|
"epoch": 4.713150926743159, |
|
"grad_norm": 1.4332928684103645, |
|
"learning_rate": 5.698698741249676e-06, |
|
"loss": 0.0709, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 4.716681376875552, |
|
"grad_norm": 1.1461862595511798, |
|
"learning_rate": 5.6930320404845475e-06, |
|
"loss": 0.0689, |
|
"step": 1336 |
|
}, |
|
{ |
|
"epoch": 4.720211827007944, |
|
"grad_norm": 1.141265802102587, |
|
"learning_rate": 5.687364431965121e-06, |
|
"loss": 0.0836, |
|
"step": 1337 |
|
}, |
|
{ |
|
"epoch": 4.723742277140335, |
|
"grad_norm": 1.503230303094657, |
|
"learning_rate": 5.681695923115002e-06, |
|
"loss": 0.0819, |
|
"step": 1338 |
|
}, |
|
{ |
|
"epoch": 4.7272727272727275, |
|
"grad_norm": 1.1626902290316583, |
|
"learning_rate": 5.676026521358973e-06, |
|
"loss": 0.0781, |
|
"step": 1339 |
|
}, |
|
{ |
|
"epoch": 4.730803177405119, |
|
"grad_norm": 1.1620016918977858, |
|
"learning_rate": 5.6703562341229886e-06, |
|
"loss": 0.0805, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 4.734333627537511, |
|
"grad_norm": 0.980205333032981, |
|
"learning_rate": 5.664685068834164e-06, |
|
"loss": 0.06, |
|
"step": 1341 |
|
}, |
|
{ |
|
"epoch": 4.737864077669903, |
|
"grad_norm": 1.0919120901198711, |
|
"learning_rate": 5.659013032920758e-06, |
|
"loss": 0.0667, |
|
"step": 1342 |
|
}, |
|
{ |
|
"epoch": 4.741394527802295, |
|
"grad_norm": 1.058369377556783, |
|
"learning_rate": 5.653340133812179e-06, |
|
"loss": 0.0678, |
|
"step": 1343 |
|
}, |
|
{ |
|
"epoch": 4.744924977934687, |
|
"grad_norm": 1.2729504832126617, |
|
"learning_rate": 5.647666378938959e-06, |
|
"loss": 0.0746, |
|
"step": 1344 |
|
}, |
|
{ |
|
"epoch": 4.7484554280670785, |
|
"grad_norm": 1.2480081282387157, |
|
"learning_rate": 5.641991775732756e-06, |
|
"loss": 0.0711, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 4.75198587819947, |
|
"grad_norm": 1.1576335449365858, |
|
"learning_rate": 5.636316331626332e-06, |
|
"loss": 0.074, |
|
"step": 1346 |
|
}, |
|
{ |
|
"epoch": 4.755516328331862, |
|
"grad_norm": 1.1637647202647143, |
|
"learning_rate": 5.630640054053562e-06, |
|
"loss": 0.0695, |
|
"step": 1347 |
|
}, |
|
{ |
|
"epoch": 4.7590467784642545, |
|
"grad_norm": 1.1265709555415027, |
|
"learning_rate": 5.624962950449402e-06, |
|
"loss": 0.0716, |
|
"step": 1348 |
|
}, |
|
{ |
|
"epoch": 4.762577228596646, |
|
"grad_norm": 1.2129102271686025, |
|
"learning_rate": 5.6192850282498925e-06, |
|
"loss": 0.0728, |
|
"step": 1349 |
|
}, |
|
{ |
|
"epoch": 4.766107678729038, |
|
"grad_norm": 1.4098311899227094, |
|
"learning_rate": 5.6136062948921486e-06, |
|
"loss": 0.0695, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 4.76963812886143, |
|
"grad_norm": 1.0254747813352643, |
|
"learning_rate": 5.607926757814347e-06, |
|
"loss": 0.072, |
|
"step": 1351 |
|
}, |
|
{ |
|
"epoch": 4.773168578993822, |
|
"grad_norm": 1.3062783164855614, |
|
"learning_rate": 5.6022464244557175e-06, |
|
"loss": 0.0836, |
|
"step": 1352 |
|
}, |
|
{ |
|
"epoch": 4.776699029126213, |
|
"grad_norm": 1.1787618293744981, |
|
"learning_rate": 5.596565302256531e-06, |
|
"loss": 0.0644, |
|
"step": 1353 |
|
}, |
|
{ |
|
"epoch": 4.7802294792586055, |
|
"grad_norm": 1.4645406649621868, |
|
"learning_rate": 5.590883398658095e-06, |
|
"loss": 0.0786, |
|
"step": 1354 |
|
}, |
|
{ |
|
"epoch": 4.783759929390998, |
|
"grad_norm": 1.3381703719371458, |
|
"learning_rate": 5.585200721102737e-06, |
|
"loss": 0.0855, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 4.787290379523389, |
|
"grad_norm": 1.5162385054662686, |
|
"learning_rate": 5.579517277033799e-06, |
|
"loss": 0.0727, |
|
"step": 1356 |
|
}, |
|
{ |
|
"epoch": 4.7908208296557815, |
|
"grad_norm": 1.333030859041725, |
|
"learning_rate": 5.573833073895629e-06, |
|
"loss": 0.0872, |
|
"step": 1357 |
|
}, |
|
{ |
|
"epoch": 4.794351279788173, |
|
"grad_norm": 1.2513683261705562, |
|
"learning_rate": 5.568148119133566e-06, |
|
"loss": 0.0661, |
|
"step": 1358 |
|
}, |
|
{ |
|
"epoch": 4.797881729920565, |
|
"grad_norm": 1.204520949412178, |
|
"learning_rate": 5.562462420193939e-06, |
|
"loss": 0.0769, |
|
"step": 1359 |
|
}, |
|
{ |
|
"epoch": 4.801412180052957, |
|
"grad_norm": 1.2922206835673506, |
|
"learning_rate": 5.556775984524044e-06, |
|
"loss": 0.0718, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 4.804942630185349, |
|
"grad_norm": 1.4591640646940425, |
|
"learning_rate": 5.55108881957215e-06, |
|
"loss": 0.0648, |
|
"step": 1361 |
|
}, |
|
{ |
|
"epoch": 4.80847308031774, |
|
"grad_norm": 1.2720780355230428, |
|
"learning_rate": 5.545400932787476e-06, |
|
"loss": 0.0735, |
|
"step": 1362 |
|
}, |
|
{ |
|
"epoch": 4.8120035304501325, |
|
"grad_norm": 1.2599249791592424, |
|
"learning_rate": 5.539712331620186e-06, |
|
"loss": 0.062, |
|
"step": 1363 |
|
}, |
|
{ |
|
"epoch": 4.815533980582524, |
|
"grad_norm": 1.0986643518131567, |
|
"learning_rate": 5.534023023521384e-06, |
|
"loss": 0.0701, |
|
"step": 1364 |
|
}, |
|
{ |
|
"epoch": 4.819064430714916, |
|
"grad_norm": 1.194220635218748, |
|
"learning_rate": 5.528333015943097e-06, |
|
"loss": 0.0663, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 4.8225948808473085, |
|
"grad_norm": 1.1823226864675451, |
|
"learning_rate": 5.522642316338268e-06, |
|
"loss": 0.059, |
|
"step": 1366 |
|
}, |
|
{ |
|
"epoch": 4.8261253309797, |
|
"grad_norm": 1.225355888651712, |
|
"learning_rate": 5.516950932160748e-06, |
|
"loss": 0.0755, |
|
"step": 1367 |
|
}, |
|
{ |
|
"epoch": 4.829655781112092, |
|
"grad_norm": 1.417142754172787, |
|
"learning_rate": 5.511258870865283e-06, |
|
"loss": 0.0907, |
|
"step": 1368 |
|
}, |
|
{ |
|
"epoch": 4.8331862312444835, |
|
"grad_norm": 1.315652572963653, |
|
"learning_rate": 5.50556613990751e-06, |
|
"loss": 0.0807, |
|
"step": 1369 |
|
}, |
|
{ |
|
"epoch": 4.836716681376876, |
|
"grad_norm": 1.4848341875180038, |
|
"learning_rate": 5.499872746743935e-06, |
|
"loss": 0.1056, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 4.840247131509267, |
|
"grad_norm": 1.5291378895324677, |
|
"learning_rate": 5.49417869883194e-06, |
|
"loss": 0.0837, |
|
"step": 1371 |
|
}, |
|
{ |
|
"epoch": 4.8437775816416595, |
|
"grad_norm": 1.226839957471945, |
|
"learning_rate": 5.488484003629759e-06, |
|
"loss": 0.0779, |
|
"step": 1372 |
|
}, |
|
{ |
|
"epoch": 4.847308031774051, |
|
"grad_norm": 1.3463992269178928, |
|
"learning_rate": 5.4827886685964774e-06, |
|
"loss": 0.0976, |
|
"step": 1373 |
|
}, |
|
{ |
|
"epoch": 4.850838481906443, |
|
"grad_norm": 1.1494326044050664, |
|
"learning_rate": 5.477092701192014e-06, |
|
"loss": 0.0649, |
|
"step": 1374 |
|
}, |
|
{ |
|
"epoch": 4.854368932038835, |
|
"grad_norm": 1.2447420578267399, |
|
"learning_rate": 5.471396108877123e-06, |
|
"loss": 0.072, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 4.857899382171227, |
|
"grad_norm": 1.3612258681021532, |
|
"learning_rate": 5.465698899113371e-06, |
|
"loss": 0.0807, |
|
"step": 1376 |
|
}, |
|
{ |
|
"epoch": 4.861429832303619, |
|
"grad_norm": 1.4193052420845504, |
|
"learning_rate": 5.460001079363135e-06, |
|
"loss": 0.0811, |
|
"step": 1377 |
|
}, |
|
{ |
|
"epoch": 4.8649602824360105, |
|
"grad_norm": 1.2743805042555498, |
|
"learning_rate": 5.454302657089592e-06, |
|
"loss": 0.08, |
|
"step": 1378 |
|
}, |
|
{ |
|
"epoch": 4.868490732568403, |
|
"grad_norm": 1.078809447494733, |
|
"learning_rate": 5.448603639756708e-06, |
|
"loss": 0.062, |
|
"step": 1379 |
|
}, |
|
{ |
|
"epoch": 4.872021182700794, |
|
"grad_norm": 1.0799797925482022, |
|
"learning_rate": 5.442904034829226e-06, |
|
"loss": 0.075, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 4.8755516328331865, |
|
"grad_norm": 1.141767854824888, |
|
"learning_rate": 5.437203849772664e-06, |
|
"loss": 0.0636, |
|
"step": 1381 |
|
}, |
|
{ |
|
"epoch": 4.879082082965578, |
|
"grad_norm": 1.123958849288234, |
|
"learning_rate": 5.431503092053294e-06, |
|
"loss": 0.077, |
|
"step": 1382 |
|
}, |
|
{ |
|
"epoch": 4.88261253309797, |
|
"grad_norm": 0.9850627842600758, |
|
"learning_rate": 5.4258017691381415e-06, |
|
"loss": 0.0602, |
|
"step": 1383 |
|
}, |
|
{ |
|
"epoch": 4.886142983230362, |
|
"grad_norm": 1.1111371089369984, |
|
"learning_rate": 5.420099888494972e-06, |
|
"loss": 0.072, |
|
"step": 1384 |
|
}, |
|
{ |
|
"epoch": 4.889673433362754, |
|
"grad_norm": 1.116994328139007, |
|
"learning_rate": 5.414397457592278e-06, |
|
"loss": 0.0769, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 4.893203883495145, |
|
"grad_norm": 1.4416102109718356, |
|
"learning_rate": 5.408694483899281e-06, |
|
"loss": 0.0813, |
|
"step": 1386 |
|
}, |
|
{ |
|
"epoch": 4.8967343336275375, |
|
"grad_norm": 1.4613051629608635, |
|
"learning_rate": 5.402990974885904e-06, |
|
"loss": 0.0821, |
|
"step": 1387 |
|
}, |
|
{ |
|
"epoch": 4.90026478375993, |
|
"grad_norm": 1.1868315059556709, |
|
"learning_rate": 5.3972869380227765e-06, |
|
"loss": 0.0804, |
|
"step": 1388 |
|
}, |
|
{ |
|
"epoch": 4.903795233892321, |
|
"grad_norm": 1.1666204562256506, |
|
"learning_rate": 5.3915823807812196e-06, |
|
"loss": 0.0785, |
|
"step": 1389 |
|
}, |
|
{ |
|
"epoch": 4.9073256840247135, |
|
"grad_norm": 1.1215733098359313, |
|
"learning_rate": 5.385877310633233e-06, |
|
"loss": 0.0745, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 4.910856134157105, |
|
"grad_norm": 1.585401934451052, |
|
"learning_rate": 5.38017173505149e-06, |
|
"loss": 0.0797, |
|
"step": 1391 |
|
}, |
|
{ |
|
"epoch": 4.914386584289497, |
|
"grad_norm": 1.3919507079497462, |
|
"learning_rate": 5.374465661509326e-06, |
|
"loss": 0.0853, |
|
"step": 1392 |
|
}, |
|
{ |
|
"epoch": 4.9179170344218885, |
|
"grad_norm": 1.1020309883219317, |
|
"learning_rate": 5.368759097480728e-06, |
|
"loss": 0.0738, |
|
"step": 1393 |
|
}, |
|
{ |
|
"epoch": 4.921447484554281, |
|
"grad_norm": 1.2530954184669498, |
|
"learning_rate": 5.363052050440327e-06, |
|
"loss": 0.0649, |
|
"step": 1394 |
|
}, |
|
{ |
|
"epoch": 4.924977934686672, |
|
"grad_norm": 1.3563436418670745, |
|
"learning_rate": 5.357344527863385e-06, |
|
"loss": 0.0848, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 4.9285083848190645, |
|
"grad_norm": 1.2161047750969436, |
|
"learning_rate": 5.35163653722579e-06, |
|
"loss": 0.0711, |
|
"step": 1396 |
|
}, |
|
{ |
|
"epoch": 4.932038834951456, |
|
"grad_norm": 1.1405877554596378, |
|
"learning_rate": 5.345928086004034e-06, |
|
"loss": 0.0759, |
|
"step": 1397 |
|
}, |
|
{ |
|
"epoch": 4.935569285083848, |
|
"grad_norm": 1.5340642291990807, |
|
"learning_rate": 5.340219181675227e-06, |
|
"loss": 0.0756, |
|
"step": 1398 |
|
}, |
|
{ |
|
"epoch": 4.9390997352162405, |
|
"grad_norm": 1.4009707021891677, |
|
"learning_rate": 5.334509831717058e-06, |
|
"loss": 0.0773, |
|
"step": 1399 |
|
}, |
|
{ |
|
"epoch": 4.942630185348632, |
|
"grad_norm": 1.2971256965481643, |
|
"learning_rate": 5.328800043607809e-06, |
|
"loss": 0.0891, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 4.946160635481024, |
|
"grad_norm": 1.378471603765613, |
|
"learning_rate": 5.323089824826332e-06, |
|
"loss": 0.0812, |
|
"step": 1401 |
|
}, |
|
{ |
|
"epoch": 4.9496910856134155, |
|
"grad_norm": 0.9246558166305276, |
|
"learning_rate": 5.317379182852044e-06, |
|
"loss": 0.0652, |
|
"step": 1402 |
|
}, |
|
{ |
|
"epoch": 4.953221535745808, |
|
"grad_norm": 1.1283491586166101, |
|
"learning_rate": 5.311668125164916e-06, |
|
"loss": 0.0747, |
|
"step": 1403 |
|
}, |
|
{ |
|
"epoch": 4.956751985878199, |
|
"grad_norm": 1.1947879187764758, |
|
"learning_rate": 5.305956659245466e-06, |
|
"loss": 0.0871, |
|
"step": 1404 |
|
}, |
|
{ |
|
"epoch": 4.9602824360105915, |
|
"grad_norm": 1.1562516815747672, |
|
"learning_rate": 5.300244792574743e-06, |
|
"loss": 0.0818, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 4.963812886142983, |
|
"grad_norm": 1.2658874903814799, |
|
"learning_rate": 5.294532532634321e-06, |
|
"loss": 0.0784, |
|
"step": 1406 |
|
}, |
|
{ |
|
"epoch": 4.967343336275375, |
|
"grad_norm": 1.0490558876913674, |
|
"learning_rate": 5.288819886906293e-06, |
|
"loss": 0.0618, |
|
"step": 1407 |
|
}, |
|
{ |
|
"epoch": 4.970873786407767, |
|
"grad_norm": 1.2147121704532486, |
|
"learning_rate": 5.283106862873253e-06, |
|
"loss": 0.0799, |
|
"step": 1408 |
|
}, |
|
{ |
|
"epoch": 4.974404236540159, |
|
"grad_norm": 1.1451504862662256, |
|
"learning_rate": 5.2773934680182935e-06, |
|
"loss": 0.0668, |
|
"step": 1409 |
|
}, |
|
{ |
|
"epoch": 4.977934686672551, |
|
"grad_norm": 1.1648655228082387, |
|
"learning_rate": 5.271679709824993e-06, |
|
"loss": 0.0742, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 4.9814651368049425, |
|
"grad_norm": 1.1046312634759032, |
|
"learning_rate": 5.265965595777402e-06, |
|
"loss": 0.0814, |
|
"step": 1411 |
|
}, |
|
{ |
|
"epoch": 4.984995586937335, |
|
"grad_norm": 1.1273037101841474, |
|
"learning_rate": 5.2602511333600414e-06, |
|
"loss": 0.0785, |
|
"step": 1412 |
|
}, |
|
{ |
|
"epoch": 4.988526037069726, |
|
"grad_norm": 1.354622519190956, |
|
"learning_rate": 5.254536330057884e-06, |
|
"loss": 0.0785, |
|
"step": 1413 |
|
}, |
|
{ |
|
"epoch": 4.9920564872021185, |
|
"grad_norm": 1.3583127081670356, |
|
"learning_rate": 5.248821193356356e-06, |
|
"loss": 0.0789, |
|
"step": 1414 |
|
}, |
|
{ |
|
"epoch": 4.99558693733451, |
|
"grad_norm": 1.2297540102464188, |
|
"learning_rate": 5.243105730741312e-06, |
|
"loss": 0.0782, |
|
"step": 1415 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 2830, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 1415, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 110584559550464.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|