{ "best_metric": 1.6337618827819824, "best_model_checkpoint": "outputs/checkpoint-298", "epoch": 0.9987368421052631, "eval_steps": 149, "global_step": 593, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0016842105263157896, "grad_norm": 0.6960500478744507, "learning_rate": 0.001, "loss": 2.2378, "step": 1 }, { "epoch": 0.003368421052631579, "grad_norm": 1.7757582664489746, "learning_rate": 0.001, "loss": 2.3598, "step": 2 }, { "epoch": 0.0050526315789473685, "grad_norm": 4.837254047393799, "learning_rate": 0.001, "loss": 3.0649, "step": 3 }, { "epoch": 0.006736842105263158, "grad_norm": 1.8388794660568237, "learning_rate": 0.001, "loss": 2.4077, "step": 4 }, { "epoch": 0.008421052631578947, "grad_norm": 2.6212971210479736, "learning_rate": 0.001, "loss": 1.8794, "step": 5 }, { "epoch": 0.010105263157894737, "grad_norm": 3.2032198905944824, "learning_rate": 0.001, "loss": 2.2291, "step": 6 }, { "epoch": 0.011789473684210527, "grad_norm": 0.9379774332046509, "learning_rate": 0.001, "loss": 1.9836, "step": 7 }, { "epoch": 0.013473684210526317, "grad_norm": 0.8930391073226929, "learning_rate": 0.001, "loss": 1.9855, "step": 8 }, { "epoch": 0.015157894736842105, "grad_norm": 1.121151089668274, "learning_rate": 0.001, "loss": 2.142, "step": 9 }, { "epoch": 0.016842105263157894, "grad_norm": 1.3604674339294434, "learning_rate": 0.001, "loss": 2.0745, "step": 10 }, { "epoch": 0.018526315789473686, "grad_norm": 0.9475807547569275, "learning_rate": 0.001, "loss": 2.2516, "step": 11 }, { "epoch": 0.020210526315789474, "grad_norm": 1.56779146194458, "learning_rate": 0.001, "loss": 1.9924, "step": 12 }, { "epoch": 0.021894736842105262, "grad_norm": 1.14510178565979, "learning_rate": 0.001, "loss": 1.8669, "step": 13 }, { "epoch": 0.023578947368421053, "grad_norm": 0.9147089719772339, "learning_rate": 0.001, "loss": 2.1365, "step": 14 }, { "epoch": 0.02526315789473684, "grad_norm": 0.7278650403022766, "learning_rate": 0.001, "loss": 2.2927, "step": 15 }, { "epoch": 0.026947368421052633, "grad_norm": 0.6441658735275269, "learning_rate": 0.001, "loss": 2.1152, "step": 16 }, { "epoch": 0.02863157894736842, "grad_norm": 0.6069326996803284, "learning_rate": 0.001, "loss": 1.8998, "step": 17 }, { "epoch": 0.03031578947368421, "grad_norm": 0.7715803384780884, "learning_rate": 0.001, "loss": 2.3267, "step": 18 }, { "epoch": 0.032, "grad_norm": 1.274883508682251, "learning_rate": 0.001, "loss": 1.9622, "step": 19 }, { "epoch": 0.03368421052631579, "grad_norm": 1.315987467765808, "learning_rate": 0.001, "loss": 2.0188, "step": 20 }, { "epoch": 0.03536842105263158, "grad_norm": 0.7619612216949463, "learning_rate": 0.001, "loss": 2.1692, "step": 21 }, { "epoch": 0.03705263157894737, "grad_norm": 0.6206510066986084, "learning_rate": 0.001, "loss": 1.9139, "step": 22 }, { "epoch": 0.03873684210526316, "grad_norm": 0.7554602026939392, "learning_rate": 0.001, "loss": 2.0588, "step": 23 }, { "epoch": 0.04042105263157895, "grad_norm": 0.6345531344413757, "learning_rate": 0.001, "loss": 2.1671, "step": 24 }, { "epoch": 0.042105263157894736, "grad_norm": 0.6695383191108704, "learning_rate": 0.001, "loss": 2.02, "step": 25 }, { "epoch": 0.043789473684210524, "grad_norm": 0.4833696186542511, "learning_rate": 0.001, "loss": 1.75, "step": 26 }, { "epoch": 0.04547368421052632, "grad_norm": 0.7931104302406311, "learning_rate": 0.001, "loss": 2.1626, "step": 27 }, { "epoch": 0.04715789473684211, "grad_norm": 1.5443921089172363, "learning_rate": 0.001, "loss": 2.0443, "step": 28 }, { "epoch": 0.048842105263157895, "grad_norm": 0.6693266034126282, "learning_rate": 0.001, "loss": 2.2165, "step": 29 }, { "epoch": 0.05052631578947368, "grad_norm": 0.6431847810745239, "learning_rate": 0.001, "loss": 2.1791, "step": 30 }, { "epoch": 0.05221052631578947, "grad_norm": 0.5706735253334045, "learning_rate": 0.001, "loss": 1.9472, "step": 31 }, { "epoch": 0.053894736842105266, "grad_norm": 0.7617158889770508, "learning_rate": 0.001, "loss": 2.0268, "step": 32 }, { "epoch": 0.055578947368421054, "grad_norm": 0.5798472762107849, "learning_rate": 0.001, "loss": 1.9278, "step": 33 }, { "epoch": 0.05726315789473684, "grad_norm": 0.5660713315010071, "learning_rate": 0.001, "loss": 1.7519, "step": 34 }, { "epoch": 0.05894736842105263, "grad_norm": 0.5779318809509277, "learning_rate": 0.001, "loss": 1.9023, "step": 35 }, { "epoch": 0.06063157894736842, "grad_norm": 0.6591325998306274, "learning_rate": 0.001, "loss": 1.959, "step": 36 }, { "epoch": 0.06231578947368421, "grad_norm": 0.6363794803619385, "learning_rate": 0.001, "loss": 2.1155, "step": 37 }, { "epoch": 0.064, "grad_norm": 0.6779230833053589, "learning_rate": 0.001, "loss": 1.9465, "step": 38 }, { "epoch": 0.06568421052631579, "grad_norm": 0.6995664238929749, "learning_rate": 0.001, "loss": 1.8616, "step": 39 }, { "epoch": 0.06736842105263158, "grad_norm": 0.7718273997306824, "learning_rate": 0.001, "loss": 2.1503, "step": 40 }, { "epoch": 0.06905263157894737, "grad_norm": 0.5334057807922363, "learning_rate": 0.001, "loss": 1.7283, "step": 41 }, { "epoch": 0.07073684210526315, "grad_norm": 0.5298858880996704, "learning_rate": 0.001, "loss": 1.6779, "step": 42 }, { "epoch": 0.07242105263157894, "grad_norm": 0.5175891518592834, "learning_rate": 0.001, "loss": 1.9119, "step": 43 }, { "epoch": 0.07410526315789474, "grad_norm": 0.5781148076057434, "learning_rate": 0.001, "loss": 1.7144, "step": 44 }, { "epoch": 0.07578947368421053, "grad_norm": 0.5868591666221619, "learning_rate": 0.001, "loss": 1.8802, "step": 45 }, { "epoch": 0.07747368421052632, "grad_norm": 0.5802803635597229, "learning_rate": 0.001, "loss": 1.7823, "step": 46 }, { "epoch": 0.07915789473684211, "grad_norm": 0.6214857697486877, "learning_rate": 0.001, "loss": 1.6589, "step": 47 }, { "epoch": 0.0808421052631579, "grad_norm": 0.6045804619789124, "learning_rate": 0.001, "loss": 1.9772, "step": 48 }, { "epoch": 0.08252631578947368, "grad_norm": 0.6322258114814758, "learning_rate": 0.001, "loss": 1.7932, "step": 49 }, { "epoch": 0.08421052631578947, "grad_norm": 0.6137242317199707, "learning_rate": 0.001, "loss": 1.8022, "step": 50 }, { "epoch": 0.08589473684210526, "grad_norm": 0.5677167773246765, "learning_rate": 0.001, "loss": 1.5019, "step": 51 }, { "epoch": 0.08757894736842105, "grad_norm": 0.6881672143936157, "learning_rate": 0.001, "loss": 1.9834, "step": 52 }, { "epoch": 0.08926315789473684, "grad_norm": 0.6887582540512085, "learning_rate": 0.001, "loss": 2.0563, "step": 53 }, { "epoch": 0.09094736842105264, "grad_norm": 0.6267439723014832, "learning_rate": 0.001, "loss": 1.9473, "step": 54 }, { "epoch": 0.09263157894736843, "grad_norm": 0.7325497269630432, "learning_rate": 0.001, "loss": 1.7903, "step": 55 }, { "epoch": 0.09431578947368421, "grad_norm": 0.6531684994697571, "learning_rate": 0.001, "loss": 1.8948, "step": 56 }, { "epoch": 0.096, "grad_norm": 0.6358041763305664, "learning_rate": 0.001, "loss": 1.7967, "step": 57 }, { "epoch": 0.09768421052631579, "grad_norm": 0.6489672660827637, "learning_rate": 0.001, "loss": 1.9836, "step": 58 }, { "epoch": 0.09936842105263158, "grad_norm": 0.6798167824745178, "learning_rate": 0.001, "loss": 1.4925, "step": 59 }, { "epoch": 0.10105263157894737, "grad_norm": 0.71540367603302, "learning_rate": 0.001, "loss": 1.9803, "step": 60 }, { "epoch": 0.10273684210526315, "grad_norm": 0.7388565540313721, "learning_rate": 0.001, "loss": 1.994, "step": 61 }, { "epoch": 0.10442105263157894, "grad_norm": 0.5921775698661804, "learning_rate": 0.001, "loss": 1.7347, "step": 62 }, { "epoch": 0.10610526315789473, "grad_norm": 0.6923938393592834, "learning_rate": 0.001, "loss": 1.8954, "step": 63 }, { "epoch": 0.10778947368421053, "grad_norm": 0.6679465770721436, "learning_rate": 0.001, "loss": 1.6889, "step": 64 }, { "epoch": 0.10947368421052632, "grad_norm": 0.6221023797988892, "learning_rate": 0.001, "loss": 1.8324, "step": 65 }, { "epoch": 0.11115789473684211, "grad_norm": 0.6746177673339844, "learning_rate": 0.001, "loss": 1.822, "step": 66 }, { "epoch": 0.1128421052631579, "grad_norm": 0.6581069231033325, "learning_rate": 0.001, "loss": 2.2108, "step": 67 }, { "epoch": 0.11452631578947368, "grad_norm": 0.6342150568962097, "learning_rate": 0.001, "loss": 1.7536, "step": 68 }, { "epoch": 0.11621052631578947, "grad_norm": 0.6652107834815979, "learning_rate": 0.001, "loss": 2.0814, "step": 69 }, { "epoch": 0.11789473684210526, "grad_norm": 0.5661808848381042, "learning_rate": 0.001, "loss": 1.9419, "step": 70 }, { "epoch": 0.11957894736842105, "grad_norm": 0.8789241313934326, "learning_rate": 0.001, "loss": 1.8391, "step": 71 }, { "epoch": 0.12126315789473684, "grad_norm": 0.665008008480072, "learning_rate": 0.001, "loss": 1.9936, "step": 72 }, { "epoch": 0.12294736842105262, "grad_norm": 0.647729218006134, "learning_rate": 0.001, "loss": 1.8883, "step": 73 }, { "epoch": 0.12463157894736843, "grad_norm": 0.5845763087272644, "learning_rate": 0.001, "loss": 1.8779, "step": 74 }, { "epoch": 0.12631578947368421, "grad_norm": 0.6005629897117615, "learning_rate": 0.001, "loss": 2.1395, "step": 75 }, { "epoch": 0.128, "grad_norm": 0.570796012878418, "learning_rate": 0.001, "loss": 1.7652, "step": 76 }, { "epoch": 0.1296842105263158, "grad_norm": 0.652999222278595, "learning_rate": 0.001, "loss": 1.9479, "step": 77 }, { "epoch": 0.13136842105263158, "grad_norm": 0.7086900472640991, "learning_rate": 0.001, "loss": 1.924, "step": 78 }, { "epoch": 0.13305263157894737, "grad_norm": 0.8962117433547974, "learning_rate": 0.001, "loss": 1.7471, "step": 79 }, { "epoch": 0.13473684210526315, "grad_norm": 0.9083784222602844, "learning_rate": 0.001, "loss": 1.4415, "step": 80 }, { "epoch": 0.13642105263157894, "grad_norm": 0.6662907600402832, "learning_rate": 0.001, "loss": 1.5885, "step": 81 }, { "epoch": 0.13810526315789473, "grad_norm": 0.748068630695343, "learning_rate": 0.001, "loss": 1.9043, "step": 82 }, { "epoch": 0.13978947368421052, "grad_norm": 0.653835117816925, "learning_rate": 0.001, "loss": 1.9271, "step": 83 }, { "epoch": 0.1414736842105263, "grad_norm": 0.5937058925628662, "learning_rate": 0.001, "loss": 1.6577, "step": 84 }, { "epoch": 0.1431578947368421, "grad_norm": 0.5573813319206238, "learning_rate": 0.001, "loss": 1.7685, "step": 85 }, { "epoch": 0.14484210526315788, "grad_norm": 0.7234801054000854, "learning_rate": 0.001, "loss": 2.1049, "step": 86 }, { "epoch": 0.14652631578947367, "grad_norm": 0.7858671545982361, "learning_rate": 0.001, "loss": 1.8385, "step": 87 }, { "epoch": 0.1482105263157895, "grad_norm": 0.5881790518760681, "learning_rate": 0.001, "loss": 1.6881, "step": 88 }, { "epoch": 0.14989473684210528, "grad_norm": 0.6036899089813232, "learning_rate": 0.001, "loss": 2.0763, "step": 89 }, { "epoch": 0.15157894736842106, "grad_norm": 0.6678960919380188, "learning_rate": 0.001, "loss": 1.7332, "step": 90 }, { "epoch": 0.15326315789473685, "grad_norm": 0.6993541717529297, "learning_rate": 0.001, "loss": 2.0437, "step": 91 }, { "epoch": 0.15494736842105264, "grad_norm": 1.0041996240615845, "learning_rate": 0.001, "loss": 2.0819, "step": 92 }, { "epoch": 0.15663157894736843, "grad_norm": 1.3755688667297363, "learning_rate": 0.001, "loss": 1.8761, "step": 93 }, { "epoch": 0.15831578947368422, "grad_norm": 0.9399350881576538, "learning_rate": 0.001, "loss": 1.7925, "step": 94 }, { "epoch": 0.16, "grad_norm": 0.6488239169120789, "learning_rate": 0.001, "loss": 2.009, "step": 95 }, { "epoch": 0.1616842105263158, "grad_norm": 0.9083341360092163, "learning_rate": 0.001, "loss": 2.2331, "step": 96 }, { "epoch": 0.16336842105263158, "grad_norm": 0.6239296197891235, "learning_rate": 0.001, "loss": 1.8789, "step": 97 }, { "epoch": 0.16505263157894737, "grad_norm": 0.7653887867927551, "learning_rate": 0.001, "loss": 2.1016, "step": 98 }, { "epoch": 0.16673684210526316, "grad_norm": 0.6791508793830872, "learning_rate": 0.001, "loss": 2.0147, "step": 99 }, { "epoch": 0.16842105263157894, "grad_norm": 0.6757349967956543, "learning_rate": 0.001, "loss": 1.909, "step": 100 }, { "epoch": 0.17010526315789473, "grad_norm": 0.5010210275650024, "learning_rate": 0.001, "loss": 2.0148, "step": 101 }, { "epoch": 0.17178947368421052, "grad_norm": 0.6564686894416809, "learning_rate": 0.001, "loss": 1.707, "step": 102 }, { "epoch": 0.1734736842105263, "grad_norm": 0.6926625370979309, "learning_rate": 0.001, "loss": 2.1313, "step": 103 }, { "epoch": 0.1751578947368421, "grad_norm": 0.8134363293647766, "learning_rate": 0.001, "loss": 1.8948, "step": 104 }, { "epoch": 0.17684210526315788, "grad_norm": 0.8722719550132751, "learning_rate": 0.001, "loss": 1.9564, "step": 105 }, { "epoch": 0.17852631578947367, "grad_norm": 1.5459606647491455, "learning_rate": 0.001, "loss": 1.4568, "step": 106 }, { "epoch": 0.18021052631578946, "grad_norm": 0.672356367111206, "learning_rate": 0.001, "loss": 2.2486, "step": 107 }, { "epoch": 0.18189473684210528, "grad_norm": 0.6597303152084351, "learning_rate": 0.001, "loss": 2.1888, "step": 108 }, { "epoch": 0.18357894736842106, "grad_norm": 0.6516699194908142, "learning_rate": 0.001, "loss": 1.7791, "step": 109 }, { "epoch": 0.18526315789473685, "grad_norm": 0.6535261273384094, "learning_rate": 0.001, "loss": 1.5753, "step": 110 }, { "epoch": 0.18694736842105264, "grad_norm": 0.5394155979156494, "learning_rate": 0.001, "loss": 1.7874, "step": 111 }, { "epoch": 0.18863157894736843, "grad_norm": 0.6403316855430603, "learning_rate": 0.001, "loss": 1.8044, "step": 112 }, { "epoch": 0.19031578947368422, "grad_norm": 0.6894748210906982, "learning_rate": 0.001, "loss": 1.7434, "step": 113 }, { "epoch": 0.192, "grad_norm": 0.5362414717674255, "learning_rate": 0.001, "loss": 1.6198, "step": 114 }, { "epoch": 0.1936842105263158, "grad_norm": 0.5218887329101562, "learning_rate": 0.001, "loss": 1.7941, "step": 115 }, { "epoch": 0.19536842105263158, "grad_norm": 0.5951269865036011, "learning_rate": 0.001, "loss": 2.115, "step": 116 }, { "epoch": 0.19705263157894737, "grad_norm": 1.0313245058059692, "learning_rate": 0.001, "loss": 1.9144, "step": 117 }, { "epoch": 0.19873684210526316, "grad_norm": 0.6776890754699707, "learning_rate": 0.001, "loss": 1.6852, "step": 118 }, { "epoch": 0.20042105263157894, "grad_norm": 0.5906718373298645, "learning_rate": 0.001, "loss": 1.7103, "step": 119 }, { "epoch": 0.20210526315789473, "grad_norm": 0.6788285970687866, "learning_rate": 0.001, "loss": 2.1048, "step": 120 }, { "epoch": 0.20378947368421052, "grad_norm": 0.7527502179145813, "learning_rate": 0.001, "loss": 1.8199, "step": 121 }, { "epoch": 0.2054736842105263, "grad_norm": 0.5279136896133423, "learning_rate": 0.001, "loss": 1.5327, "step": 122 }, { "epoch": 0.2071578947368421, "grad_norm": 0.7087485194206238, "learning_rate": 0.001, "loss": 1.8483, "step": 123 }, { "epoch": 0.20884210526315788, "grad_norm": 0.7274911999702454, "learning_rate": 0.001, "loss": 2.4062, "step": 124 }, { "epoch": 0.21052631578947367, "grad_norm": 0.5436732769012451, "learning_rate": 0.001, "loss": 1.7369, "step": 125 }, { "epoch": 0.21221052631578946, "grad_norm": 0.5522803664207458, "learning_rate": 0.001, "loss": 1.8024, "step": 126 }, { "epoch": 0.21389473684210528, "grad_norm": 0.7198563814163208, "learning_rate": 0.001, "loss": 2.2969, "step": 127 }, { "epoch": 0.21557894736842106, "grad_norm": 0.6230013370513916, "learning_rate": 0.001, "loss": 1.8566, "step": 128 }, { "epoch": 0.21726315789473685, "grad_norm": 0.5977436304092407, "learning_rate": 0.001, "loss": 1.339, "step": 129 }, { "epoch": 0.21894736842105264, "grad_norm": 0.5400142669677734, "learning_rate": 0.001, "loss": 1.6346, "step": 130 }, { "epoch": 0.22063157894736843, "grad_norm": 0.6537740230560303, "learning_rate": 0.001, "loss": 1.981, "step": 131 }, { "epoch": 0.22231578947368422, "grad_norm": 0.6012418866157532, "learning_rate": 0.001, "loss": 1.9957, "step": 132 }, { "epoch": 0.224, "grad_norm": 0.6363667845726013, "learning_rate": 0.001, "loss": 2.0906, "step": 133 }, { "epoch": 0.2256842105263158, "grad_norm": 0.7009410262107849, "learning_rate": 0.001, "loss": 2.0259, "step": 134 }, { "epoch": 0.22736842105263158, "grad_norm": 0.6076754331588745, "learning_rate": 0.001, "loss": 2.0623, "step": 135 }, { "epoch": 0.22905263157894737, "grad_norm": 0.6430286169052124, "learning_rate": 0.001, "loss": 1.9965, "step": 136 }, { "epoch": 0.23073684210526316, "grad_norm": 0.7743528485298157, "learning_rate": 0.001, "loss": 1.6476, "step": 137 }, { "epoch": 0.23242105263157894, "grad_norm": 0.7490441203117371, "learning_rate": 0.001, "loss": 2.1016, "step": 138 }, { "epoch": 0.23410526315789473, "grad_norm": 0.6852337121963501, "learning_rate": 0.001, "loss": 1.7251, "step": 139 }, { "epoch": 0.23578947368421052, "grad_norm": 0.6012661457061768, "learning_rate": 0.001, "loss": 1.985, "step": 140 }, { "epoch": 0.2374736842105263, "grad_norm": 0.48385104537010193, "learning_rate": 0.001, "loss": 1.7926, "step": 141 }, { "epoch": 0.2391578947368421, "grad_norm": 0.5751200914382935, "learning_rate": 0.001, "loss": 1.8795, "step": 142 }, { "epoch": 0.24084210526315789, "grad_norm": 0.571426510810852, "learning_rate": 0.001, "loss": 1.6457, "step": 143 }, { "epoch": 0.24252631578947367, "grad_norm": 0.6982892751693726, "learning_rate": 0.001, "loss": 2.0313, "step": 144 }, { "epoch": 0.24421052631578946, "grad_norm": 0.7380142211914062, "learning_rate": 0.001, "loss": 2.1306, "step": 145 }, { "epoch": 0.24589473684210525, "grad_norm": 0.97590571641922, "learning_rate": 0.001, "loss": 1.9926, "step": 146 }, { "epoch": 0.24757894736842107, "grad_norm": 0.8416200876235962, "learning_rate": 0.001, "loss": 1.733, "step": 147 }, { "epoch": 0.24926315789473685, "grad_norm": 0.6639004945755005, "learning_rate": 0.001, "loss": 1.9836, "step": 148 }, { "epoch": 0.25094736842105264, "grad_norm": 0.6488214135169983, "learning_rate": 0.001, "loss": 1.8941, "step": 149 }, { "epoch": 0.25094736842105264, "eval_loss": 1.7599804401397705, "eval_runtime": 0.1738, "eval_samples_per_second": 5.754, "eval_steps_per_second": 5.754, "step": 149 }, { "epoch": 0.25263157894736843, "grad_norm": 0.524825930595398, "learning_rate": 0.001, "loss": 1.7767, "step": 150 }, { "epoch": 0.2543157894736842, "grad_norm": 0.49492335319519043, "learning_rate": 0.001, "loss": 1.775, "step": 151 }, { "epoch": 0.256, "grad_norm": 0.5911272168159485, "learning_rate": 0.001, "loss": 1.8286, "step": 152 }, { "epoch": 0.2576842105263158, "grad_norm": 0.8157614469528198, "learning_rate": 0.001, "loss": 1.8913, "step": 153 }, { "epoch": 0.2593684210526316, "grad_norm": 0.7529662847518921, "learning_rate": 0.001, "loss": 1.8988, "step": 154 }, { "epoch": 0.26105263157894737, "grad_norm": 0.8185762763023376, "learning_rate": 0.001, "loss": 2.1685, "step": 155 }, { "epoch": 0.26273684210526316, "grad_norm": 0.7138445377349854, "learning_rate": 0.001, "loss": 1.8507, "step": 156 }, { "epoch": 0.26442105263157895, "grad_norm": 0.5665900707244873, "learning_rate": 0.001, "loss": 1.525, "step": 157 }, { "epoch": 0.26610526315789473, "grad_norm": 0.6799633502960205, "learning_rate": 0.001, "loss": 1.9605, "step": 158 }, { "epoch": 0.2677894736842105, "grad_norm": 0.6787411570549011, "learning_rate": 0.001, "loss": 1.6556, "step": 159 }, { "epoch": 0.2694736842105263, "grad_norm": 0.8427496552467346, "learning_rate": 0.001, "loss": 1.8083, "step": 160 }, { "epoch": 0.2711578947368421, "grad_norm": 0.6665315628051758, "learning_rate": 0.001, "loss": 1.966, "step": 161 }, { "epoch": 0.2728421052631579, "grad_norm": 0.6209701895713806, "learning_rate": 0.001, "loss": 1.79, "step": 162 }, { "epoch": 0.2745263157894737, "grad_norm": 0.5687562823295593, "learning_rate": 0.001, "loss": 1.779, "step": 163 }, { "epoch": 0.27621052631578946, "grad_norm": 0.5852699279785156, "learning_rate": 0.001, "loss": 1.4817, "step": 164 }, { "epoch": 0.27789473684210525, "grad_norm": 0.6601601839065552, "learning_rate": 0.001, "loss": 1.9279, "step": 165 }, { "epoch": 0.27957894736842104, "grad_norm": 0.5629734992980957, "learning_rate": 0.001, "loss": 1.821, "step": 166 }, { "epoch": 0.2812631578947368, "grad_norm": 0.7956101894378662, "learning_rate": 0.001, "loss": 1.9925, "step": 167 }, { "epoch": 0.2829473684210526, "grad_norm": 0.7143905758857727, "learning_rate": 0.001, "loss": 2.0572, "step": 168 }, { "epoch": 0.2846315789473684, "grad_norm": 0.7645180821418762, "learning_rate": 0.001, "loss": 1.9208, "step": 169 }, { "epoch": 0.2863157894736842, "grad_norm": 0.7295411825180054, "learning_rate": 0.001, "loss": 1.962, "step": 170 }, { "epoch": 0.288, "grad_norm": 0.7587769031524658, "learning_rate": 0.001, "loss": 1.9372, "step": 171 }, { "epoch": 0.28968421052631577, "grad_norm": 0.6111007332801819, "learning_rate": 0.001, "loss": 1.5688, "step": 172 }, { "epoch": 0.29136842105263155, "grad_norm": 0.7311589121818542, "learning_rate": 0.001, "loss": 1.8835, "step": 173 }, { "epoch": 0.29305263157894734, "grad_norm": 0.6812251210212708, "learning_rate": 0.001, "loss": 1.8754, "step": 174 }, { "epoch": 0.29473684210526313, "grad_norm": 0.6704198718070984, "learning_rate": 0.001, "loss": 1.8445, "step": 175 }, { "epoch": 0.296421052631579, "grad_norm": 0.7953410148620605, "learning_rate": 0.001, "loss": 2.0227, "step": 176 }, { "epoch": 0.29810526315789476, "grad_norm": 0.8933955430984497, "learning_rate": 0.001, "loss": 1.9599, "step": 177 }, { "epoch": 0.29978947368421055, "grad_norm": 0.7686247825622559, "learning_rate": 0.001, "loss": 1.8684, "step": 178 }, { "epoch": 0.30147368421052634, "grad_norm": 0.6903300881385803, "learning_rate": 0.001, "loss": 1.9449, "step": 179 }, { "epoch": 0.3031578947368421, "grad_norm": 0.6970056295394897, "learning_rate": 0.001, "loss": 1.6342, "step": 180 }, { "epoch": 0.3048421052631579, "grad_norm": 0.5772702693939209, "learning_rate": 0.001, "loss": 1.7464, "step": 181 }, { "epoch": 0.3065263157894737, "grad_norm": 0.6376874446868896, "learning_rate": 0.001, "loss": 1.82, "step": 182 }, { "epoch": 0.3082105263157895, "grad_norm": 0.761457085609436, "learning_rate": 0.001, "loss": 1.7796, "step": 183 }, { "epoch": 0.3098947368421053, "grad_norm": 0.6312285661697388, "learning_rate": 0.001, "loss": 2.0363, "step": 184 }, { "epoch": 0.31157894736842107, "grad_norm": 0.8040784001350403, "learning_rate": 0.001, "loss": 2.0369, "step": 185 }, { "epoch": 0.31326315789473685, "grad_norm": 0.7520210146903992, "learning_rate": 0.001, "loss": 2.1518, "step": 186 }, { "epoch": 0.31494736842105264, "grad_norm": 0.6293883919715881, "learning_rate": 0.001, "loss": 1.9187, "step": 187 }, { "epoch": 0.31663157894736843, "grad_norm": 0.7219449877738953, "learning_rate": 0.001, "loss": 1.5069, "step": 188 }, { "epoch": 0.3183157894736842, "grad_norm": 0.8080244660377502, "learning_rate": 0.001, "loss": 1.6169, "step": 189 }, { "epoch": 0.32, "grad_norm": 0.8044946193695068, "learning_rate": 0.001, "loss": 1.8206, "step": 190 }, { "epoch": 0.3216842105263158, "grad_norm": 0.894588828086853, "learning_rate": 0.001, "loss": 2.0623, "step": 191 }, { "epoch": 0.3233684210526316, "grad_norm": 0.6865862607955933, "learning_rate": 0.001, "loss": 1.7629, "step": 192 }, { "epoch": 0.32505263157894737, "grad_norm": 0.9992401003837585, "learning_rate": 0.001, "loss": 2.2471, "step": 193 }, { "epoch": 0.32673684210526316, "grad_norm": 0.8792619705200195, "learning_rate": 0.001, "loss": 2.0458, "step": 194 }, { "epoch": 0.32842105263157895, "grad_norm": 0.8508814573287964, "learning_rate": 0.001, "loss": 1.8746, "step": 195 }, { "epoch": 0.33010526315789473, "grad_norm": 0.6977102756500244, "learning_rate": 0.001, "loss": 2.0411, "step": 196 }, { "epoch": 0.3317894736842105, "grad_norm": 0.8430894017219543, "learning_rate": 0.001, "loss": 2.0066, "step": 197 }, { "epoch": 0.3334736842105263, "grad_norm": 0.8048614859580994, "learning_rate": 0.001, "loss": 1.9879, "step": 198 }, { "epoch": 0.3351578947368421, "grad_norm": 0.8604184985160828, "learning_rate": 0.001, "loss": 2.0952, "step": 199 }, { "epoch": 0.3368421052631579, "grad_norm": 1.0472347736358643, "learning_rate": 0.001, "loss": 1.8939, "step": 200 }, { "epoch": 0.3385263157894737, "grad_norm": 0.761587381362915, "learning_rate": 0.001, "loss": 1.9897, "step": 201 }, { "epoch": 0.34021052631578946, "grad_norm": 0.7086905837059021, "learning_rate": 0.001, "loss": 1.8068, "step": 202 }, { "epoch": 0.34189473684210525, "grad_norm": 0.8213825225830078, "learning_rate": 0.001, "loss": 1.682, "step": 203 }, { "epoch": 0.34357894736842104, "grad_norm": 0.8660598397254944, "learning_rate": 0.001, "loss": 2.0042, "step": 204 }, { "epoch": 0.3452631578947368, "grad_norm": 0.9913591742515564, "learning_rate": 0.001, "loss": 1.9211, "step": 205 }, { "epoch": 0.3469473684210526, "grad_norm": 0.7640036940574646, "learning_rate": 0.001, "loss": 2.0904, "step": 206 }, { "epoch": 0.3486315789473684, "grad_norm": 0.7359378337860107, "learning_rate": 0.001, "loss": 1.9424, "step": 207 }, { "epoch": 0.3503157894736842, "grad_norm": 0.7246221303939819, "learning_rate": 0.001, "loss": 1.9887, "step": 208 }, { "epoch": 0.352, "grad_norm": 0.8169429302215576, "learning_rate": 0.001, "loss": 1.8335, "step": 209 }, { "epoch": 0.35368421052631577, "grad_norm": 0.903678834438324, "learning_rate": 0.001, "loss": 1.6703, "step": 210 }, { "epoch": 0.35536842105263156, "grad_norm": 0.7328379154205322, "learning_rate": 0.001, "loss": 2.1565, "step": 211 }, { "epoch": 0.35705263157894734, "grad_norm": 0.8003093004226685, "learning_rate": 0.001, "loss": 1.7172, "step": 212 }, { "epoch": 0.35873684210526313, "grad_norm": 0.7532063722610474, "learning_rate": 0.001, "loss": 2.0264, "step": 213 }, { "epoch": 0.3604210526315789, "grad_norm": 0.7619852423667908, "learning_rate": 0.001, "loss": 1.7766, "step": 214 }, { "epoch": 0.36210526315789476, "grad_norm": 0.7145585417747498, "learning_rate": 0.001, "loss": 1.8829, "step": 215 }, { "epoch": 0.36378947368421055, "grad_norm": 0.739275336265564, "learning_rate": 0.001, "loss": 1.6963, "step": 216 }, { "epoch": 0.36547368421052634, "grad_norm": 0.8174360990524292, "learning_rate": 0.001, "loss": 2.028, "step": 217 }, { "epoch": 0.3671578947368421, "grad_norm": 0.7873148322105408, "learning_rate": 0.001, "loss": 1.9912, "step": 218 }, { "epoch": 0.3688421052631579, "grad_norm": 0.7683485746383667, "learning_rate": 0.001, "loss": 2.0106, "step": 219 }, { "epoch": 0.3705263157894737, "grad_norm": 0.841464102268219, "learning_rate": 0.001, "loss": 1.3909, "step": 220 }, { "epoch": 0.3722105263157895, "grad_norm": 0.9224113821983337, "learning_rate": 0.001, "loss": 2.0642, "step": 221 }, { "epoch": 0.3738947368421053, "grad_norm": 1.3310387134552002, "learning_rate": 0.001, "loss": 2.0075, "step": 222 }, { "epoch": 0.37557894736842107, "grad_norm": 0.8936915397644043, "learning_rate": 0.001, "loss": 1.8838, "step": 223 }, { "epoch": 0.37726315789473686, "grad_norm": 0.7084046602249146, "learning_rate": 0.001, "loss": 2.0105, "step": 224 }, { "epoch": 0.37894736842105264, "grad_norm": 0.802139401435852, "learning_rate": 0.001, "loss": 2.1188, "step": 225 }, { "epoch": 0.38063157894736843, "grad_norm": 0.8018360137939453, "learning_rate": 0.001, "loss": 1.8581, "step": 226 }, { "epoch": 0.3823157894736842, "grad_norm": 0.8070486187934875, "learning_rate": 0.001, "loss": 1.9121, "step": 227 }, { "epoch": 0.384, "grad_norm": 0.7557722330093384, "learning_rate": 0.001, "loss": 1.8515, "step": 228 }, { "epoch": 0.3856842105263158, "grad_norm": 0.8111100196838379, "learning_rate": 0.001, "loss": 2.0789, "step": 229 }, { "epoch": 0.3873684210526316, "grad_norm": 0.9642356038093567, "learning_rate": 0.001, "loss": 2.2856, "step": 230 }, { "epoch": 0.38905263157894737, "grad_norm": 0.9470245242118835, "learning_rate": 0.001, "loss": 2.0114, "step": 231 }, { "epoch": 0.39073684210526316, "grad_norm": 0.8576509952545166, "learning_rate": 0.001, "loss": 1.8915, "step": 232 }, { "epoch": 0.39242105263157895, "grad_norm": 0.8524518013000488, "learning_rate": 0.001, "loss": 1.9937, "step": 233 }, { "epoch": 0.39410526315789474, "grad_norm": 1.523067831993103, "learning_rate": 0.001, "loss": 1.641, "step": 234 }, { "epoch": 0.3957894736842105, "grad_norm": 0.9369080066680908, "learning_rate": 0.001, "loss": 2.0977, "step": 235 }, { "epoch": 0.3974736842105263, "grad_norm": 0.8704274296760559, "learning_rate": 0.001, "loss": 2.0126, "step": 236 }, { "epoch": 0.3991578947368421, "grad_norm": 0.8420674800872803, "learning_rate": 0.001, "loss": 1.9577, "step": 237 }, { "epoch": 0.4008421052631579, "grad_norm": 0.7344264388084412, "learning_rate": 0.001, "loss": 1.8681, "step": 238 }, { "epoch": 0.4025263157894737, "grad_norm": 0.7144782543182373, "learning_rate": 0.001, "loss": 1.9696, "step": 239 }, { "epoch": 0.40421052631578946, "grad_norm": 0.8455988168716431, "learning_rate": 0.001, "loss": 1.7568, "step": 240 }, { "epoch": 0.40589473684210525, "grad_norm": 0.807806134223938, "learning_rate": 0.001, "loss": 2.1119, "step": 241 }, { "epoch": 0.40757894736842104, "grad_norm": 0.8274264335632324, "learning_rate": 0.001, "loss": 1.9695, "step": 242 }, { "epoch": 0.40926315789473683, "grad_norm": 0.9100606441497803, "learning_rate": 0.001, "loss": 1.9766, "step": 243 }, { "epoch": 0.4109473684210526, "grad_norm": 1.1465590000152588, "learning_rate": 0.001, "loss": 1.8377, "step": 244 }, { "epoch": 0.4126315789473684, "grad_norm": 0.7355701327323914, "learning_rate": 0.001, "loss": 1.3323, "step": 245 }, { "epoch": 0.4143157894736842, "grad_norm": 0.8275692462921143, "learning_rate": 0.001, "loss": 1.9701, "step": 246 }, { "epoch": 0.416, "grad_norm": 0.848210334777832, "learning_rate": 0.001, "loss": 2.0912, "step": 247 }, { "epoch": 0.41768421052631577, "grad_norm": 0.8262030482292175, "learning_rate": 0.001, "loss": 2.115, "step": 248 }, { "epoch": 0.41936842105263156, "grad_norm": 0.6998792886734009, "learning_rate": 0.001, "loss": 1.8737, "step": 249 }, { "epoch": 0.42105263157894735, "grad_norm": 0.8081846237182617, "learning_rate": 0.001, "loss": 2.2837, "step": 250 }, { "epoch": 0.42273684210526313, "grad_norm": 0.8310023546218872, "learning_rate": 0.001, "loss": 2.2355, "step": 251 }, { "epoch": 0.4244210526315789, "grad_norm": 1.157334804534912, "learning_rate": 0.001, "loss": 2.3762, "step": 252 }, { "epoch": 0.4261052631578947, "grad_norm": 0.7893511056900024, "learning_rate": 0.001, "loss": 1.8737, "step": 253 }, { "epoch": 0.42778947368421055, "grad_norm": 0.959355354309082, "learning_rate": 0.001, "loss": 2.0047, "step": 254 }, { "epoch": 0.42947368421052634, "grad_norm": 0.8904256224632263, "learning_rate": 0.001, "loss": 1.8809, "step": 255 }, { "epoch": 0.43115789473684213, "grad_norm": 0.9054950475692749, "learning_rate": 0.001, "loss": 2.3487, "step": 256 }, { "epoch": 0.4328421052631579, "grad_norm": 0.8389487862586975, "learning_rate": 0.001, "loss": 2.0993, "step": 257 }, { "epoch": 0.4345263157894737, "grad_norm": 1.0516859292984009, "learning_rate": 0.001, "loss": 1.9482, "step": 258 }, { "epoch": 0.4362105263157895, "grad_norm": 1.0936013460159302, "learning_rate": 0.001, "loss": 2.2888, "step": 259 }, { "epoch": 0.4378947368421053, "grad_norm": 0.894350528717041, "learning_rate": 0.001, "loss": 1.9703, "step": 260 }, { "epoch": 0.43957894736842107, "grad_norm": 0.8147197961807251, "learning_rate": 0.001, "loss": 1.8964, "step": 261 }, { "epoch": 0.44126315789473686, "grad_norm": 0.6683039665222168, "learning_rate": 0.001, "loss": 1.8701, "step": 262 }, { "epoch": 0.44294736842105265, "grad_norm": 0.83613121509552, "learning_rate": 0.001, "loss": 2.3627, "step": 263 }, { "epoch": 0.44463157894736843, "grad_norm": 0.724908173084259, "learning_rate": 0.001, "loss": 1.8411, "step": 264 }, { "epoch": 0.4463157894736842, "grad_norm": 0.7576204538345337, "learning_rate": 0.001, "loss": 2.1275, "step": 265 }, { "epoch": 0.448, "grad_norm": 0.7902230620384216, "learning_rate": 0.001, "loss": 1.663, "step": 266 }, { "epoch": 0.4496842105263158, "grad_norm": 1.5043684244155884, "learning_rate": 0.001, "loss": 2.1615, "step": 267 }, { "epoch": 0.4513684210526316, "grad_norm": 0.8250028491020203, "learning_rate": 0.001, "loss": 1.8787, "step": 268 }, { "epoch": 0.4530526315789474, "grad_norm": 0.7849893569946289, "learning_rate": 0.001, "loss": 2.0168, "step": 269 }, { "epoch": 0.45473684210526316, "grad_norm": 0.8021153807640076, "learning_rate": 0.001, "loss": 2.2486, "step": 270 }, { "epoch": 0.45642105263157895, "grad_norm": 0.7869856953620911, "learning_rate": 0.001, "loss": 1.5522, "step": 271 }, { "epoch": 0.45810526315789474, "grad_norm": 0.813165009021759, "learning_rate": 0.001, "loss": 2.0801, "step": 272 }, { "epoch": 0.4597894736842105, "grad_norm": 0.8223312497138977, "learning_rate": 0.001, "loss": 1.9111, "step": 273 }, { "epoch": 0.4614736842105263, "grad_norm": 0.8650989532470703, "learning_rate": 0.001, "loss": 1.716, "step": 274 }, { "epoch": 0.4631578947368421, "grad_norm": 0.7596947550773621, "learning_rate": 0.001, "loss": 1.9046, "step": 275 }, { "epoch": 0.4648421052631579, "grad_norm": 0.7211440801620483, "learning_rate": 0.001, "loss": 1.7714, "step": 276 }, { "epoch": 0.4665263157894737, "grad_norm": 1.0673142671585083, "learning_rate": 0.001, "loss": 2.1321, "step": 277 }, { "epoch": 0.46821052631578947, "grad_norm": 0.7947107553482056, "learning_rate": 0.001, "loss": 1.9446, "step": 278 }, { "epoch": 0.46989473684210525, "grad_norm": 0.8121020197868347, "learning_rate": 0.001, "loss": 2.217, "step": 279 }, { "epoch": 0.47157894736842104, "grad_norm": 0.7495191097259521, "learning_rate": 0.001, "loss": 1.891, "step": 280 }, { "epoch": 0.47326315789473683, "grad_norm": 0.7859931588172913, "learning_rate": 0.001, "loss": 1.5517, "step": 281 }, { "epoch": 0.4749473684210526, "grad_norm": 0.8961056470870972, "learning_rate": 0.001, "loss": 2.0886, "step": 282 }, { "epoch": 0.4766315789473684, "grad_norm": 0.7971674799919128, "learning_rate": 0.001, "loss": 1.8596, "step": 283 }, { "epoch": 0.4783157894736842, "grad_norm": 0.881367564201355, "learning_rate": 0.001, "loss": 1.9999, "step": 284 }, { "epoch": 0.48, "grad_norm": 0.883185863494873, "learning_rate": 0.001, "loss": 2.0156, "step": 285 }, { "epoch": 0.48168421052631577, "grad_norm": 0.8560335636138916, "learning_rate": 0.001, "loss": 2.0237, "step": 286 }, { "epoch": 0.48336842105263156, "grad_norm": 1.038077712059021, "learning_rate": 0.001, "loss": 1.6664, "step": 287 }, { "epoch": 0.48505263157894735, "grad_norm": 0.7434845566749573, "learning_rate": 0.001, "loss": 1.9518, "step": 288 }, { "epoch": 0.48673684210526313, "grad_norm": 1.099915862083435, "learning_rate": 0.001, "loss": 2.1582, "step": 289 }, { "epoch": 0.4884210526315789, "grad_norm": 0.7814631462097168, "learning_rate": 0.001, "loss": 1.7189, "step": 290 }, { "epoch": 0.4901052631578947, "grad_norm": 0.9618262052536011, "learning_rate": 0.001, "loss": 2.1185, "step": 291 }, { "epoch": 0.4917894736842105, "grad_norm": 0.7547399401664734, "learning_rate": 0.001, "loss": 1.8551, "step": 292 }, { "epoch": 0.49347368421052634, "grad_norm": 0.891696035861969, "learning_rate": 0.001, "loss": 2.1148, "step": 293 }, { "epoch": 0.49515789473684213, "grad_norm": 0.9156106114387512, "learning_rate": 0.001, "loss": 1.829, "step": 294 }, { "epoch": 0.4968421052631579, "grad_norm": 0.8770383596420288, "learning_rate": 0.001, "loss": 1.9336, "step": 295 }, { "epoch": 0.4985263157894737, "grad_norm": 0.7889037132263184, "learning_rate": 0.001, "loss": 1.7122, "step": 296 }, { "epoch": 0.5002105263157894, "grad_norm": 1.0132378339767456, "learning_rate": 0.001, "loss": 2.0484, "step": 297 }, { "epoch": 0.5018947368421053, "grad_norm": 0.852583110332489, "learning_rate": 0.001, "loss": 1.9974, "step": 298 }, { "epoch": 0.5018947368421053, "eval_loss": 1.6337618827819824, "eval_runtime": 0.0834, "eval_samples_per_second": 11.984, "eval_steps_per_second": 11.984, "step": 298 }, { "epoch": 0.503578947368421, "grad_norm": 1.0435866117477417, "learning_rate": 0.001, "loss": 2.105, "step": 299 }, { "epoch": 0.5052631578947369, "grad_norm": 0.754615843296051, "learning_rate": 0.001, "loss": 2.0089, "step": 300 }, { "epoch": 0.5069473684210526, "grad_norm": 1.012373924255371, "learning_rate": 0.001, "loss": 2.3203, "step": 301 }, { "epoch": 0.5086315789473684, "grad_norm": 0.7808589935302734, "learning_rate": 0.001, "loss": 2.1087, "step": 302 }, { "epoch": 0.5103157894736842, "grad_norm": 0.8035853505134583, "learning_rate": 0.001, "loss": 2.1473, "step": 303 }, { "epoch": 0.512, "grad_norm": 0.7854329943656921, "learning_rate": 0.001, "loss": 1.9042, "step": 304 }, { "epoch": 0.5136842105263157, "grad_norm": 0.8837404251098633, "learning_rate": 0.001, "loss": 1.6176, "step": 305 }, { "epoch": 0.5153684210526316, "grad_norm": 0.9439155459403992, "learning_rate": 0.001, "loss": 2.1619, "step": 306 }, { "epoch": 0.5170526315789473, "grad_norm": 0.836586058139801, "learning_rate": 0.001, "loss": 1.9484, "step": 307 }, { "epoch": 0.5187368421052632, "grad_norm": 0.8734055161476135, "learning_rate": 0.001, "loss": 2.0695, "step": 308 }, { "epoch": 0.5204210526315789, "grad_norm": 0.8716776967048645, "learning_rate": 0.001, "loss": 2.1273, "step": 309 }, { "epoch": 0.5221052631578947, "grad_norm": 0.9540092349052429, "learning_rate": 0.001, "loss": 1.8999, "step": 310 }, { "epoch": 0.5237894736842105, "grad_norm": 1.1694831848144531, "learning_rate": 0.001, "loss": 1.728, "step": 311 }, { "epoch": 0.5254736842105263, "grad_norm": 0.7269738912582397, "learning_rate": 0.001, "loss": 2.0951, "step": 312 }, { "epoch": 0.5271578947368422, "grad_norm": 0.7646914720535278, "learning_rate": 0.001, "loss": 1.8231, "step": 313 }, { "epoch": 0.5288421052631579, "grad_norm": 0.8613254427909851, "learning_rate": 0.001, "loss": 1.9105, "step": 314 }, { "epoch": 0.5305263157894737, "grad_norm": 1.0191853046417236, "learning_rate": 0.001, "loss": 1.7064, "step": 315 }, { "epoch": 0.5322105263157895, "grad_norm": 1.2197155952453613, "learning_rate": 0.001, "loss": 1.7123, "step": 316 }, { "epoch": 0.5338947368421053, "grad_norm": 0.818133533000946, "learning_rate": 0.001, "loss": 1.7865, "step": 317 }, { "epoch": 0.535578947368421, "grad_norm": 0.8760883808135986, "learning_rate": 0.001, "loss": 2.1209, "step": 318 }, { "epoch": 0.5372631578947369, "grad_norm": 1.0778782367706299, "learning_rate": 0.001, "loss": 2.0902, "step": 319 }, { "epoch": 0.5389473684210526, "grad_norm": 0.8181326985359192, "learning_rate": 0.001, "loss": 1.9372, "step": 320 }, { "epoch": 0.5406315789473685, "grad_norm": 0.9272657036781311, "learning_rate": 0.001, "loss": 2.0478, "step": 321 }, { "epoch": 0.5423157894736842, "grad_norm": 0.9218736290931702, "learning_rate": 0.001, "loss": 2.108, "step": 322 }, { "epoch": 0.544, "grad_norm": 1.0985972881317139, "learning_rate": 0.001, "loss": 2.0267, "step": 323 }, { "epoch": 0.5456842105263158, "grad_norm": 0.8310480713844299, "learning_rate": 0.001, "loss": 1.7772, "step": 324 }, { "epoch": 0.5473684210526316, "grad_norm": 0.8774259090423584, "learning_rate": 0.001, "loss": 2.0248, "step": 325 }, { "epoch": 0.5490526315789473, "grad_norm": 1.0681616067886353, "learning_rate": 0.001, "loss": 2.2355, "step": 326 }, { "epoch": 0.5507368421052632, "grad_norm": 0.9428539872169495, "learning_rate": 0.001, "loss": 1.9988, "step": 327 }, { "epoch": 0.5524210526315789, "grad_norm": 1.0054833889007568, "learning_rate": 0.001, "loss": 1.9063, "step": 328 }, { "epoch": 0.5541052631578948, "grad_norm": 0.8005337715148926, "learning_rate": 0.001, "loss": 2.1752, "step": 329 }, { "epoch": 0.5557894736842105, "grad_norm": 1.109134554862976, "learning_rate": 0.001, "loss": 2.3235, "step": 330 }, { "epoch": 0.5574736842105263, "grad_norm": 0.9584336280822754, "learning_rate": 0.001, "loss": 1.7009, "step": 331 }, { "epoch": 0.5591578947368421, "grad_norm": 1.2622302770614624, "learning_rate": 0.001, "loss": 2.0998, "step": 332 }, { "epoch": 0.5608421052631579, "grad_norm": 0.92564457654953, "learning_rate": 0.001, "loss": 1.7039, "step": 333 }, { "epoch": 0.5625263157894737, "grad_norm": 0.7569521069526672, "learning_rate": 0.001, "loss": 1.7963, "step": 334 }, { "epoch": 0.5642105263157895, "grad_norm": 0.7915797233581543, "learning_rate": 0.001, "loss": 1.645, "step": 335 }, { "epoch": 0.5658947368421052, "grad_norm": 0.7300320863723755, "learning_rate": 0.001, "loss": 1.7476, "step": 336 }, { "epoch": 0.5675789473684211, "grad_norm": 1.1384440660476685, "learning_rate": 0.001, "loss": 2.1937, "step": 337 }, { "epoch": 0.5692631578947368, "grad_norm": 0.8770859241485596, "learning_rate": 0.001, "loss": 1.9664, "step": 338 }, { "epoch": 0.5709473684210526, "grad_norm": 0.9081368446350098, "learning_rate": 0.001, "loss": 1.4632, "step": 339 }, { "epoch": 0.5726315789473684, "grad_norm": 0.8865834474563599, "learning_rate": 0.001, "loss": 1.7578, "step": 340 }, { "epoch": 0.5743157894736842, "grad_norm": 0.8756502866744995, "learning_rate": 0.001, "loss": 1.9929, "step": 341 }, { "epoch": 0.576, "grad_norm": 0.8333286643028259, "learning_rate": 0.001, "loss": 2.3068, "step": 342 }, { "epoch": 0.5776842105263158, "grad_norm": 0.8217945098876953, "learning_rate": 0.001, "loss": 1.9818, "step": 343 }, { "epoch": 0.5793684210526315, "grad_norm": 0.8414101004600525, "learning_rate": 0.001, "loss": 2.0111, "step": 344 }, { "epoch": 0.5810526315789474, "grad_norm": 0.9645239114761353, "learning_rate": 0.001, "loss": 2.2026, "step": 345 }, { "epoch": 0.5827368421052631, "grad_norm": 0.9366424083709717, "learning_rate": 0.001, "loss": 2.1754, "step": 346 }, { "epoch": 0.584421052631579, "grad_norm": 0.839468240737915, "learning_rate": 0.001, "loss": 1.8966, "step": 347 }, { "epoch": 0.5861052631578947, "grad_norm": 0.9215678572654724, "learning_rate": 0.001, "loss": 2.1501, "step": 348 }, { "epoch": 0.5877894736842105, "grad_norm": 1.0060967206954956, "learning_rate": 0.001, "loss": 1.9594, "step": 349 }, { "epoch": 0.5894736842105263, "grad_norm": 0.9866886734962463, "learning_rate": 0.001, "loss": 2.0878, "step": 350 }, { "epoch": 0.5911578947368421, "grad_norm": 1.0554858446121216, "learning_rate": 0.001, "loss": 1.7109, "step": 351 }, { "epoch": 0.592842105263158, "grad_norm": 0.9574116468429565, "learning_rate": 0.001, "loss": 2.1836, "step": 352 }, { "epoch": 0.5945263157894737, "grad_norm": 0.9625939726829529, "learning_rate": 0.001, "loss": 2.0379, "step": 353 }, { "epoch": 0.5962105263157895, "grad_norm": 0.9140836000442505, "learning_rate": 0.001, "loss": 2.0564, "step": 354 }, { "epoch": 0.5978947368421053, "grad_norm": 0.9520573616027832, "learning_rate": 0.001, "loss": 1.9383, "step": 355 }, { "epoch": 0.5995789473684211, "grad_norm": 0.875503659248352, "learning_rate": 0.001, "loss": 1.6936, "step": 356 }, { "epoch": 0.6012631578947368, "grad_norm": 1.461020827293396, "learning_rate": 0.001, "loss": 2.2684, "step": 357 }, { "epoch": 0.6029473684210527, "grad_norm": 0.8192405700683594, "learning_rate": 0.001, "loss": 1.8995, "step": 358 }, { "epoch": 0.6046315789473684, "grad_norm": 1.4530872106552124, "learning_rate": 0.001, "loss": 1.8539, "step": 359 }, { "epoch": 0.6063157894736843, "grad_norm": 0.959186315536499, "learning_rate": 0.001, "loss": 2.0851, "step": 360 }, { "epoch": 0.608, "grad_norm": 0.8276315331459045, "learning_rate": 0.001, "loss": 1.9521, "step": 361 }, { "epoch": 0.6096842105263158, "grad_norm": 1.2478163242340088, "learning_rate": 0.001, "loss": 1.9309, "step": 362 }, { "epoch": 0.6113684210526316, "grad_norm": 1.1320995092391968, "learning_rate": 0.001, "loss": 1.9807, "step": 363 }, { "epoch": 0.6130526315789474, "grad_norm": 0.9767136573791504, "learning_rate": 0.001, "loss": 1.9802, "step": 364 }, { "epoch": 0.6147368421052631, "grad_norm": 0.8936948776245117, "learning_rate": 0.001, "loss": 1.9986, "step": 365 }, { "epoch": 0.616421052631579, "grad_norm": 0.7911234498023987, "learning_rate": 0.001, "loss": 2.0896, "step": 366 }, { "epoch": 0.6181052631578947, "grad_norm": 0.7824344635009766, "learning_rate": 0.001, "loss": 1.4767, "step": 367 }, { "epoch": 0.6197894736842106, "grad_norm": 0.9858822822570801, "learning_rate": 0.001, "loss": 1.9758, "step": 368 }, { "epoch": 0.6214736842105263, "grad_norm": 0.942699670791626, "learning_rate": 0.001, "loss": 1.9038, "step": 369 }, { "epoch": 0.6231578947368421, "grad_norm": 1.0846315622329712, "learning_rate": 0.001, "loss": 1.8084, "step": 370 }, { "epoch": 0.6248421052631579, "grad_norm": 0.9172139167785645, "learning_rate": 0.001, "loss": 1.8795, "step": 371 }, { "epoch": 0.6265263157894737, "grad_norm": 0.8866816163063049, "learning_rate": 0.001, "loss": 1.8835, "step": 372 }, { "epoch": 0.6282105263157894, "grad_norm": 0.8923367261886597, "learning_rate": 0.001, "loss": 2.1602, "step": 373 }, { "epoch": 0.6298947368421053, "grad_norm": 0.9485911130905151, "learning_rate": 0.001, "loss": 2.0832, "step": 374 }, { "epoch": 0.631578947368421, "grad_norm": 1.026877999305725, "learning_rate": 0.001, "loss": 2.283, "step": 375 }, { "epoch": 0.6332631578947369, "grad_norm": 0.9710808396339417, "learning_rate": 0.001, "loss": 1.6655, "step": 376 }, { "epoch": 0.6349473684210526, "grad_norm": 1.1433035135269165, "learning_rate": 0.001, "loss": 1.9597, "step": 377 }, { "epoch": 0.6366315789473684, "grad_norm": 0.9025890827178955, "learning_rate": 0.001, "loss": 2.0539, "step": 378 }, { "epoch": 0.6383157894736842, "grad_norm": 0.9245177507400513, "learning_rate": 0.001, "loss": 1.7705, "step": 379 }, { "epoch": 0.64, "grad_norm": 0.932959258556366, "learning_rate": 0.001, "loss": 2.0923, "step": 380 }, { "epoch": 0.6416842105263157, "grad_norm": 0.9858509302139282, "learning_rate": 0.001, "loss": 1.9338, "step": 381 }, { "epoch": 0.6433684210526316, "grad_norm": 1.0888968706130981, "learning_rate": 0.001, "loss": 1.8463, "step": 382 }, { "epoch": 0.6450526315789473, "grad_norm": 0.9424766302108765, "learning_rate": 0.001, "loss": 1.8135, "step": 383 }, { "epoch": 0.6467368421052632, "grad_norm": 0.955096960067749, "learning_rate": 0.001, "loss": 1.9355, "step": 384 }, { "epoch": 0.6484210526315789, "grad_norm": 0.9020712375640869, "learning_rate": 0.001, "loss": 2.2235, "step": 385 }, { "epoch": 0.6501052631578947, "grad_norm": 1.2948638200759888, "learning_rate": 0.001, "loss": 1.8841, "step": 386 }, { "epoch": 0.6517894736842105, "grad_norm": 1.1215901374816895, "learning_rate": 0.001, "loss": 2.3204, "step": 387 }, { "epoch": 0.6534736842105263, "grad_norm": 1.000780701637268, "learning_rate": 0.001, "loss": 1.9545, "step": 388 }, { "epoch": 0.655157894736842, "grad_norm": 1.0688225030899048, "learning_rate": 0.001, "loss": 2.1332, "step": 389 }, { "epoch": 0.6568421052631579, "grad_norm": 0.8454869985580444, "learning_rate": 0.001, "loss": 2.2442, "step": 390 }, { "epoch": 0.6585263157894737, "grad_norm": 1.0029394626617432, "learning_rate": 0.001, "loss": 1.9708, "step": 391 }, { "epoch": 0.6602105263157895, "grad_norm": 1.2006776332855225, "learning_rate": 0.001, "loss": 1.9889, "step": 392 }, { "epoch": 0.6618947368421053, "grad_norm": 0.8848575949668884, "learning_rate": 0.001, "loss": 1.8719, "step": 393 }, { "epoch": 0.663578947368421, "grad_norm": 0.9433349370956421, "learning_rate": 0.001, "loss": 1.807, "step": 394 }, { "epoch": 0.6652631578947369, "grad_norm": 1.1462281942367554, "learning_rate": 0.001, "loss": 1.9315, "step": 395 }, { "epoch": 0.6669473684210526, "grad_norm": 0.9325450658798218, "learning_rate": 0.001, "loss": 1.701, "step": 396 }, { "epoch": 0.6686315789473685, "grad_norm": 1.1033800840377808, "learning_rate": 0.001, "loss": 2.0728, "step": 397 }, { "epoch": 0.6703157894736842, "grad_norm": 1.0380494594573975, "learning_rate": 0.001, "loss": 1.9656, "step": 398 }, { "epoch": 0.672, "grad_norm": 0.8989469408988953, "learning_rate": 0.001, "loss": 1.9998, "step": 399 }, { "epoch": 0.6736842105263158, "grad_norm": 1.1360111236572266, "learning_rate": 0.001, "loss": 2.1114, "step": 400 }, { "epoch": 0.6753684210526316, "grad_norm": 0.9334345459938049, "learning_rate": 0.001, "loss": 1.7843, "step": 401 }, { "epoch": 0.6770526315789474, "grad_norm": 0.8565030097961426, "learning_rate": 0.001, "loss": 1.8634, "step": 402 }, { "epoch": 0.6787368421052632, "grad_norm": 0.926354706287384, "learning_rate": 0.001, "loss": 2.1583, "step": 403 }, { "epoch": 0.6804210526315789, "grad_norm": 1.041408896446228, "learning_rate": 0.001, "loss": 2.0652, "step": 404 }, { "epoch": 0.6821052631578948, "grad_norm": 0.8957986235618591, "learning_rate": 0.001, "loss": 1.9705, "step": 405 }, { "epoch": 0.6837894736842105, "grad_norm": 0.866303026676178, "learning_rate": 0.001, "loss": 1.9173, "step": 406 }, { "epoch": 0.6854736842105263, "grad_norm": 0.9298515915870667, "learning_rate": 0.001, "loss": 1.4851, "step": 407 }, { "epoch": 0.6871578947368421, "grad_norm": 0.8325154781341553, "learning_rate": 0.001, "loss": 1.9217, "step": 408 }, { "epoch": 0.6888421052631579, "grad_norm": 1.0331366062164307, "learning_rate": 0.001, "loss": 2.2276, "step": 409 }, { "epoch": 0.6905263157894737, "grad_norm": 0.9623380899429321, "learning_rate": 0.001, "loss": 1.9401, "step": 410 }, { "epoch": 0.6922105263157895, "grad_norm": 0.9502870440483093, "learning_rate": 0.001, "loss": 2.0698, "step": 411 }, { "epoch": 0.6938947368421052, "grad_norm": 0.7351365089416504, "learning_rate": 0.001, "loss": 1.8196, "step": 412 }, { "epoch": 0.6955789473684211, "grad_norm": 1.4265284538269043, "learning_rate": 0.001, "loss": 2.1069, "step": 413 }, { "epoch": 0.6972631578947368, "grad_norm": 0.9151477813720703, "learning_rate": 0.001, "loss": 1.9652, "step": 414 }, { "epoch": 0.6989473684210527, "grad_norm": 1.0058677196502686, "learning_rate": 0.001, "loss": 1.9402, "step": 415 }, { "epoch": 0.7006315789473684, "grad_norm": 0.8107333183288574, "learning_rate": 0.001, "loss": 1.8946, "step": 416 }, { "epoch": 0.7023157894736842, "grad_norm": 0.9497429132461548, "learning_rate": 0.001, "loss": 2.0289, "step": 417 }, { "epoch": 0.704, "grad_norm": 0.9337472319602966, "learning_rate": 0.001, "loss": 1.9517, "step": 418 }, { "epoch": 0.7056842105263158, "grad_norm": 0.872475802898407, "learning_rate": 0.001, "loss": 2.0209, "step": 419 }, { "epoch": 0.7073684210526315, "grad_norm": 0.9438268542289734, "learning_rate": 0.001, "loss": 1.9209, "step": 420 }, { "epoch": 0.7090526315789474, "grad_norm": 1.2881578207015991, "learning_rate": 0.001, "loss": 2.1235, "step": 421 }, { "epoch": 0.7107368421052631, "grad_norm": 0.8764305114746094, "learning_rate": 0.001, "loss": 1.9836, "step": 422 }, { "epoch": 0.712421052631579, "grad_norm": 1.232689619064331, "learning_rate": 0.001, "loss": 2.2588, "step": 423 }, { "epoch": 0.7141052631578947, "grad_norm": 0.9619866013526917, "learning_rate": 0.001, "loss": 2.1062, "step": 424 }, { "epoch": 0.7157894736842105, "grad_norm": 0.9023774266242981, "learning_rate": 0.001, "loss": 1.9278, "step": 425 }, { "epoch": 0.7174736842105263, "grad_norm": 1.1033554077148438, "learning_rate": 0.001, "loss": 1.7297, "step": 426 }, { "epoch": 0.7191578947368421, "grad_norm": 0.7463766932487488, "learning_rate": 0.001, "loss": 1.725, "step": 427 }, { "epoch": 0.7208421052631578, "grad_norm": 0.9457252621650696, "learning_rate": 0.001, "loss": 1.7739, "step": 428 }, { "epoch": 0.7225263157894737, "grad_norm": 0.934600293636322, "learning_rate": 0.001, "loss": 2.1775, "step": 429 }, { "epoch": 0.7242105263157895, "grad_norm": 0.8955628871917725, "learning_rate": 0.001, "loss": 2.2009, "step": 430 }, { "epoch": 0.7258947368421053, "grad_norm": 0.94439697265625, "learning_rate": 0.001, "loss": 2.244, "step": 431 }, { "epoch": 0.7275789473684211, "grad_norm": 0.8255289793014526, "learning_rate": 0.001, "loss": 1.8666, "step": 432 }, { "epoch": 0.7292631578947368, "grad_norm": 0.8670260906219482, "learning_rate": 0.001, "loss": 1.7115, "step": 433 }, { "epoch": 0.7309473684210527, "grad_norm": 0.8938360214233398, "learning_rate": 0.001, "loss": 1.6273, "step": 434 }, { "epoch": 0.7326315789473684, "grad_norm": 1.236402988433838, "learning_rate": 0.001, "loss": 2.0744, "step": 435 }, { "epoch": 0.7343157894736843, "grad_norm": 0.8387994170188904, "learning_rate": 0.001, "loss": 1.9027, "step": 436 }, { "epoch": 0.736, "grad_norm": 0.8984929323196411, "learning_rate": 0.001, "loss": 2.1411, "step": 437 }, { "epoch": 0.7376842105263158, "grad_norm": 0.9328674077987671, "learning_rate": 0.001, "loss": 2.2223, "step": 438 }, { "epoch": 0.7393684210526316, "grad_norm": 0.8986714482307434, "learning_rate": 0.001, "loss": 1.9568, "step": 439 }, { "epoch": 0.7410526315789474, "grad_norm": 0.9492053389549255, "learning_rate": 0.001, "loss": 2.2018, "step": 440 }, { "epoch": 0.7427368421052631, "grad_norm": 0.8261067867279053, "learning_rate": 0.001, "loss": 2.0828, "step": 441 }, { "epoch": 0.744421052631579, "grad_norm": 1.0723634958267212, "learning_rate": 0.001, "loss": 2.2274, "step": 442 }, { "epoch": 0.7461052631578947, "grad_norm": 0.933506190776825, "learning_rate": 0.001, "loss": 2.0962, "step": 443 }, { "epoch": 0.7477894736842106, "grad_norm": 0.7961298227310181, "learning_rate": 0.001, "loss": 2.1599, "step": 444 }, { "epoch": 0.7494736842105263, "grad_norm": 0.9804624319076538, "learning_rate": 0.001, "loss": 2.2872, "step": 445 }, { "epoch": 0.7511578947368421, "grad_norm": 0.8495241403579712, "learning_rate": 0.001, "loss": 1.9647, "step": 446 }, { "epoch": 0.7528421052631579, "grad_norm": 0.9075875878334045, "learning_rate": 0.001, "loss": 2.1734, "step": 447 }, { "epoch": 0.7528421052631579, "eval_loss": 1.7126047611236572, "eval_runtime": 0.0791, "eval_samples_per_second": 12.634, "eval_steps_per_second": 12.634, "step": 447 }, { "epoch": 0.7545263157894737, "grad_norm": 0.8501783013343811, "learning_rate": 0.001, "loss": 1.7057, "step": 448 }, { "epoch": 0.7562105263157894, "grad_norm": 0.9169737100601196, "learning_rate": 0.001, "loss": 2.0444, "step": 449 }, { "epoch": 0.7578947368421053, "grad_norm": 0.8813796043395996, "learning_rate": 0.001, "loss": 1.4839, "step": 450 }, { "epoch": 0.759578947368421, "grad_norm": 0.9467214345932007, "learning_rate": 0.001, "loss": 1.8931, "step": 451 }, { "epoch": 0.7612631578947369, "grad_norm": 0.8730418086051941, "learning_rate": 0.001, "loss": 1.7104, "step": 452 }, { "epoch": 0.7629473684210526, "grad_norm": 0.8568257093429565, "learning_rate": 0.001, "loss": 1.998, "step": 453 }, { "epoch": 0.7646315789473684, "grad_norm": 0.8972058892250061, "learning_rate": 0.001, "loss": 1.8552, "step": 454 }, { "epoch": 0.7663157894736842, "grad_norm": 0.9764978289604187, "learning_rate": 0.001, "loss": 2.0078, "step": 455 }, { "epoch": 0.768, "grad_norm": 0.9850934147834778, "learning_rate": 0.001, "loss": 1.9529, "step": 456 }, { "epoch": 0.7696842105263157, "grad_norm": 0.8368688225746155, "learning_rate": 0.001, "loss": 2.1118, "step": 457 }, { "epoch": 0.7713684210526316, "grad_norm": 0.9366937875747681, "learning_rate": 0.001, "loss": 1.7256, "step": 458 }, { "epoch": 0.7730526315789473, "grad_norm": 0.9020842909812927, "learning_rate": 0.001, "loss": 1.697, "step": 459 }, { "epoch": 0.7747368421052632, "grad_norm": 0.9683967232704163, "learning_rate": 0.001, "loss": 1.8288, "step": 460 }, { "epoch": 0.7764210526315789, "grad_norm": 0.926777720451355, "learning_rate": 0.001, "loss": 1.7249, "step": 461 }, { "epoch": 0.7781052631578947, "grad_norm": 1.0532008409500122, "learning_rate": 0.001, "loss": 2.3132, "step": 462 }, { "epoch": 0.7797894736842105, "grad_norm": 0.8814186453819275, "learning_rate": 0.001, "loss": 1.9781, "step": 463 }, { "epoch": 0.7814736842105263, "grad_norm": 0.78118896484375, "learning_rate": 0.001, "loss": 1.7407, "step": 464 }, { "epoch": 0.783157894736842, "grad_norm": 0.9273678064346313, "learning_rate": 0.001, "loss": 2.2367, "step": 465 }, { "epoch": 0.7848421052631579, "grad_norm": 0.9391574263572693, "learning_rate": 0.001, "loss": 1.9642, "step": 466 }, { "epoch": 0.7865263157894736, "grad_norm": 1.0936115980148315, "learning_rate": 0.001, "loss": 2.2852, "step": 467 }, { "epoch": 0.7882105263157895, "grad_norm": 1.0484883785247803, "learning_rate": 0.001, "loss": 1.8051, "step": 468 }, { "epoch": 0.7898947368421053, "grad_norm": 1.0337159633636475, "learning_rate": 0.001, "loss": 2.0115, "step": 469 }, { "epoch": 0.791578947368421, "grad_norm": 0.8993913531303406, "learning_rate": 0.001, "loss": 1.7269, "step": 470 }, { "epoch": 0.7932631578947369, "grad_norm": 0.8433123826980591, "learning_rate": 0.001, "loss": 2.0798, "step": 471 }, { "epoch": 0.7949473684210526, "grad_norm": 0.9056971669197083, "learning_rate": 0.001, "loss": 2.3624, "step": 472 }, { "epoch": 0.7966315789473685, "grad_norm": 0.8916146159172058, "learning_rate": 0.001, "loss": 1.6266, "step": 473 }, { "epoch": 0.7983157894736842, "grad_norm": 0.8712090253829956, "learning_rate": 0.001, "loss": 2.2228, "step": 474 }, { "epoch": 0.8, "grad_norm": 0.8916635513305664, "learning_rate": 0.001, "loss": 1.9197, "step": 475 }, { "epoch": 0.8016842105263158, "grad_norm": 0.9295555949211121, "learning_rate": 0.001, "loss": 1.7184, "step": 476 }, { "epoch": 0.8033684210526316, "grad_norm": 0.8718081116676331, "learning_rate": 0.001, "loss": 1.9394, "step": 477 }, { "epoch": 0.8050526315789474, "grad_norm": 1.0214195251464844, "learning_rate": 0.001, "loss": 1.8555, "step": 478 }, { "epoch": 0.8067368421052632, "grad_norm": 0.9039567708969116, "learning_rate": 0.001, "loss": 2.0961, "step": 479 }, { "epoch": 0.8084210526315789, "grad_norm": 1.088826298713684, "learning_rate": 0.001, "loss": 2.1166, "step": 480 }, { "epoch": 0.8101052631578948, "grad_norm": 0.9610921144485474, "learning_rate": 0.001, "loss": 2.1177, "step": 481 }, { "epoch": 0.8117894736842105, "grad_norm": 0.8916026949882507, "learning_rate": 0.001, "loss": 1.9126, "step": 482 }, { "epoch": 0.8134736842105263, "grad_norm": 0.9478291273117065, "learning_rate": 0.001, "loss": 2.0522, "step": 483 }, { "epoch": 0.8151578947368421, "grad_norm": 1.1988763809204102, "learning_rate": 0.001, "loss": 2.1891, "step": 484 }, { "epoch": 0.8168421052631579, "grad_norm": 1.06550133228302, "learning_rate": 0.001, "loss": 2.0575, "step": 485 }, { "epoch": 0.8185263157894737, "grad_norm": 1.0411326885223389, "learning_rate": 0.001, "loss": 2.2342, "step": 486 }, { "epoch": 0.8202105263157895, "grad_norm": 0.996446967124939, "learning_rate": 0.001, "loss": 2.3195, "step": 487 }, { "epoch": 0.8218947368421052, "grad_norm": 0.9175570011138916, "learning_rate": 0.001, "loss": 1.9672, "step": 488 }, { "epoch": 0.8235789473684211, "grad_norm": 0.9615358710289001, "learning_rate": 0.001, "loss": 1.7088, "step": 489 }, { "epoch": 0.8252631578947368, "grad_norm": 1.0030488967895508, "learning_rate": 0.001, "loss": 1.9093, "step": 490 }, { "epoch": 0.8269473684210527, "grad_norm": 0.8371875882148743, "learning_rate": 0.001, "loss": 2.0981, "step": 491 }, { "epoch": 0.8286315789473684, "grad_norm": 1.0094941854476929, "learning_rate": 0.001, "loss": 1.9831, "step": 492 }, { "epoch": 0.8303157894736842, "grad_norm": 0.6985566020011902, "learning_rate": 0.001, "loss": 1.481, "step": 493 }, { "epoch": 0.832, "grad_norm": 0.8714520335197449, "learning_rate": 0.001, "loss": 1.8478, "step": 494 }, { "epoch": 0.8336842105263158, "grad_norm": 0.9632995128631592, "learning_rate": 0.001, "loss": 1.9434, "step": 495 }, { "epoch": 0.8353684210526315, "grad_norm": 0.989193320274353, "learning_rate": 0.001, "loss": 2.1274, "step": 496 }, { "epoch": 0.8370526315789474, "grad_norm": 0.9159345626831055, "learning_rate": 0.001, "loss": 2.0572, "step": 497 }, { "epoch": 0.8387368421052631, "grad_norm": 0.8769975304603577, "learning_rate": 0.001, "loss": 1.6761, "step": 498 }, { "epoch": 0.840421052631579, "grad_norm": 1.0717746019363403, "learning_rate": 0.001, "loss": 1.964, "step": 499 }, { "epoch": 0.8421052631578947, "grad_norm": 1.0184441804885864, "learning_rate": 0.001, "loss": 2.0598, "step": 500 }, { "epoch": 0.8437894736842105, "grad_norm": 1.1409798860549927, "learning_rate": 0.001, "loss": 1.57, "step": 501 }, { "epoch": 0.8454736842105263, "grad_norm": 0.917641818523407, "learning_rate": 0.001, "loss": 1.7864, "step": 502 }, { "epoch": 0.8471578947368421, "grad_norm": 0.9784872531890869, "learning_rate": 0.001, "loss": 2.2752, "step": 503 }, { "epoch": 0.8488421052631578, "grad_norm": 1.012035846710205, "learning_rate": 0.001, "loss": 1.9572, "step": 504 }, { "epoch": 0.8505263157894737, "grad_norm": 0.8496696949005127, "learning_rate": 0.001, "loss": 1.5138, "step": 505 }, { "epoch": 0.8522105263157894, "grad_norm": 1.0498113632202148, "learning_rate": 0.001, "loss": 2.1815, "step": 506 }, { "epoch": 0.8538947368421053, "grad_norm": 0.8694950938224792, "learning_rate": 0.001, "loss": 1.9297, "step": 507 }, { "epoch": 0.8555789473684211, "grad_norm": 0.9349990487098694, "learning_rate": 0.001, "loss": 1.9476, "step": 508 }, { "epoch": 0.8572631578947368, "grad_norm": 0.9136828184127808, "learning_rate": 0.001, "loss": 1.9822, "step": 509 }, { "epoch": 0.8589473684210527, "grad_norm": 0.9763911366462708, "learning_rate": 0.001, "loss": 2.0084, "step": 510 }, { "epoch": 0.8606315789473684, "grad_norm": 0.9221206307411194, "learning_rate": 0.001, "loss": 1.9916, "step": 511 }, { "epoch": 0.8623157894736843, "grad_norm": 0.9817140698432922, "learning_rate": 0.001, "loss": 1.9887, "step": 512 }, { "epoch": 0.864, "grad_norm": 1.166142463684082, "learning_rate": 0.001, "loss": 2.0389, "step": 513 }, { "epoch": 0.8656842105263158, "grad_norm": 0.9571674466133118, "learning_rate": 0.001, "loss": 1.6596, "step": 514 }, { "epoch": 0.8673684210526316, "grad_norm": 1.113754391670227, "learning_rate": 0.001, "loss": 1.7947, "step": 515 }, { "epoch": 0.8690526315789474, "grad_norm": 0.8172054290771484, "learning_rate": 0.001, "loss": 2.1411, "step": 516 }, { "epoch": 0.8707368421052631, "grad_norm": 0.7713819742202759, "learning_rate": 0.001, "loss": 1.6631, "step": 517 }, { "epoch": 0.872421052631579, "grad_norm": 1.0675171613693237, "learning_rate": 0.001, "loss": 2.115, "step": 518 }, { "epoch": 0.8741052631578947, "grad_norm": 0.9892959594726562, "learning_rate": 0.001, "loss": 2.0591, "step": 519 }, { "epoch": 0.8757894736842106, "grad_norm": 1.017602562904358, "learning_rate": 0.001, "loss": 1.8098, "step": 520 }, { "epoch": 0.8774736842105263, "grad_norm": 1.0160664319992065, "learning_rate": 0.001, "loss": 2.0635, "step": 521 }, { "epoch": 0.8791578947368421, "grad_norm": 0.7594957947731018, "learning_rate": 0.001, "loss": 1.9662, "step": 522 }, { "epoch": 0.8808421052631579, "grad_norm": 0.9666638374328613, "learning_rate": 0.001, "loss": 2.3076, "step": 523 }, { "epoch": 0.8825263157894737, "grad_norm": 1.0217084884643555, "learning_rate": 0.001, "loss": 2.1043, "step": 524 }, { "epoch": 0.8842105263157894, "grad_norm": 0.9620029330253601, "learning_rate": 0.001, "loss": 1.9836, "step": 525 }, { "epoch": 0.8858947368421053, "grad_norm": 0.9031502604484558, "learning_rate": 0.001, "loss": 1.8648, "step": 526 }, { "epoch": 0.887578947368421, "grad_norm": 1.0853564739227295, "learning_rate": 0.001, "loss": 2.2491, "step": 527 }, { "epoch": 0.8892631578947369, "grad_norm": 0.850796639919281, "learning_rate": 0.001, "loss": 1.8988, "step": 528 }, { "epoch": 0.8909473684210526, "grad_norm": 0.9813094735145569, "learning_rate": 0.001, "loss": 2.0652, "step": 529 }, { "epoch": 0.8926315789473684, "grad_norm": 1.1805915832519531, "learning_rate": 0.001, "loss": 1.9761, "step": 530 }, { "epoch": 0.8943157894736842, "grad_norm": 0.91289222240448, "learning_rate": 0.001, "loss": 2.2265, "step": 531 }, { "epoch": 0.896, "grad_norm": 0.8116694092750549, "learning_rate": 0.001, "loss": 2.0667, "step": 532 }, { "epoch": 0.8976842105263158, "grad_norm": 1.0094329118728638, "learning_rate": 0.001, "loss": 2.2099, "step": 533 }, { "epoch": 0.8993684210526316, "grad_norm": 0.9709912538528442, "learning_rate": 0.001, "loss": 1.7337, "step": 534 }, { "epoch": 0.9010526315789473, "grad_norm": 1.1453299522399902, "learning_rate": 0.001, "loss": 2.0997, "step": 535 }, { "epoch": 0.9027368421052632, "grad_norm": 1.097075343132019, "learning_rate": 0.001, "loss": 2.1082, "step": 536 }, { "epoch": 0.9044210526315789, "grad_norm": 0.9422668218612671, "learning_rate": 0.001, "loss": 2.2451, "step": 537 }, { "epoch": 0.9061052631578947, "grad_norm": 0.8550103902816772, "learning_rate": 0.001, "loss": 1.8164, "step": 538 }, { "epoch": 0.9077894736842105, "grad_norm": 1.0287948846817017, "learning_rate": 0.001, "loss": 1.8799, "step": 539 }, { "epoch": 0.9094736842105263, "grad_norm": 1.0734996795654297, "learning_rate": 0.001, "loss": 2.1292, "step": 540 }, { "epoch": 0.9111578947368421, "grad_norm": 0.9072704315185547, "learning_rate": 0.001, "loss": 1.9709, "step": 541 }, { "epoch": 0.9128421052631579, "grad_norm": 0.9148054718971252, "learning_rate": 0.001, "loss": 1.7025, "step": 542 }, { "epoch": 0.9145263157894736, "grad_norm": 1.094970941543579, "learning_rate": 0.001, "loss": 2.0636, "step": 543 }, { "epoch": 0.9162105263157895, "grad_norm": 0.8233382701873779, "learning_rate": 0.001, "loss": 2.0036, "step": 544 }, { "epoch": 0.9178947368421052, "grad_norm": 0.789847731590271, "learning_rate": 0.001, "loss": 1.9641, "step": 545 }, { "epoch": 0.919578947368421, "grad_norm": 1.1164312362670898, "learning_rate": 0.001, "loss": 2.0094, "step": 546 }, { "epoch": 0.9212631578947369, "grad_norm": 0.9254114031791687, "learning_rate": 0.001, "loss": 2.0071, "step": 547 }, { "epoch": 0.9229473684210526, "grad_norm": 0.772678017616272, "learning_rate": 0.001, "loss": 1.6871, "step": 548 }, { "epoch": 0.9246315789473685, "grad_norm": 0.8702642321586609, "learning_rate": 0.001, "loss": 1.86, "step": 549 }, { "epoch": 0.9263157894736842, "grad_norm": 0.9259825348854065, "learning_rate": 0.001, "loss": 2.2161, "step": 550 }, { "epoch": 0.928, "grad_norm": 1.095923662185669, "learning_rate": 0.001, "loss": 2.2506, "step": 551 }, { "epoch": 0.9296842105263158, "grad_norm": 0.8884152770042419, "learning_rate": 0.001, "loss": 1.9812, "step": 552 }, { "epoch": 0.9313684210526316, "grad_norm": 1.201453685760498, "learning_rate": 0.001, "loss": 2.3967, "step": 553 }, { "epoch": 0.9330526315789474, "grad_norm": 0.8036067485809326, "learning_rate": 0.001, "loss": 2.0411, "step": 554 }, { "epoch": 0.9347368421052632, "grad_norm": 1.1204805374145508, "learning_rate": 0.001, "loss": 1.9963, "step": 555 }, { "epoch": 0.9364210526315789, "grad_norm": 0.9753801822662354, "learning_rate": 0.001, "loss": 2.0774, "step": 556 }, { "epoch": 0.9381052631578948, "grad_norm": 0.9044502973556519, "learning_rate": 0.001, "loss": 2.2171, "step": 557 }, { "epoch": 0.9397894736842105, "grad_norm": 0.872684121131897, "learning_rate": 0.001, "loss": 2.1116, "step": 558 }, { "epoch": 0.9414736842105264, "grad_norm": 0.9896319508552551, "learning_rate": 0.001, "loss": 1.9894, "step": 559 }, { "epoch": 0.9431578947368421, "grad_norm": 0.8896704912185669, "learning_rate": 0.001, "loss": 1.83, "step": 560 }, { "epoch": 0.9448421052631579, "grad_norm": 1.1786295175552368, "learning_rate": 0.001, "loss": 1.3999, "step": 561 }, { "epoch": 0.9465263157894737, "grad_norm": 1.0260367393493652, "learning_rate": 0.001, "loss": 1.7856, "step": 562 }, { "epoch": 0.9482105263157895, "grad_norm": 1.1223475933074951, "learning_rate": 0.001, "loss": 1.8988, "step": 563 }, { "epoch": 0.9498947368421052, "grad_norm": 1.2666854858398438, "learning_rate": 0.001, "loss": 2.4208, "step": 564 }, { "epoch": 0.9515789473684211, "grad_norm": 0.8746899366378784, "learning_rate": 0.001, "loss": 2.0621, "step": 565 }, { "epoch": 0.9532631578947368, "grad_norm": 1.1559736728668213, "learning_rate": 0.001, "loss": 1.9889, "step": 566 }, { "epoch": 0.9549473684210527, "grad_norm": 0.8954889178276062, "learning_rate": 0.001, "loss": 2.1674, "step": 567 }, { "epoch": 0.9566315789473684, "grad_norm": 0.848497748374939, "learning_rate": 0.001, "loss": 2.1309, "step": 568 }, { "epoch": 0.9583157894736842, "grad_norm": 0.9832907319068909, "learning_rate": 0.001, "loss": 1.9696, "step": 569 }, { "epoch": 0.96, "grad_norm": 1.004216194152832, "learning_rate": 0.001, "loss": 2.1556, "step": 570 }, { "epoch": 0.9616842105263158, "grad_norm": 1.2567979097366333, "learning_rate": 0.001, "loss": 1.925, "step": 571 }, { "epoch": 0.9633684210526315, "grad_norm": 1.056889533996582, "learning_rate": 0.001, "loss": 1.7742, "step": 572 }, { "epoch": 0.9650526315789474, "grad_norm": 0.7364183664321899, "learning_rate": 0.001, "loss": 1.7558, "step": 573 }, { "epoch": 0.9667368421052631, "grad_norm": 1.1390528678894043, "learning_rate": 0.001, "loss": 1.8911, "step": 574 }, { "epoch": 0.968421052631579, "grad_norm": 1.0370616912841797, "learning_rate": 0.001, "loss": 2.091, "step": 575 }, { "epoch": 0.9701052631578947, "grad_norm": 0.8506829738616943, "learning_rate": 0.001, "loss": 2.2052, "step": 576 }, { "epoch": 0.9717894736842105, "grad_norm": 0.8810485601425171, "learning_rate": 0.001, "loss": 2.2049, "step": 577 }, { "epoch": 0.9734736842105263, "grad_norm": 1.0657389163970947, "learning_rate": 0.001, "loss": 2.1982, "step": 578 }, { "epoch": 0.9751578947368421, "grad_norm": 0.9130502343177795, "learning_rate": 0.001, "loss": 1.7093, "step": 579 }, { "epoch": 0.9768421052631578, "grad_norm": 0.9334926605224609, "learning_rate": 0.001, "loss": 1.8507, "step": 580 }, { "epoch": 0.9785263157894737, "grad_norm": 0.9801141023635864, "learning_rate": 0.001, "loss": 2.0824, "step": 581 }, { "epoch": 0.9802105263157894, "grad_norm": 1.1863903999328613, "learning_rate": 0.001, "loss": 2.0316, "step": 582 }, { "epoch": 0.9818947368421053, "grad_norm": 0.9224819540977478, "learning_rate": 0.001, "loss": 2.1538, "step": 583 }, { "epoch": 0.983578947368421, "grad_norm": 0.9597621560096741, "learning_rate": 0.001, "loss": 2.0054, "step": 584 }, { "epoch": 0.9852631578947368, "grad_norm": 1.1641483306884766, "learning_rate": 0.001, "loss": 1.9722, "step": 585 }, { "epoch": 0.9869473684210527, "grad_norm": 0.9333254098892212, "learning_rate": 0.001, "loss": 2.0654, "step": 586 }, { "epoch": 0.9886315789473684, "grad_norm": 1.0137170553207397, "learning_rate": 0.001, "loss": 2.198, "step": 587 }, { "epoch": 0.9903157894736843, "grad_norm": 0.9962888360023499, "learning_rate": 0.001, "loss": 1.9295, "step": 588 }, { "epoch": 0.992, "grad_norm": 1.1276189088821411, "learning_rate": 0.001, "loss": 1.9579, "step": 589 }, { "epoch": 0.9936842105263158, "grad_norm": 0.9350583553314209, "learning_rate": 0.001, "loss": 1.9998, "step": 590 }, { "epoch": 0.9953684210526316, "grad_norm": 0.97312992811203, "learning_rate": 0.001, "loss": 1.7338, "step": 591 }, { "epoch": 0.9970526315789474, "grad_norm": 1.0777076482772827, "learning_rate": 0.001, "loss": 1.7413, "step": 592 }, { "epoch": 0.9987368421052631, "grad_norm": 1.031515121459961, "learning_rate": 0.001, "loss": 2.3266, "step": 593 }, { "epoch": 0.9987368421052631, "step": 593, "total_flos": 3.231264518066995e+16, "train_loss": 1.9546177544038976, "train_runtime": 1901.8595, "train_samples_per_second": 4.995, "train_steps_per_second": 0.312 } ], "logging_steps": 1, "max_steps": 593, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 149, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 100, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.231264518066995e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }