Dolphin3.0-Mistral-24B / trainer_state.json
ehartford's picture
Upload folder using huggingface_hub
590d923 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.9970559371933267,
"eval_steps": 128,
"global_step": 1018,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.001962708537782139,
"grad_norm": 62.355553075278046,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.8187,
"step": 1
},
{
"epoch": 0.001962708537782139,
"eval_loss": 0.8275482654571533,
"eval_runtime": 246.1117,
"eval_samples_per_second": 123.33,
"eval_steps_per_second": 3.856,
"step": 1
},
{
"epoch": 0.003925417075564278,
"grad_norm": 62.468487550636745,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.8042,
"step": 2
},
{
"epoch": 0.005888125613346418,
"grad_norm": 4.2406831230151605,
"learning_rate": 3e-06,
"loss": 0.638,
"step": 3
},
{
"epoch": 0.007850834151128557,
"grad_norm": 0.6232773471755494,
"learning_rate": 4.000000000000001e-06,
"loss": 0.5818,
"step": 4
},
{
"epoch": 0.009813542688910697,
"grad_norm": 1.2988116476459395,
"learning_rate": 5e-06,
"loss": 0.5652,
"step": 5
},
{
"epoch": 0.011776251226692836,
"grad_norm": 19.326748341786118,
"learning_rate": 6e-06,
"loss": 0.988,
"step": 6
},
{
"epoch": 0.013738959764474975,
"grad_norm": 1.2669160024303416,
"learning_rate": 7e-06,
"loss": 0.6037,
"step": 7
},
{
"epoch": 0.015701668302257114,
"grad_norm": 1.2907800971966077,
"learning_rate": 8.000000000000001e-06,
"loss": 0.6027,
"step": 8
},
{
"epoch": 0.017664376840039256,
"grad_norm": 0.7888911759146412,
"learning_rate": 9e-06,
"loss": 0.5711,
"step": 9
},
{
"epoch": 0.019627085377821395,
"grad_norm": 0.7260839669652744,
"learning_rate": 1e-05,
"loss": 0.5624,
"step": 10
},
{
"epoch": 0.021589793915603533,
"grad_norm": 0.6532838806100564,
"learning_rate": 1.1000000000000001e-05,
"loss": 0.5612,
"step": 11
},
{
"epoch": 0.023552502453385672,
"grad_norm": 0.7559746737848189,
"learning_rate": 1.2e-05,
"loss": 0.5493,
"step": 12
},
{
"epoch": 0.02551521099116781,
"grad_norm": 0.5482270203080323,
"learning_rate": 1.3000000000000001e-05,
"loss": 0.5435,
"step": 13
},
{
"epoch": 0.02747791952894995,
"grad_norm": 0.4970646557650867,
"learning_rate": 1.4e-05,
"loss": 0.5408,
"step": 14
},
{
"epoch": 0.029440628066732092,
"grad_norm": 0.391725935764316,
"learning_rate": 1.5000000000000002e-05,
"loss": 0.5241,
"step": 15
},
{
"epoch": 0.03140333660451423,
"grad_norm": 0.43457047378514496,
"learning_rate": 1.6000000000000003e-05,
"loss": 0.5488,
"step": 16
},
{
"epoch": 0.033366045142296366,
"grad_norm": 0.3789337362265595,
"learning_rate": 1.7e-05,
"loss": 0.5266,
"step": 17
},
{
"epoch": 0.03532875368007851,
"grad_norm": 0.3331484005850679,
"learning_rate": 1.8e-05,
"loss": 0.5112,
"step": 18
},
{
"epoch": 0.03729146221786065,
"grad_norm": 0.3444864658048081,
"learning_rate": 1.9e-05,
"loss": 0.5051,
"step": 19
},
{
"epoch": 0.03925417075564279,
"grad_norm": 0.2887437422361683,
"learning_rate": 2e-05,
"loss": 0.5289,
"step": 20
},
{
"epoch": 0.04121687929342493,
"grad_norm": 0.278825983002472,
"learning_rate": 1.9999987858045354e-05,
"loss": 0.5289,
"step": 21
},
{
"epoch": 0.04317958783120707,
"grad_norm": 0.39102343961692493,
"learning_rate": 1.9999951432210905e-05,
"loss": 0.5076,
"step": 22
},
{
"epoch": 0.045142296368989206,
"grad_norm": 0.5521017110831627,
"learning_rate": 1.9999890722585106e-05,
"loss": 0.5108,
"step": 23
},
{
"epoch": 0.047105004906771344,
"grad_norm": 0.9336591410925016,
"learning_rate": 1.9999805729315383e-05,
"loss": 0.5316,
"step": 24
},
{
"epoch": 0.04906771344455348,
"grad_norm": 0.8299578992443886,
"learning_rate": 1.9999696452608135e-05,
"loss": 0.5334,
"step": 25
},
{
"epoch": 0.05103042198233562,
"grad_norm": 4.751183353378357,
"learning_rate": 1.999956289272873e-05,
"loss": 0.5285,
"step": 26
},
{
"epoch": 0.05299313052011776,
"grad_norm": 0.860325982956048,
"learning_rate": 1.99994050500015e-05,
"loss": 0.539,
"step": 27
},
{
"epoch": 0.0549558390578999,
"grad_norm": 0.4897569644596324,
"learning_rate": 1.999922292480975e-05,
"loss": 0.5358,
"step": 28
},
{
"epoch": 0.05691854759568204,
"grad_norm": 0.5398108233225083,
"learning_rate": 1.9999016517595752e-05,
"loss": 0.5351,
"step": 29
},
{
"epoch": 0.058881256133464184,
"grad_norm": 0.4780297052699839,
"learning_rate": 1.9998785828860744e-05,
"loss": 0.5185,
"step": 30
},
{
"epoch": 0.06084396467124632,
"grad_norm": 0.33693365999492547,
"learning_rate": 1.9998530859164926e-05,
"loss": 0.5104,
"step": 31
},
{
"epoch": 0.06280667320902845,
"grad_norm": 0.4261145101597809,
"learning_rate": 1.9998251609127465e-05,
"loss": 0.5005,
"step": 32
},
{
"epoch": 0.0647693817468106,
"grad_norm": 0.3067512941150541,
"learning_rate": 1.999794807942649e-05,
"loss": 0.508,
"step": 33
},
{
"epoch": 0.06673209028459273,
"grad_norm": 0.3756361155229969,
"learning_rate": 1.999762027079909e-05,
"loss": 0.5021,
"step": 34
},
{
"epoch": 0.06869479882237488,
"grad_norm": 0.3065423142817771,
"learning_rate": 1.9997268184041318e-05,
"loss": 0.5242,
"step": 35
},
{
"epoch": 0.07065750736015702,
"grad_norm": 0.27326520045279085,
"learning_rate": 1.9996891820008165e-05,
"loss": 0.5133,
"step": 36
},
{
"epoch": 0.07262021589793916,
"grad_norm": 0.2964526747556581,
"learning_rate": 1.9996491179613597e-05,
"loss": 0.492,
"step": 37
},
{
"epoch": 0.0745829244357213,
"grad_norm": 0.2446010040256801,
"learning_rate": 1.9996066263830533e-05,
"loss": 0.4952,
"step": 38
},
{
"epoch": 0.07654563297350343,
"grad_norm": 0.22448559939725016,
"learning_rate": 1.999561707369082e-05,
"loss": 0.4917,
"step": 39
},
{
"epoch": 0.07850834151128558,
"grad_norm": 0.22417659964598924,
"learning_rate": 1.9995143610285275e-05,
"loss": 0.4978,
"step": 40
},
{
"epoch": 0.08047105004906771,
"grad_norm": 0.24556309982732719,
"learning_rate": 1.9994645874763657e-05,
"loss": 0.5123,
"step": 41
},
{
"epoch": 0.08243375858684986,
"grad_norm": 0.2078136882784064,
"learning_rate": 1.9994123868334655e-05,
"loss": 0.5008,
"step": 42
},
{
"epoch": 0.08439646712463199,
"grad_norm": 0.2126670968682465,
"learning_rate": 1.999357759226591e-05,
"loss": 0.5028,
"step": 43
},
{
"epoch": 0.08635917566241413,
"grad_norm": 0.19399581978765076,
"learning_rate": 1.9993007047883988e-05,
"loss": 0.4937,
"step": 44
},
{
"epoch": 0.08832188420019627,
"grad_norm": 0.1873229933356675,
"learning_rate": 1.9992412236574396e-05,
"loss": 0.4798,
"step": 45
},
{
"epoch": 0.09028459273797841,
"grad_norm": 0.19589361661093946,
"learning_rate": 1.999179315978157e-05,
"loss": 0.4794,
"step": 46
},
{
"epoch": 0.09224730127576054,
"grad_norm": 0.174199073484297,
"learning_rate": 1.999114981900887e-05,
"loss": 0.4735,
"step": 47
},
{
"epoch": 0.09421000981354269,
"grad_norm": 0.16944247667745355,
"learning_rate": 1.999048221581858e-05,
"loss": 0.4843,
"step": 48
},
{
"epoch": 0.09617271835132483,
"grad_norm": 0.1854184220985238,
"learning_rate": 1.9989790351831898e-05,
"loss": 0.4774,
"step": 49
},
{
"epoch": 0.09813542688910697,
"grad_norm": 0.16797468279664923,
"learning_rate": 1.9989074228728942e-05,
"loss": 0.4684,
"step": 50
},
{
"epoch": 0.10009813542688911,
"grad_norm": 0.16773676945215363,
"learning_rate": 1.998833384824874e-05,
"loss": 0.4613,
"step": 51
},
{
"epoch": 0.10206084396467124,
"grad_norm": 0.1645777226690783,
"learning_rate": 1.9987569212189224e-05,
"loss": 0.477,
"step": 52
},
{
"epoch": 0.10402355250245339,
"grad_norm": 0.17954527682275045,
"learning_rate": 1.998678032240723e-05,
"loss": 0.4718,
"step": 53
},
{
"epoch": 0.10598626104023552,
"grad_norm": 0.16162280011064276,
"learning_rate": 1.9985967180818493e-05,
"loss": 0.4927,
"step": 54
},
{
"epoch": 0.10794896957801767,
"grad_norm": 0.16773939638216998,
"learning_rate": 1.9985129789397633e-05,
"loss": 0.4797,
"step": 55
},
{
"epoch": 0.1099116781157998,
"grad_norm": 0.18424558774001681,
"learning_rate": 1.998426815017817e-05,
"loss": 0.4747,
"step": 56
},
{
"epoch": 0.11187438665358194,
"grad_norm": 0.1388232795248698,
"learning_rate": 1.9983382265252494e-05,
"loss": 0.4633,
"step": 57
},
{
"epoch": 0.11383709519136408,
"grad_norm": 0.17752346108818134,
"learning_rate": 1.998247213677188e-05,
"loss": 0.4613,
"step": 58
},
{
"epoch": 0.11579980372914622,
"grad_norm": 0.14030828634333592,
"learning_rate": 1.9981537766946486e-05,
"loss": 0.4726,
"step": 59
},
{
"epoch": 0.11776251226692837,
"grad_norm": 0.17867869880789194,
"learning_rate": 1.9980579158045322e-05,
"loss": 0.4659,
"step": 60
},
{
"epoch": 0.1197252208047105,
"grad_norm": 0.17555178239215463,
"learning_rate": 1.9979596312396258e-05,
"loss": 0.4882,
"step": 61
},
{
"epoch": 0.12168792934249265,
"grad_norm": 0.1525166881836132,
"learning_rate": 1.9978589232386036e-05,
"loss": 0.4668,
"step": 62
},
{
"epoch": 0.12365063788027478,
"grad_norm": 0.16104011998686385,
"learning_rate": 1.9977557920460237e-05,
"loss": 0.5001,
"step": 63
},
{
"epoch": 0.1256133464180569,
"grad_norm": 0.16256381990686963,
"learning_rate": 1.997650237912329e-05,
"loss": 0.4615,
"step": 64
},
{
"epoch": 0.12757605495583907,
"grad_norm": 0.14954423248271237,
"learning_rate": 1.9975422610938463e-05,
"loss": 0.4686,
"step": 65
},
{
"epoch": 0.1295387634936212,
"grad_norm": 0.19293534758098466,
"learning_rate": 1.997431861852785e-05,
"loss": 0.4607,
"step": 66
},
{
"epoch": 0.13150147203140333,
"grad_norm": 0.13976174634255645,
"learning_rate": 1.997319040457238e-05,
"loss": 0.4727,
"step": 67
},
{
"epoch": 0.13346418056918546,
"grad_norm": 0.18195008776452798,
"learning_rate": 1.9972037971811802e-05,
"loss": 0.4681,
"step": 68
},
{
"epoch": 0.13542688910696762,
"grad_norm": 0.13607271529544315,
"learning_rate": 1.9970861323044667e-05,
"loss": 0.4747,
"step": 69
},
{
"epoch": 0.13738959764474976,
"grad_norm": 0.14246714469054536,
"learning_rate": 1.996966046112834e-05,
"loss": 0.457,
"step": 70
},
{
"epoch": 0.1393523061825319,
"grad_norm": 0.15545571681772094,
"learning_rate": 1.9968435388978986e-05,
"loss": 0.4654,
"step": 71
},
{
"epoch": 0.14131501472031405,
"grad_norm": 0.1631020324563178,
"learning_rate": 1.996718610957155e-05,
"loss": 0.4663,
"step": 72
},
{
"epoch": 0.14327772325809618,
"grad_norm": 0.14479401685365606,
"learning_rate": 1.996591262593978e-05,
"loss": 0.4522,
"step": 73
},
{
"epoch": 0.1452404317958783,
"grad_norm": 0.15470054571534417,
"learning_rate": 1.9964614941176194e-05,
"loss": 0.4596,
"step": 74
},
{
"epoch": 0.14720314033366044,
"grad_norm": 0.13412246717171877,
"learning_rate": 1.9963293058432066e-05,
"loss": 0.4465,
"step": 75
},
{
"epoch": 0.1491658488714426,
"grad_norm": 0.14095810281768514,
"learning_rate": 1.9961946980917457e-05,
"loss": 0.4537,
"step": 76
},
{
"epoch": 0.15112855740922473,
"grad_norm": 0.16396930622944708,
"learning_rate": 1.996057671190116e-05,
"loss": 0.453,
"step": 77
},
{
"epoch": 0.15309126594700687,
"grad_norm": 0.5108895960415857,
"learning_rate": 1.995918225471073e-05,
"loss": 0.4819,
"step": 78
},
{
"epoch": 0.155053974484789,
"grad_norm": 0.16436026012077695,
"learning_rate": 1.995776361273245e-05,
"loss": 0.4481,
"step": 79
},
{
"epoch": 0.15701668302257116,
"grad_norm": 0.18055053584795847,
"learning_rate": 1.9956320789411338e-05,
"loss": 0.4561,
"step": 80
},
{
"epoch": 0.1589793915603533,
"grad_norm": 0.17300584614095313,
"learning_rate": 1.9954853788251137e-05,
"loss": 0.4626,
"step": 81
},
{
"epoch": 0.16094210009813542,
"grad_norm": 0.18537091420945745,
"learning_rate": 1.9953362612814294e-05,
"loss": 0.467,
"step": 82
},
{
"epoch": 0.16290480863591755,
"grad_norm": 0.1831953893733552,
"learning_rate": 1.995184726672197e-05,
"loss": 0.4533,
"step": 83
},
{
"epoch": 0.1648675171736997,
"grad_norm": 0.1785397387970331,
"learning_rate": 1.9950307753654016e-05,
"loss": 0.4714,
"step": 84
},
{
"epoch": 0.16683022571148184,
"grad_norm": 0.1696225474699336,
"learning_rate": 1.994874407734897e-05,
"loss": 0.4695,
"step": 85
},
{
"epoch": 0.16879293424926398,
"grad_norm": 0.20269683913377595,
"learning_rate": 1.994715624160405e-05,
"loss": 0.4648,
"step": 86
},
{
"epoch": 0.17075564278704614,
"grad_norm": 0.18739457130939605,
"learning_rate": 1.9945544250275147e-05,
"loss": 0.4608,
"step": 87
},
{
"epoch": 0.17271835132482827,
"grad_norm": 0.21960029028555894,
"learning_rate": 1.99439081072768e-05,
"loss": 0.4576,
"step": 88
},
{
"epoch": 0.1746810598626104,
"grad_norm": 0.1959495489832016,
"learning_rate": 1.9942247816582206e-05,
"loss": 0.4596,
"step": 89
},
{
"epoch": 0.17664376840039253,
"grad_norm": 0.17212969050820215,
"learning_rate": 1.9940563382223196e-05,
"loss": 0.4522,
"step": 90
},
{
"epoch": 0.1786064769381747,
"grad_norm": 0.1663706916020894,
"learning_rate": 1.9938854808290244e-05,
"loss": 0.4504,
"step": 91
},
{
"epoch": 0.18056918547595682,
"grad_norm": 0.31290735737771297,
"learning_rate": 1.9937122098932428e-05,
"loss": 0.4608,
"step": 92
},
{
"epoch": 0.18253189401373895,
"grad_norm": 0.23409310633529254,
"learning_rate": 1.9935365258357446e-05,
"loss": 0.4559,
"step": 93
},
{
"epoch": 0.1844946025515211,
"grad_norm": 0.2596933197806358,
"learning_rate": 1.9933584290831593e-05,
"loss": 0.4698,
"step": 94
},
{
"epoch": 0.18645731108930325,
"grad_norm": 0.19766333203388936,
"learning_rate": 1.9931779200679754e-05,
"loss": 0.443,
"step": 95
},
{
"epoch": 0.18842001962708538,
"grad_norm": 0.18968390210321073,
"learning_rate": 1.9929949992285397e-05,
"loss": 0.4623,
"step": 96
},
{
"epoch": 0.1903827281648675,
"grad_norm": 0.16223204516832543,
"learning_rate": 1.9928096670090552e-05,
"loss": 0.4772,
"step": 97
},
{
"epoch": 0.19234543670264967,
"grad_norm": 0.18715665258083516,
"learning_rate": 1.992621923859581e-05,
"loss": 0.4771,
"step": 98
},
{
"epoch": 0.1943081452404318,
"grad_norm": 0.2670823974648894,
"learning_rate": 1.992431770236031e-05,
"loss": 0.4625,
"step": 99
},
{
"epoch": 0.19627085377821393,
"grad_norm": 0.2389458544722749,
"learning_rate": 1.9922392066001724e-05,
"loss": 0.4496,
"step": 100
},
{
"epoch": 0.19823356231599606,
"grad_norm": 0.18959477739852718,
"learning_rate": 1.9920442334196248e-05,
"loss": 0.4685,
"step": 101
},
{
"epoch": 0.20019627085377822,
"grad_norm": 0.1872065955224367,
"learning_rate": 1.99184685116786e-05,
"loss": 0.4724,
"step": 102
},
{
"epoch": 0.20215897939156036,
"grad_norm": 0.2002235120172999,
"learning_rate": 1.991647060324198e-05,
"loss": 0.456,
"step": 103
},
{
"epoch": 0.2041216879293425,
"grad_norm": 0.2764357846718622,
"learning_rate": 1.9914448613738107e-05,
"loss": 0.4496,
"step": 104
},
{
"epoch": 0.20608439646712462,
"grad_norm": 0.1916898969998294,
"learning_rate": 1.991240254807715e-05,
"loss": 0.4584,
"step": 105
},
{
"epoch": 0.20804710500490678,
"grad_norm": 0.3108126102936448,
"learning_rate": 1.991033241122776e-05,
"loss": 0.4595,
"step": 106
},
{
"epoch": 0.2100098135426889,
"grad_norm": 0.21101958681311,
"learning_rate": 1.990823820821704e-05,
"loss": 0.4911,
"step": 107
},
{
"epoch": 0.21197252208047104,
"grad_norm": 0.42110628942523765,
"learning_rate": 1.9906119944130527e-05,
"loss": 0.4605,
"step": 108
},
{
"epoch": 0.2139352306182532,
"grad_norm": 0.17728266885723962,
"learning_rate": 1.9903977624112204e-05,
"loss": 0.461,
"step": 109
},
{
"epoch": 0.21589793915603533,
"grad_norm": 0.26344485864001554,
"learning_rate": 1.9901811253364458e-05,
"loss": 0.456,
"step": 110
},
{
"epoch": 0.21786064769381747,
"grad_norm": 0.24928437921890148,
"learning_rate": 1.989962083714808e-05,
"loss": 0.4565,
"step": 111
},
{
"epoch": 0.2198233562315996,
"grad_norm": 0.22760146691028113,
"learning_rate": 1.9897406380782262e-05,
"loss": 0.4617,
"step": 112
},
{
"epoch": 0.22178606476938176,
"grad_norm": 0.33505042507064053,
"learning_rate": 1.9895167889644568e-05,
"loss": 0.4619,
"step": 113
},
{
"epoch": 0.2237487733071639,
"grad_norm": 0.23380989191893622,
"learning_rate": 1.989290536917093e-05,
"loss": 0.4682,
"step": 114
},
{
"epoch": 0.22571148184494602,
"grad_norm": 0.16292574235785917,
"learning_rate": 1.9890618824855624e-05,
"loss": 0.4609,
"step": 115
},
{
"epoch": 0.22767419038272815,
"grad_norm": 0.21123922747241927,
"learning_rate": 1.9888308262251286e-05,
"loss": 0.4535,
"step": 116
},
{
"epoch": 0.2296368989205103,
"grad_norm": 0.19216925224306786,
"learning_rate": 1.988597368696886e-05,
"loss": 0.4666,
"step": 117
},
{
"epoch": 0.23159960745829244,
"grad_norm": 0.1978727333868667,
"learning_rate": 1.988361510467761e-05,
"loss": 0.4483,
"step": 118
},
{
"epoch": 0.23356231599607458,
"grad_norm": 0.2024852375366097,
"learning_rate": 1.988123252110509e-05,
"loss": 0.4575,
"step": 119
},
{
"epoch": 0.23552502453385674,
"grad_norm": 0.2275812980861413,
"learning_rate": 1.9878825942037147e-05,
"loss": 0.4733,
"step": 120
},
{
"epoch": 0.23748773307163887,
"grad_norm": 0.16622185526393202,
"learning_rate": 1.98763953733179e-05,
"loss": 0.4638,
"step": 121
},
{
"epoch": 0.239450441609421,
"grad_norm": 0.1599896679961423,
"learning_rate": 1.9873940820849714e-05,
"loss": 0.4676,
"step": 122
},
{
"epoch": 0.24141315014720313,
"grad_norm": 0.17524027525754993,
"learning_rate": 1.9871462290593206e-05,
"loss": 0.4671,
"step": 123
},
{
"epoch": 0.2433758586849853,
"grad_norm": 0.18451386556539923,
"learning_rate": 1.9868959788567213e-05,
"loss": 0.4714,
"step": 124
},
{
"epoch": 0.24533856722276742,
"grad_norm": 0.20332604727761022,
"learning_rate": 1.9866433320848793e-05,
"loss": 0.4395,
"step": 125
},
{
"epoch": 0.24730127576054955,
"grad_norm": 0.13917105010229833,
"learning_rate": 1.9863882893573188e-05,
"loss": 0.4568,
"step": 126
},
{
"epoch": 0.2492639842983317,
"grad_norm": 0.17751758499748868,
"learning_rate": 1.9861308512933846e-05,
"loss": 0.4746,
"step": 127
},
{
"epoch": 0.2512266928361138,
"grad_norm": 0.15877578177692772,
"learning_rate": 1.985871018518236e-05,
"loss": 0.4515,
"step": 128
},
{
"epoch": 0.2512266928361138,
"eval_loss": 0.4546394944190979,
"eval_runtime": 245.4163,
"eval_samples_per_second": 123.68,
"eval_steps_per_second": 3.867,
"step": 128
},
{
"epoch": 0.25318940137389595,
"grad_norm": 0.16746870749832415,
"learning_rate": 1.9856087916628487e-05,
"loss": 0.4564,
"step": 129
},
{
"epoch": 0.25515210991167814,
"grad_norm": 0.32948890713463225,
"learning_rate": 1.9853441713640123e-05,
"loss": 0.4604,
"step": 130
},
{
"epoch": 0.25711481844946027,
"grad_norm": 0.19647029953806538,
"learning_rate": 1.985077158264328e-05,
"loss": 0.4439,
"step": 131
},
{
"epoch": 0.2590775269872424,
"grad_norm": 0.251393507519347,
"learning_rate": 1.9848077530122083e-05,
"loss": 0.4574,
"step": 132
},
{
"epoch": 0.26104023552502453,
"grad_norm": 0.24198770425177948,
"learning_rate": 1.984535956261874e-05,
"loss": 0.4725,
"step": 133
},
{
"epoch": 0.26300294406280667,
"grad_norm": 0.19883583006200609,
"learning_rate": 1.9842617686733546e-05,
"loss": 0.4509,
"step": 134
},
{
"epoch": 0.2649656526005888,
"grad_norm": 0.19609522995459108,
"learning_rate": 1.983985190912484e-05,
"loss": 0.4513,
"step": 135
},
{
"epoch": 0.26692836113837093,
"grad_norm": 0.18132847552230033,
"learning_rate": 1.9837062236509013e-05,
"loss": 0.4614,
"step": 136
},
{
"epoch": 0.2688910696761531,
"grad_norm": 0.2617064903935584,
"learning_rate": 1.9834248675660484e-05,
"loss": 0.4729,
"step": 137
},
{
"epoch": 0.27085377821393525,
"grad_norm": 0.15567291724986626,
"learning_rate": 1.983141123341168e-05,
"loss": 0.4496,
"step": 138
},
{
"epoch": 0.2728164867517174,
"grad_norm": 0.8337178932954963,
"learning_rate": 1.9828549916653013e-05,
"loss": 0.4464,
"step": 139
},
{
"epoch": 0.2747791952894995,
"grad_norm": 0.13791288977874855,
"learning_rate": 1.9825664732332886e-05,
"loss": 0.4484,
"step": 140
},
{
"epoch": 0.27674190382728164,
"grad_norm": 0.15828576470912684,
"learning_rate": 1.9822755687457645e-05,
"loss": 0.4568,
"step": 141
},
{
"epoch": 0.2787046123650638,
"grad_norm": 0.18001770804302206,
"learning_rate": 1.9819822789091597e-05,
"loss": 0.4651,
"step": 142
},
{
"epoch": 0.2806673209028459,
"grad_norm": 0.39836155933646317,
"learning_rate": 1.9816866044356968e-05,
"loss": 0.4481,
"step": 143
},
{
"epoch": 0.2826300294406281,
"grad_norm": 0.15389096751590023,
"learning_rate": 1.981388546043388e-05,
"loss": 0.4604,
"step": 144
},
{
"epoch": 0.2845927379784102,
"grad_norm": 0.23663992095817404,
"learning_rate": 1.981088104456036e-05,
"loss": 0.4587,
"step": 145
},
{
"epoch": 0.28655544651619236,
"grad_norm": 0.18263790394872537,
"learning_rate": 1.9807852804032306e-05,
"loss": 0.4606,
"step": 146
},
{
"epoch": 0.2885181550539745,
"grad_norm": 0.1790389596426215,
"learning_rate": 1.980480074620347e-05,
"loss": 0.4492,
"step": 147
},
{
"epoch": 0.2904808635917566,
"grad_norm": 0.27451162166681453,
"learning_rate": 1.9801724878485438e-05,
"loss": 0.458,
"step": 148
},
{
"epoch": 0.29244357212953875,
"grad_norm": 0.17498754542123285,
"learning_rate": 1.9798625208347627e-05,
"loss": 0.4583,
"step": 149
},
{
"epoch": 0.2944062806673209,
"grad_norm": 0.9166003434108919,
"learning_rate": 1.979550174331724e-05,
"loss": 0.4599,
"step": 150
},
{
"epoch": 0.296368989205103,
"grad_norm": 1.7151984325661747,
"learning_rate": 1.9792354490979275e-05,
"loss": 0.4652,
"step": 151
},
{
"epoch": 0.2983316977428852,
"grad_norm": 3.7084263741544903,
"learning_rate": 1.9789183458976485e-05,
"loss": 0.5222,
"step": 152
},
{
"epoch": 0.30029440628066734,
"grad_norm": 0.23854084460339886,
"learning_rate": 1.9785988655009386e-05,
"loss": 0.4612,
"step": 153
},
{
"epoch": 0.30225711481844947,
"grad_norm": 0.4504871270083281,
"learning_rate": 1.97827700868362e-05,
"loss": 0.4812,
"step": 154
},
{
"epoch": 0.3042198233562316,
"grad_norm": 0.4201391978088661,
"learning_rate": 1.9779527762272877e-05,
"loss": 0.4855,
"step": 155
},
{
"epoch": 0.30618253189401373,
"grad_norm": 0.44160943541930997,
"learning_rate": 1.977626168919305e-05,
"loss": 0.4723,
"step": 156
},
{
"epoch": 0.30814524043179586,
"grad_norm": 0.4231731091405931,
"learning_rate": 1.977297187552801e-05,
"loss": 0.4648,
"step": 157
},
{
"epoch": 0.310107948969578,
"grad_norm": 0.3638067041093047,
"learning_rate": 1.9769658329266718e-05,
"loss": 0.4545,
"step": 158
},
{
"epoch": 0.3120706575073602,
"grad_norm": 0.4239566722192982,
"learning_rate": 1.976632105845576e-05,
"loss": 0.4849,
"step": 159
},
{
"epoch": 0.3140333660451423,
"grad_norm": 0.3461796288026477,
"learning_rate": 1.9762960071199334e-05,
"loss": 0.4794,
"step": 160
},
{
"epoch": 0.31599607458292445,
"grad_norm": 0.24887697538008421,
"learning_rate": 1.9759575375659232e-05,
"loss": 0.4814,
"step": 161
},
{
"epoch": 0.3179587831207066,
"grad_norm": 0.24050744980645108,
"learning_rate": 1.9756166980054812e-05,
"loss": 0.4516,
"step": 162
},
{
"epoch": 0.3199214916584887,
"grad_norm": 0.2345271899376135,
"learning_rate": 1.9752734892663e-05,
"loss": 0.4779,
"step": 163
},
{
"epoch": 0.32188420019627084,
"grad_norm": 0.27248178409002094,
"learning_rate": 1.9749279121818235e-05,
"loss": 0.4677,
"step": 164
},
{
"epoch": 0.323846908734053,
"grad_norm": 0.19021861355008923,
"learning_rate": 1.9745799675912492e-05,
"loss": 0.4491,
"step": 165
},
{
"epoch": 0.3258096172718351,
"grad_norm": 0.20109086134349027,
"learning_rate": 1.9742296563395218e-05,
"loss": 0.4485,
"step": 166
},
{
"epoch": 0.3277723258096173,
"grad_norm": 0.19466230218717548,
"learning_rate": 1.9738769792773338e-05,
"loss": 0.4429,
"step": 167
},
{
"epoch": 0.3297350343473994,
"grad_norm": 0.22618689389367447,
"learning_rate": 1.9735219372611232e-05,
"loss": 0.4776,
"step": 168
},
{
"epoch": 0.33169774288518156,
"grad_norm": 0.2548001164602319,
"learning_rate": 1.9731645311530718e-05,
"loss": 0.4544,
"step": 169
},
{
"epoch": 0.3336604514229637,
"grad_norm": 0.1791237461163614,
"learning_rate": 1.9728047618210995e-05,
"loss": 0.4554,
"step": 170
},
{
"epoch": 0.3356231599607458,
"grad_norm": 0.3948418555429021,
"learning_rate": 1.9724426301388683e-05,
"loss": 0.4638,
"step": 171
},
{
"epoch": 0.33758586849852795,
"grad_norm": 0.17378090038653812,
"learning_rate": 1.9720781369857747e-05,
"loss": 0.4454,
"step": 172
},
{
"epoch": 0.3395485770363101,
"grad_norm": 0.3222426037289685,
"learning_rate": 1.971711283246951e-05,
"loss": 0.4535,
"step": 173
},
{
"epoch": 0.34151128557409227,
"grad_norm": 0.18445752660643416,
"learning_rate": 1.9713420698132614e-05,
"loss": 0.4529,
"step": 174
},
{
"epoch": 0.3434739941118744,
"grad_norm": 0.16185923929968957,
"learning_rate": 1.9709704975813007e-05,
"loss": 0.4368,
"step": 175
},
{
"epoch": 0.34543670264965654,
"grad_norm": 1.1894973875213863,
"learning_rate": 1.970596567453391e-05,
"loss": 0.4544,
"step": 176
},
{
"epoch": 0.34739941118743867,
"grad_norm": 0.16134561772472725,
"learning_rate": 1.9702202803375813e-05,
"loss": 0.4622,
"step": 177
},
{
"epoch": 0.3493621197252208,
"grad_norm": 0.1970049090626823,
"learning_rate": 1.9698416371476434e-05,
"loss": 0.4671,
"step": 178
},
{
"epoch": 0.35132482826300293,
"grad_norm": 0.36331006012853106,
"learning_rate": 1.9694606388030715e-05,
"loss": 0.457,
"step": 179
},
{
"epoch": 0.35328753680078506,
"grad_norm": 0.17936278528599436,
"learning_rate": 1.969077286229078e-05,
"loss": 0.4548,
"step": 180
},
{
"epoch": 0.35525024533856725,
"grad_norm": 1.9159888836797607,
"learning_rate": 1.9686915803565934e-05,
"loss": 0.4645,
"step": 181
},
{
"epoch": 0.3572129538763494,
"grad_norm": 10.74627429291283,
"learning_rate": 1.9683035221222617e-05,
"loss": 0.6072,
"step": 182
},
{
"epoch": 0.3591756624141315,
"grad_norm": 52.13270281331323,
"learning_rate": 1.9679131124684403e-05,
"loss": 1.8264,
"step": 183
},
{
"epoch": 0.36113837095191365,
"grad_norm": 4.418687411757653,
"learning_rate": 1.9675203523431964e-05,
"loss": 0.5769,
"step": 184
},
{
"epoch": 0.3631010794896958,
"grad_norm": 0.5887657411717333,
"learning_rate": 1.9671252427003052e-05,
"loss": 0.5168,
"step": 185
},
{
"epoch": 0.3650637880274779,
"grad_norm": 0.674215963992249,
"learning_rate": 1.9667277844992476e-05,
"loss": 0.5026,
"step": 186
},
{
"epoch": 0.36702649656526004,
"grad_norm": 1.0528021833617283,
"learning_rate": 1.966327978705207e-05,
"loss": 0.5349,
"step": 187
},
{
"epoch": 0.3689892051030422,
"grad_norm": 0.9422062018416699,
"learning_rate": 1.9659258262890683e-05,
"loss": 0.5042,
"step": 188
},
{
"epoch": 0.37095191364082436,
"grad_norm": 0.38055499460304,
"learning_rate": 1.9655213282274153e-05,
"loss": 0.494,
"step": 189
},
{
"epoch": 0.3729146221786065,
"grad_norm": 0.7565736806021138,
"learning_rate": 1.9651144855025265e-05,
"loss": 0.4757,
"step": 190
},
{
"epoch": 0.3748773307163886,
"grad_norm": 0.6604416755715676,
"learning_rate": 1.964705299102376e-05,
"loss": 0.4694,
"step": 191
},
{
"epoch": 0.37684003925417076,
"grad_norm": 0.5198922569733609,
"learning_rate": 1.964293770020628e-05,
"loss": 0.4896,
"step": 192
},
{
"epoch": 0.3788027477919529,
"grad_norm": 0.3341751217641192,
"learning_rate": 1.9638798992566354e-05,
"loss": 0.482,
"step": 193
},
{
"epoch": 0.380765456329735,
"grad_norm": 0.34613355879167534,
"learning_rate": 1.9634636878154393e-05,
"loss": 0.4903,
"step": 194
},
{
"epoch": 0.38272816486751715,
"grad_norm": 0.29136481910346357,
"learning_rate": 1.963045136707763e-05,
"loss": 0.4795,
"step": 195
},
{
"epoch": 0.38469087340529934,
"grad_norm": 0.34118862521510146,
"learning_rate": 1.962624246950012e-05,
"loss": 0.4826,
"step": 196
},
{
"epoch": 0.38665358194308147,
"grad_norm": 0.45727416620673994,
"learning_rate": 1.9622010195642724e-05,
"loss": 0.4653,
"step": 197
},
{
"epoch": 0.3886162904808636,
"grad_norm": 0.312131527349307,
"learning_rate": 1.9617754555783045e-05,
"loss": 0.4761,
"step": 198
},
{
"epoch": 0.39057899901864573,
"grad_norm": 0.26137988821611824,
"learning_rate": 1.9613475560255445e-05,
"loss": 0.4792,
"step": 199
},
{
"epoch": 0.39254170755642787,
"grad_norm": 0.2571733191752843,
"learning_rate": 1.9609173219450998e-05,
"loss": 0.4636,
"step": 200
},
{
"epoch": 0.39450441609421,
"grad_norm": 0.286975647231177,
"learning_rate": 1.9604847543817467e-05,
"loss": 0.4733,
"step": 201
},
{
"epoch": 0.39646712463199213,
"grad_norm": 0.18272338104532077,
"learning_rate": 1.960049854385929e-05,
"loss": 0.437,
"step": 202
},
{
"epoch": 0.39842983316977426,
"grad_norm": 0.22862620377890577,
"learning_rate": 1.9596126230137532e-05,
"loss": 0.4634,
"step": 203
},
{
"epoch": 0.40039254170755645,
"grad_norm": 0.3492075746901584,
"learning_rate": 1.9591730613269878e-05,
"loss": 0.4776,
"step": 204
},
{
"epoch": 0.4023552502453386,
"grad_norm": 0.21584801782728166,
"learning_rate": 1.9587311703930615e-05,
"loss": 0.45,
"step": 205
},
{
"epoch": 0.4043179587831207,
"grad_norm": 0.1775551780611007,
"learning_rate": 1.9582869512850576e-05,
"loss": 0.4588,
"step": 206
},
{
"epoch": 0.40628066732090284,
"grad_norm": 0.19395395115771327,
"learning_rate": 1.9578404050817135e-05,
"loss": 0.4476,
"step": 207
},
{
"epoch": 0.408243375858685,
"grad_norm": 0.1756543157586606,
"learning_rate": 1.957391532867418e-05,
"loss": 0.4541,
"step": 208
},
{
"epoch": 0.4102060843964671,
"grad_norm": 0.2033504241274638,
"learning_rate": 1.956940335732209e-05,
"loss": 0.4517,
"step": 209
},
{
"epoch": 0.41216879293424924,
"grad_norm": 0.17283372188002102,
"learning_rate": 1.956486814771769e-05,
"loss": 0.4388,
"step": 210
},
{
"epoch": 0.4141315014720314,
"grad_norm": 0.3603837587378834,
"learning_rate": 1.956030971087424e-05,
"loss": 0.4589,
"step": 211
},
{
"epoch": 0.41609421000981356,
"grad_norm": 0.14136642540755467,
"learning_rate": 1.955572805786141e-05,
"loss": 0.4458,
"step": 212
},
{
"epoch": 0.4180569185475957,
"grad_norm": 0.5877313778359132,
"learning_rate": 1.9551123199805243e-05,
"loss": 0.4697,
"step": 213
},
{
"epoch": 0.4200196270853778,
"grad_norm": 0.14159168090968688,
"learning_rate": 1.9546495147888134e-05,
"loss": 0.4615,
"step": 214
},
{
"epoch": 0.42198233562315995,
"grad_norm": 0.14811573102512038,
"learning_rate": 1.9541843913348804e-05,
"loss": 0.4434,
"step": 215
},
{
"epoch": 0.4239450441609421,
"grad_norm": 0.1892192591601388,
"learning_rate": 1.953716950748227e-05,
"loss": 0.4502,
"step": 216
},
{
"epoch": 0.4259077526987242,
"grad_norm": 0.18436679871777018,
"learning_rate": 1.9532471941639816e-05,
"loss": 0.4583,
"step": 217
},
{
"epoch": 0.4278704612365064,
"grad_norm": 0.16076460592232203,
"learning_rate": 1.9527751227228964e-05,
"loss": 0.4782,
"step": 218
},
{
"epoch": 0.42983316977428854,
"grad_norm": 0.16974596439970596,
"learning_rate": 1.952300737571346e-05,
"loss": 0.4417,
"step": 219
},
{
"epoch": 0.43179587831207067,
"grad_norm": 0.1489166172695092,
"learning_rate": 1.9518240398613226e-05,
"loss": 0.45,
"step": 220
},
{
"epoch": 0.4337585868498528,
"grad_norm": 0.19974086406341826,
"learning_rate": 1.9513450307504346e-05,
"loss": 0.4457,
"step": 221
},
{
"epoch": 0.43572129538763493,
"grad_norm": 0.14988334374863602,
"learning_rate": 1.9508637114019037e-05,
"loss": 0.4416,
"step": 222
},
{
"epoch": 0.43768400392541706,
"grad_norm": 0.15223238304202846,
"learning_rate": 1.9503800829845613e-05,
"loss": 0.4463,
"step": 223
},
{
"epoch": 0.4396467124631992,
"grad_norm": 0.14747133903175189,
"learning_rate": 1.9498941466728462e-05,
"loss": 0.4408,
"step": 224
},
{
"epoch": 0.44160942100098133,
"grad_norm": 0.1526636764634093,
"learning_rate": 1.9494059036468016e-05,
"loss": 0.4499,
"step": 225
},
{
"epoch": 0.4435721295387635,
"grad_norm": 0.1628546826649206,
"learning_rate": 1.9489153550920726e-05,
"loss": 0.4439,
"step": 226
},
{
"epoch": 0.44553483807654565,
"grad_norm": 0.15901123933171535,
"learning_rate": 1.9484225021999032e-05,
"loss": 0.4412,
"step": 227
},
{
"epoch": 0.4474975466143278,
"grad_norm": 0.1490150024919776,
"learning_rate": 1.947927346167132e-05,
"loss": 0.455,
"step": 228
},
{
"epoch": 0.4494602551521099,
"grad_norm": 0.5516033577713996,
"learning_rate": 1.9474298881961918e-05,
"loss": 0.4566,
"step": 229
},
{
"epoch": 0.45142296368989204,
"grad_norm": 0.13967143480519018,
"learning_rate": 1.946930129495106e-05,
"loss": 0.4628,
"step": 230
},
{
"epoch": 0.4533856722276742,
"grad_norm": 0.1305332529656271,
"learning_rate": 1.9464280712774828e-05,
"loss": 0.4393,
"step": 231
},
{
"epoch": 0.4553483807654563,
"grad_norm": 0.1345554686370529,
"learning_rate": 1.945923714762516e-05,
"loss": 0.4545,
"step": 232
},
{
"epoch": 0.4573110893032385,
"grad_norm": 0.136671297145043,
"learning_rate": 1.9454170611749812e-05,
"loss": 0.4538,
"step": 233
},
{
"epoch": 0.4592737978410206,
"grad_norm": 0.18983035932318434,
"learning_rate": 1.9449081117452304e-05,
"loss": 0.4551,
"step": 234
},
{
"epoch": 0.46123650637880276,
"grad_norm": 0.13254449859971484,
"learning_rate": 1.9443968677091925e-05,
"loss": 0.4364,
"step": 235
},
{
"epoch": 0.4631992149165849,
"grad_norm": 0.12020638993133913,
"learning_rate": 1.9438833303083677e-05,
"loss": 0.4442,
"step": 236
},
{
"epoch": 0.465161923454367,
"grad_norm": 0.13012001270770882,
"learning_rate": 1.9433675007898255e-05,
"loss": 0.4397,
"step": 237
},
{
"epoch": 0.46712463199214915,
"grad_norm": 0.260116572268421,
"learning_rate": 1.9428493804062013e-05,
"loss": 0.4446,
"step": 238
},
{
"epoch": 0.4690873405299313,
"grad_norm": 0.1231982496143958,
"learning_rate": 1.9423289704156945e-05,
"loss": 0.4616,
"step": 239
},
{
"epoch": 0.47105004906771347,
"grad_norm": 0.11787031140608273,
"learning_rate": 1.9418062720820636e-05,
"loss": 0.4438,
"step": 240
},
{
"epoch": 0.4730127576054956,
"grad_norm": 0.13848283459243596,
"learning_rate": 1.9412812866746248e-05,
"loss": 0.4479,
"step": 241
},
{
"epoch": 0.47497546614327774,
"grad_norm": 0.12036969167700294,
"learning_rate": 1.9407540154682473e-05,
"loss": 0.4508,
"step": 242
},
{
"epoch": 0.47693817468105987,
"grad_norm": 0.17542281239029114,
"learning_rate": 1.9402244597433526e-05,
"loss": 0.4393,
"step": 243
},
{
"epoch": 0.478900883218842,
"grad_norm": 0.1137790859381083,
"learning_rate": 1.9396926207859085e-05,
"loss": 0.4444,
"step": 244
},
{
"epoch": 0.48086359175662413,
"grad_norm": 0.18308427216283138,
"learning_rate": 1.939158499887428e-05,
"loss": 0.4272,
"step": 245
},
{
"epoch": 0.48282630029440626,
"grad_norm": 0.13610805029149947,
"learning_rate": 1.9386220983449652e-05,
"loss": 0.4453,
"step": 246
},
{
"epoch": 0.4847890088321884,
"grad_norm": 0.12057983580084324,
"learning_rate": 1.938083417461113e-05,
"loss": 0.4298,
"step": 247
},
{
"epoch": 0.4867517173699706,
"grad_norm": 0.11085757126201562,
"learning_rate": 1.9375424585439994e-05,
"loss": 0.4389,
"step": 248
},
{
"epoch": 0.4887144259077527,
"grad_norm": 0.11914366344102803,
"learning_rate": 1.9369992229072834e-05,
"loss": 0.4427,
"step": 249
},
{
"epoch": 0.49067713444553485,
"grad_norm": 0.13244864999661263,
"learning_rate": 1.9364537118701542e-05,
"loss": 0.4347,
"step": 250
},
{
"epoch": 0.492639842983317,
"grad_norm": 0.19302953420237423,
"learning_rate": 1.935905926757326e-05,
"loss": 0.4499,
"step": 251
},
{
"epoch": 0.4946025515210991,
"grad_norm": 0.11571291217554695,
"learning_rate": 1.935355868899034e-05,
"loss": 0.4375,
"step": 252
},
{
"epoch": 0.49656526005888124,
"grad_norm": 0.11702904926416031,
"learning_rate": 1.9348035396310352e-05,
"loss": 0.4354,
"step": 253
},
{
"epoch": 0.4985279685966634,
"grad_norm": 0.12006868637499236,
"learning_rate": 1.9342489402945997e-05,
"loss": 0.4338,
"step": 254
},
{
"epoch": 0.5004906771344455,
"grad_norm": 0.12190055167354476,
"learning_rate": 1.9336920722365126e-05,
"loss": 0.4434,
"step": 255
},
{
"epoch": 0.5024533856722276,
"grad_norm": 0.12867582275673506,
"learning_rate": 1.9331329368090664e-05,
"loss": 0.435,
"step": 256
},
{
"epoch": 0.5024533856722276,
"eval_loss": 0.44365355372428894,
"eval_runtime": 245.6487,
"eval_samples_per_second": 123.563,
"eval_steps_per_second": 3.863,
"step": 256
},
{
"epoch": 0.5044160942100098,
"grad_norm": 0.1233612138867142,
"learning_rate": 1.932571535370061e-05,
"loss": 0.4417,
"step": 257
},
{
"epoch": 0.5063788027477919,
"grad_norm": 2.5334855914455368,
"learning_rate": 1.932007869282799e-05,
"loss": 0.4645,
"step": 258
},
{
"epoch": 0.5083415112855741,
"grad_norm": 0.15972840390446322,
"learning_rate": 1.9314419399160805e-05,
"loss": 0.4328,
"step": 259
},
{
"epoch": 0.5103042198233563,
"grad_norm": 0.21045316053724383,
"learning_rate": 1.9308737486442045e-05,
"loss": 0.437,
"step": 260
},
{
"epoch": 0.5122669283611384,
"grad_norm": 0.1353556883894671,
"learning_rate": 1.930303296846961e-05,
"loss": 0.4532,
"step": 261
},
{
"epoch": 0.5142296368989205,
"grad_norm": 0.13254413062805587,
"learning_rate": 1.9297305859096305e-05,
"loss": 0.452,
"step": 262
},
{
"epoch": 0.5161923454367027,
"grad_norm": 0.19171081658979544,
"learning_rate": 1.9291556172229784e-05,
"loss": 0.4302,
"step": 263
},
{
"epoch": 0.5181550539744848,
"grad_norm": 0.13134834595730802,
"learning_rate": 1.9285783921832537e-05,
"loss": 0.4405,
"step": 264
},
{
"epoch": 0.5201177625122669,
"grad_norm": 0.1301982247003894,
"learning_rate": 1.9279989121921846e-05,
"loss": 0.4462,
"step": 265
},
{
"epoch": 0.5220804710500491,
"grad_norm": 0.1321466010763688,
"learning_rate": 1.927417178656975e-05,
"loss": 0.4336,
"step": 266
},
{
"epoch": 0.5240431795878312,
"grad_norm": 0.1495881568007904,
"learning_rate": 1.9268331929903013e-05,
"loss": 0.4365,
"step": 267
},
{
"epoch": 0.5260058881256133,
"grad_norm": 0.2612207807396482,
"learning_rate": 1.926246956610309e-05,
"loss": 0.4457,
"step": 268
},
{
"epoch": 0.5279685966633955,
"grad_norm": 0.9154910569331776,
"learning_rate": 1.9256584709406093e-05,
"loss": 0.4553,
"step": 269
},
{
"epoch": 0.5299313052011776,
"grad_norm": 0.1534831664232736,
"learning_rate": 1.9250677374102752e-05,
"loss": 0.4441,
"step": 270
},
{
"epoch": 0.5318940137389597,
"grad_norm": 0.14984085964867322,
"learning_rate": 1.9244747574538387e-05,
"loss": 0.4306,
"step": 271
},
{
"epoch": 0.5338567222767419,
"grad_norm": 0.18760798027966705,
"learning_rate": 1.9238795325112867e-05,
"loss": 0.4538,
"step": 272
},
{
"epoch": 0.535819430814524,
"grad_norm": 0.1580188625628718,
"learning_rate": 1.923282064028059e-05,
"loss": 0.4443,
"step": 273
},
{
"epoch": 0.5377821393523062,
"grad_norm": 0.14444473834290128,
"learning_rate": 1.9226823534550418e-05,
"loss": 0.4443,
"step": 274
},
{
"epoch": 0.5397448478900884,
"grad_norm": 0.7330314247606119,
"learning_rate": 1.9220804022485674e-05,
"loss": 0.4638,
"step": 275
},
{
"epoch": 0.5417075564278705,
"grad_norm": 0.17268497255370896,
"learning_rate": 1.921476211870408e-05,
"loss": 0.4503,
"step": 276
},
{
"epoch": 0.5436702649656526,
"grad_norm": 0.18226707950341117,
"learning_rate": 1.920869783787774e-05,
"loss": 0.4383,
"step": 277
},
{
"epoch": 0.5456329735034348,
"grad_norm": 0.5632696487176844,
"learning_rate": 1.9202611194733107e-05,
"loss": 0.4452,
"step": 278
},
{
"epoch": 0.5475956820412169,
"grad_norm": 0.14836959675348457,
"learning_rate": 1.9196502204050925e-05,
"loss": 0.4314,
"step": 279
},
{
"epoch": 0.549558390578999,
"grad_norm": 0.34506199709023483,
"learning_rate": 1.9190370880666206e-05,
"loss": 0.4544,
"step": 280
},
{
"epoch": 0.5515210991167812,
"grad_norm": 0.16441747745890337,
"learning_rate": 1.9184217239468213e-05,
"loss": 0.4429,
"step": 281
},
{
"epoch": 0.5534838076545633,
"grad_norm": 0.16385435349212213,
"learning_rate": 1.9178041295400383e-05,
"loss": 0.46,
"step": 282
},
{
"epoch": 0.5554465161923454,
"grad_norm": 0.16456135455973658,
"learning_rate": 1.917184306346032e-05,
"loss": 0.4399,
"step": 283
},
{
"epoch": 0.5574092247301276,
"grad_norm": 0.14576272645037303,
"learning_rate": 1.9165622558699763e-05,
"loss": 0.4419,
"step": 284
},
{
"epoch": 0.5593719332679097,
"grad_norm": 0.15456597025956298,
"learning_rate": 1.9159379796224524e-05,
"loss": 0.449,
"step": 285
},
{
"epoch": 0.5613346418056918,
"grad_norm": 0.14297456086426616,
"learning_rate": 1.9153114791194475e-05,
"loss": 0.4371,
"step": 286
},
{
"epoch": 0.563297350343474,
"grad_norm": 0.14012901187447932,
"learning_rate": 1.914682755882349e-05,
"loss": 0.4408,
"step": 287
},
{
"epoch": 0.5652600588812562,
"grad_norm": 0.13060314277737214,
"learning_rate": 1.9140518114379433e-05,
"loss": 0.4437,
"step": 288
},
{
"epoch": 0.5672227674190383,
"grad_norm": 0.3090626051049369,
"learning_rate": 1.91341864731841e-05,
"loss": 0.4542,
"step": 289
},
{
"epoch": 0.5691854759568205,
"grad_norm": 0.2179414551780141,
"learning_rate": 1.912783265061319e-05,
"loss": 0.4408,
"step": 290
},
{
"epoch": 0.5711481844946026,
"grad_norm": 0.620377656428685,
"learning_rate": 1.912145666209627e-05,
"loss": 0.4879,
"step": 291
},
{
"epoch": 0.5731108930323847,
"grad_norm": 0.1689086534993839,
"learning_rate": 1.9115058523116734e-05,
"loss": 0.4462,
"step": 292
},
{
"epoch": 0.5750736015701668,
"grad_norm": 0.14611412522744563,
"learning_rate": 1.910863824921176e-05,
"loss": 0.4255,
"step": 293
},
{
"epoch": 0.577036310107949,
"grad_norm": 0.1882222162638847,
"learning_rate": 1.9102195855972287e-05,
"loss": 0.4425,
"step": 294
},
{
"epoch": 0.5789990186457311,
"grad_norm": 0.1656815361127913,
"learning_rate": 1.909573135904296e-05,
"loss": 0.4283,
"step": 295
},
{
"epoch": 0.5809617271835132,
"grad_norm": 0.13073397914317533,
"learning_rate": 1.908924477412211e-05,
"loss": 0.4558,
"step": 296
},
{
"epoch": 0.5829244357212954,
"grad_norm": 0.1441943832714689,
"learning_rate": 1.90827361169617e-05,
"loss": 0.4427,
"step": 297
},
{
"epoch": 0.5848871442590775,
"grad_norm": 0.12583976569411942,
"learning_rate": 1.9076205403367287e-05,
"loss": 0.4223,
"step": 298
},
{
"epoch": 0.5868498527968596,
"grad_norm": 0.1755315808099223,
"learning_rate": 1.9069652649198004e-05,
"loss": 0.4421,
"step": 299
},
{
"epoch": 0.5888125613346418,
"grad_norm": 0.12236948996858302,
"learning_rate": 1.9063077870366504e-05,
"loss": 0.4403,
"step": 300
},
{
"epoch": 0.5907752698724239,
"grad_norm": 0.13958618892557004,
"learning_rate": 1.905648108283891e-05,
"loss": 0.4304,
"step": 301
},
{
"epoch": 0.592737978410206,
"grad_norm": 0.14279319639440627,
"learning_rate": 1.90498623026348e-05,
"loss": 0.4461,
"step": 302
},
{
"epoch": 0.5947006869479883,
"grad_norm": 0.3909906198333688,
"learning_rate": 1.9043221545827172e-05,
"loss": 0.4845,
"step": 303
},
{
"epoch": 0.5966633954857704,
"grad_norm": 0.37297343958892704,
"learning_rate": 1.903655882854237e-05,
"loss": 0.4494,
"step": 304
},
{
"epoch": 0.5986261040235525,
"grad_norm": 0.8045487515732387,
"learning_rate": 1.9029874166960075e-05,
"loss": 0.4651,
"step": 305
},
{
"epoch": 0.6005888125613347,
"grad_norm": 0.17675678913043202,
"learning_rate": 1.9023167577313267e-05,
"loss": 0.4342,
"step": 306
},
{
"epoch": 0.6025515210991168,
"grad_norm": 0.4614395798632603,
"learning_rate": 1.901643907588816e-05,
"loss": 0.4666,
"step": 307
},
{
"epoch": 0.6045142296368989,
"grad_norm": 0.24112777104201236,
"learning_rate": 1.900968867902419e-05,
"loss": 0.4608,
"step": 308
},
{
"epoch": 0.6064769381746811,
"grad_norm": 0.16289640981921516,
"learning_rate": 1.900291640311396e-05,
"loss": 0.4347,
"step": 309
},
{
"epoch": 0.6084396467124632,
"grad_norm": 0.16490807898196183,
"learning_rate": 1.8996122264603202e-05,
"loss": 0.4286,
"step": 310
},
{
"epoch": 0.6104023552502453,
"grad_norm": 0.31752463155917027,
"learning_rate": 1.8989306279990736e-05,
"loss": 0.4721,
"step": 311
},
{
"epoch": 0.6123650637880275,
"grad_norm": 0.16490476925454267,
"learning_rate": 1.898246846582844e-05,
"loss": 0.4482,
"step": 312
},
{
"epoch": 0.6143277723258096,
"grad_norm": 0.15956184991382258,
"learning_rate": 1.897560883872121e-05,
"loss": 0.4417,
"step": 313
},
{
"epoch": 0.6162904808635917,
"grad_norm": 0.1831073214143379,
"learning_rate": 1.8968727415326885e-05,
"loss": 0.4441,
"step": 314
},
{
"epoch": 0.6182531894013739,
"grad_norm": 0.3402383015575775,
"learning_rate": 1.896182421235626e-05,
"loss": 0.4643,
"step": 315
},
{
"epoch": 0.620215897939156,
"grad_norm": 0.15056250902532822,
"learning_rate": 1.895489924657301e-05,
"loss": 0.4472,
"step": 316
},
{
"epoch": 0.6221786064769381,
"grad_norm": 0.17384419824788128,
"learning_rate": 1.8947952534793663e-05,
"loss": 0.4455,
"step": 317
},
{
"epoch": 0.6241413150147204,
"grad_norm": 0.18283666562150924,
"learning_rate": 1.894098409388754e-05,
"loss": 0.4425,
"step": 318
},
{
"epoch": 0.6261040235525025,
"grad_norm": 0.18363348573791377,
"learning_rate": 1.8933993940776753e-05,
"loss": 0.4347,
"step": 319
},
{
"epoch": 0.6280667320902846,
"grad_norm": 0.2173940703554356,
"learning_rate": 1.8926982092436117e-05,
"loss": 0.4426,
"step": 320
},
{
"epoch": 0.6300294406280668,
"grad_norm": 0.13338358899432057,
"learning_rate": 1.8919948565893144e-05,
"loss": 0.4348,
"step": 321
},
{
"epoch": 0.6319921491658489,
"grad_norm": 0.15291106806170288,
"learning_rate": 1.8912893378227984e-05,
"loss": 0.4441,
"step": 322
},
{
"epoch": 0.633954857703631,
"grad_norm": 0.19828098406270409,
"learning_rate": 1.8905816546573398e-05,
"loss": 0.4498,
"step": 323
},
{
"epoch": 0.6359175662414132,
"grad_norm": 0.12561855756704574,
"learning_rate": 1.8898718088114688e-05,
"loss": 0.4501,
"step": 324
},
{
"epoch": 0.6378802747791953,
"grad_norm": 0.413637068467992,
"learning_rate": 1.8891598020089695e-05,
"loss": 0.4522,
"step": 325
},
{
"epoch": 0.6398429833169774,
"grad_norm": 0.13065710530489547,
"learning_rate": 1.8884456359788725e-05,
"loss": 0.4517,
"step": 326
},
{
"epoch": 0.6418056918547596,
"grad_norm": 0.1255320805445076,
"learning_rate": 1.887729312455452e-05,
"loss": 0.4328,
"step": 327
},
{
"epoch": 0.6437684003925417,
"grad_norm": 0.17052107382561416,
"learning_rate": 1.887010833178222e-05,
"loss": 0.4335,
"step": 328
},
{
"epoch": 0.6457311089303238,
"grad_norm": 0.1463632116200419,
"learning_rate": 1.8862901998919305e-05,
"loss": 0.4489,
"step": 329
},
{
"epoch": 0.647693817468106,
"grad_norm": 0.20054042393653457,
"learning_rate": 1.8855674143465567e-05,
"loss": 0.4415,
"step": 330
},
{
"epoch": 0.6496565260058881,
"grad_norm": 0.12528395471108425,
"learning_rate": 1.8848424782973075e-05,
"loss": 0.4251,
"step": 331
},
{
"epoch": 0.6516192345436702,
"grad_norm": 0.12724271501474496,
"learning_rate": 1.8841153935046098e-05,
"loss": 0.4226,
"step": 332
},
{
"epoch": 0.6535819430814525,
"grad_norm": 0.12815480846276966,
"learning_rate": 1.8833861617341108e-05,
"loss": 0.4389,
"step": 333
},
{
"epoch": 0.6555446516192346,
"grad_norm": 0.11881092687539777,
"learning_rate": 1.8826547847566692e-05,
"loss": 0.4301,
"step": 334
},
{
"epoch": 0.6575073601570167,
"grad_norm": 0.11279466332160643,
"learning_rate": 1.881921264348355e-05,
"loss": 0.4441,
"step": 335
},
{
"epoch": 0.6594700686947988,
"grad_norm": 0.12334758226552325,
"learning_rate": 1.8811856022904423e-05,
"loss": 0.4333,
"step": 336
},
{
"epoch": 0.661432777232581,
"grad_norm": 0.1268483867046061,
"learning_rate": 1.8804478003694066e-05,
"loss": 0.4317,
"step": 337
},
{
"epoch": 0.6633954857703631,
"grad_norm": 0.12443337272902948,
"learning_rate": 1.8797078603769184e-05,
"loss": 0.4282,
"step": 338
},
{
"epoch": 0.6653581943081452,
"grad_norm": 0.12713474397441266,
"learning_rate": 1.878965784109842e-05,
"loss": 0.4482,
"step": 339
},
{
"epoch": 0.6673209028459274,
"grad_norm": 0.12143777103722954,
"learning_rate": 1.8782215733702286e-05,
"loss": 0.4342,
"step": 340
},
{
"epoch": 0.6692836113837095,
"grad_norm": 0.23691651644437595,
"learning_rate": 1.877475229965313e-05,
"loss": 0.4467,
"step": 341
},
{
"epoch": 0.6712463199214916,
"grad_norm": 0.13966524868065613,
"learning_rate": 1.876726755707508e-05,
"loss": 0.4119,
"step": 342
},
{
"epoch": 0.6732090284592738,
"grad_norm": 0.1324720435854753,
"learning_rate": 1.875976152414402e-05,
"loss": 0.4327,
"step": 343
},
{
"epoch": 0.6751717369970559,
"grad_norm": 0.1305963044017511,
"learning_rate": 1.8752234219087538e-05,
"loss": 0.4345,
"step": 344
},
{
"epoch": 0.677134445534838,
"grad_norm": 0.1256369439378997,
"learning_rate": 1.8744685660184868e-05,
"loss": 0.4373,
"step": 345
},
{
"epoch": 0.6790971540726202,
"grad_norm": 0.1306844371007657,
"learning_rate": 1.8737115865766865e-05,
"loss": 0.4326,
"step": 346
},
{
"epoch": 0.6810598626104023,
"grad_norm": 0.122862696340445,
"learning_rate": 1.8729524854215942e-05,
"loss": 0.4166,
"step": 347
},
{
"epoch": 0.6830225711481845,
"grad_norm": 0.13535420265000028,
"learning_rate": 1.8721912643966055e-05,
"loss": 0.4237,
"step": 348
},
{
"epoch": 0.6849852796859667,
"grad_norm": 0.11284294465947456,
"learning_rate": 1.8714279253502616e-05,
"loss": 0.4259,
"step": 349
},
{
"epoch": 0.6869479882237488,
"grad_norm": 0.1193357989589537,
"learning_rate": 1.8706624701362485e-05,
"loss": 0.4467,
"step": 350
},
{
"epoch": 0.6889106967615309,
"grad_norm": 0.1131129128247003,
"learning_rate": 1.8698949006133903e-05,
"loss": 0.4366,
"step": 351
},
{
"epoch": 0.6908734052993131,
"grad_norm": 0.111747855862703,
"learning_rate": 1.8691252186456465e-05,
"loss": 0.4433,
"step": 352
},
{
"epoch": 0.6928361138370952,
"grad_norm": 0.10862949730132815,
"learning_rate": 1.8683534261021058e-05,
"loss": 0.4299,
"step": 353
},
{
"epoch": 0.6947988223748773,
"grad_norm": 0.12299814873161867,
"learning_rate": 1.8675795248569816e-05,
"loss": 0.4366,
"step": 354
},
{
"epoch": 0.6967615309126595,
"grad_norm": 0.11664167882155978,
"learning_rate": 1.866803516789609e-05,
"loss": 0.4379,
"step": 355
},
{
"epoch": 0.6987242394504416,
"grad_norm": 0.12925426587807418,
"learning_rate": 1.866025403784439e-05,
"loss": 0.4333,
"step": 356
},
{
"epoch": 0.7006869479882237,
"grad_norm": 0.10893935119976447,
"learning_rate": 1.8652451877310337e-05,
"loss": 0.4293,
"step": 357
},
{
"epoch": 0.7026496565260059,
"grad_norm": 0.1227760342543677,
"learning_rate": 1.8644628705240636e-05,
"loss": 0.4255,
"step": 358
},
{
"epoch": 0.704612365063788,
"grad_norm": 0.11162933142481653,
"learning_rate": 1.8636784540633002e-05,
"loss": 0.4277,
"step": 359
},
{
"epoch": 0.7065750736015701,
"grad_norm": 0.11784511277232618,
"learning_rate": 1.862891940253613e-05,
"loss": 0.43,
"step": 360
},
{
"epoch": 0.7085377821393523,
"grad_norm": 0.19349027686476353,
"learning_rate": 1.8621033310049658e-05,
"loss": 0.4388,
"step": 361
},
{
"epoch": 0.7105004906771345,
"grad_norm": 0.11461754333230102,
"learning_rate": 1.8613126282324092e-05,
"loss": 0.4459,
"step": 362
},
{
"epoch": 0.7124631992149166,
"grad_norm": 0.10611440044540593,
"learning_rate": 1.860519833856079e-05,
"loss": 0.4286,
"step": 363
},
{
"epoch": 0.7144259077526988,
"grad_norm": 1.3919982408500475,
"learning_rate": 1.8597249498011906e-05,
"loss": 0.4271,
"step": 364
},
{
"epoch": 0.7163886162904809,
"grad_norm": 0.23104503037861346,
"learning_rate": 1.858927977998032e-05,
"loss": 0.4489,
"step": 365
},
{
"epoch": 0.718351324828263,
"grad_norm": 0.13423371418728916,
"learning_rate": 1.858128920381963e-05,
"loss": 0.4432,
"step": 366
},
{
"epoch": 0.7203140333660452,
"grad_norm": 0.14035310994358413,
"learning_rate": 1.8573277788934084e-05,
"loss": 0.4436,
"step": 367
},
{
"epoch": 0.7222767419038273,
"grad_norm": 0.12538177756048358,
"learning_rate": 1.8565245554778516e-05,
"loss": 0.4152,
"step": 368
},
{
"epoch": 0.7242394504416094,
"grad_norm": 0.1393704340215518,
"learning_rate": 1.855719252085834e-05,
"loss": 0.4599,
"step": 369
},
{
"epoch": 0.7262021589793916,
"grad_norm": 0.15221198393758803,
"learning_rate": 1.854911870672947e-05,
"loss": 0.4346,
"step": 370
},
{
"epoch": 0.7281648675171737,
"grad_norm": 0.13205252608835663,
"learning_rate": 1.8541024131998277e-05,
"loss": 0.433,
"step": 371
},
{
"epoch": 0.7301275760549558,
"grad_norm": 0.1824152481964556,
"learning_rate": 1.8532908816321557e-05,
"loss": 0.4258,
"step": 372
},
{
"epoch": 0.732090284592738,
"grad_norm": 0.11473500921109103,
"learning_rate": 1.852477277940647e-05,
"loss": 0.4364,
"step": 373
},
{
"epoch": 0.7340529931305201,
"grad_norm": 0.17550775961163126,
"learning_rate": 1.8516616041010495e-05,
"loss": 0.4311,
"step": 374
},
{
"epoch": 0.7360157016683022,
"grad_norm": 0.11491081744632524,
"learning_rate": 1.850843862094138e-05,
"loss": 0.4299,
"step": 375
},
{
"epoch": 0.7379784102060843,
"grad_norm": 0.12893078747082584,
"learning_rate": 1.8500240539057093e-05,
"loss": 0.4375,
"step": 376
},
{
"epoch": 0.7399411187438666,
"grad_norm": 0.1386244465195304,
"learning_rate": 1.849202181526579e-05,
"loss": 0.4411,
"step": 377
},
{
"epoch": 0.7419038272816487,
"grad_norm": 0.1363795250121767,
"learning_rate": 1.848378246952574e-05,
"loss": 0.4294,
"step": 378
},
{
"epoch": 0.7438665358194309,
"grad_norm": 0.11138728492984143,
"learning_rate": 1.8475522521845296e-05,
"loss": 0.4403,
"step": 379
},
{
"epoch": 0.745829244357213,
"grad_norm": 0.1238513979749556,
"learning_rate": 1.8467241992282842e-05,
"loss": 0.4267,
"step": 380
},
{
"epoch": 0.7477919528949951,
"grad_norm": 0.12373568951195847,
"learning_rate": 1.845894090094674e-05,
"loss": 0.4312,
"step": 381
},
{
"epoch": 0.7497546614327772,
"grad_norm": 0.19653148756122196,
"learning_rate": 1.8450619267995283e-05,
"loss": 0.4343,
"step": 382
},
{
"epoch": 0.7517173699705594,
"grad_norm": 0.1394570505619246,
"learning_rate": 1.8442277113636654e-05,
"loss": 0.4373,
"step": 383
},
{
"epoch": 0.7536800785083415,
"grad_norm": 0.12251934848515013,
"learning_rate": 1.843391445812886e-05,
"loss": 0.4267,
"step": 384
},
{
"epoch": 0.7536800785083415,
"eval_loss": 0.43717867136001587,
"eval_runtime": 245.9442,
"eval_samples_per_second": 123.414,
"eval_steps_per_second": 3.859,
"step": 384
},
{
"epoch": 0.7556427870461236,
"grad_norm": 0.11928267779998,
"learning_rate": 1.84255313217797e-05,
"loss": 0.4411,
"step": 385
},
{
"epoch": 0.7576054955839058,
"grad_norm": 0.10973043663025576,
"learning_rate": 1.84171277249467e-05,
"loss": 0.4266,
"step": 386
},
{
"epoch": 0.7595682041216879,
"grad_norm": 0.6751166131440709,
"learning_rate": 1.8408703688037088e-05,
"loss": 0.4442,
"step": 387
},
{
"epoch": 0.76153091265947,
"grad_norm": 0.11525859230534183,
"learning_rate": 1.8400259231507716e-05,
"loss": 0.4257,
"step": 388
},
{
"epoch": 0.7634936211972522,
"grad_norm": 0.14081912497481,
"learning_rate": 1.8391794375865025e-05,
"loss": 0.4559,
"step": 389
},
{
"epoch": 0.7654563297350343,
"grad_norm": 0.1430446337142817,
"learning_rate": 1.8383309141664992e-05,
"loss": 0.4348,
"step": 390
},
{
"epoch": 0.7674190382728164,
"grad_norm": 0.13701747456448043,
"learning_rate": 1.837480354951308e-05,
"loss": 0.4355,
"step": 391
},
{
"epoch": 0.7693817468105987,
"grad_norm": 0.15468002573828674,
"learning_rate": 1.83662776200642e-05,
"loss": 0.4247,
"step": 392
},
{
"epoch": 0.7713444553483808,
"grad_norm": 0.1359423923850112,
"learning_rate": 1.8357731374022635e-05,
"loss": 0.4378,
"step": 393
},
{
"epoch": 0.7733071638861629,
"grad_norm": 0.12156559871467941,
"learning_rate": 1.8349164832142015e-05,
"loss": 0.433,
"step": 394
},
{
"epoch": 0.7752698724239451,
"grad_norm": 1.1029116856223753,
"learning_rate": 1.834057801522525e-05,
"loss": 0.456,
"step": 395
},
{
"epoch": 0.7772325809617272,
"grad_norm": 0.1498532799383752,
"learning_rate": 1.833197094412449e-05,
"loss": 0.4416,
"step": 396
},
{
"epoch": 0.7791952894995093,
"grad_norm": 0.4425396937410584,
"learning_rate": 1.832334363974107e-05,
"loss": 0.4358,
"step": 397
},
{
"epoch": 0.7811579980372915,
"grad_norm": 0.14741438537186383,
"learning_rate": 1.8314696123025456e-05,
"loss": 0.4432,
"step": 398
},
{
"epoch": 0.7831207065750736,
"grad_norm": 1.3635940422702577,
"learning_rate": 1.8306028414977196e-05,
"loss": 0.4473,
"step": 399
},
{
"epoch": 0.7850834151128557,
"grad_norm": 0.31067568125771744,
"learning_rate": 1.8297340536644877e-05,
"loss": 0.4341,
"step": 400
},
{
"epoch": 0.7870461236506379,
"grad_norm": 0.23052926471546017,
"learning_rate": 1.8288632509126064e-05,
"loss": 0.4469,
"step": 401
},
{
"epoch": 0.78900883218842,
"grad_norm": 1.4150298223131084,
"learning_rate": 1.827990435356725e-05,
"loss": 0.4573,
"step": 402
},
{
"epoch": 0.7909715407262021,
"grad_norm": 1.2166261643274474,
"learning_rate": 1.8271156091163813e-05,
"loss": 0.4623,
"step": 403
},
{
"epoch": 0.7929342492639843,
"grad_norm": 8.194510529799096,
"learning_rate": 1.826238774315995e-05,
"loss": 0.9405,
"step": 404
},
{
"epoch": 0.7948969578017664,
"grad_norm": 1.780892469368783,
"learning_rate": 1.8253599330848638e-05,
"loss": 0.6091,
"step": 405
},
{
"epoch": 0.7968596663395485,
"grad_norm": 7.535603719678218,
"learning_rate": 1.8244790875571582e-05,
"loss": 0.9524,
"step": 406
},
{
"epoch": 0.7988223748773308,
"grad_norm": 1.7773136251051938,
"learning_rate": 1.823596239871915e-05,
"loss": 0.6362,
"step": 407
},
{
"epoch": 0.8007850834151129,
"grad_norm": 0.9044661668135322,
"learning_rate": 1.8227113921730336e-05,
"loss": 0.5439,
"step": 408
},
{
"epoch": 0.802747791952895,
"grad_norm": 2.7139783131315376,
"learning_rate": 1.8218245466092704e-05,
"loss": 0.5663,
"step": 409
},
{
"epoch": 0.8047105004906772,
"grad_norm": 0.8834413599502067,
"learning_rate": 1.8209357053342325e-05,
"loss": 0.5256,
"step": 410
},
{
"epoch": 0.8066732090284593,
"grad_norm": 1.1467415784558643,
"learning_rate": 1.8200448705063748e-05,
"loss": 0.5348,
"step": 411
},
{
"epoch": 0.8086359175662414,
"grad_norm": 0.515184372706671,
"learning_rate": 1.819152044288992e-05,
"loss": 0.5062,
"step": 412
},
{
"epoch": 0.8105986261040236,
"grad_norm": 0.6461620091644302,
"learning_rate": 1.8182572288502154e-05,
"loss": 0.5077,
"step": 413
},
{
"epoch": 0.8125613346418057,
"grad_norm": 0.4908151447534473,
"learning_rate": 1.8173604263630066e-05,
"loss": 0.4806,
"step": 414
},
{
"epoch": 0.8145240431795878,
"grad_norm": 0.3802375716765609,
"learning_rate": 1.8164616390051523e-05,
"loss": 0.4802,
"step": 415
},
{
"epoch": 0.81648675171737,
"grad_norm": 0.3805715926219177,
"learning_rate": 1.8155608689592604e-05,
"loss": 0.4814,
"step": 416
},
{
"epoch": 0.8184494602551521,
"grad_norm": 0.3945834169579088,
"learning_rate": 1.814658118412752e-05,
"loss": 0.4724,
"step": 417
},
{
"epoch": 0.8204121687929342,
"grad_norm": 0.2883495416963042,
"learning_rate": 1.8137533895578585e-05,
"loss": 0.4645,
"step": 418
},
{
"epoch": 0.8223748773307163,
"grad_norm": 0.2790010413679452,
"learning_rate": 1.8128466845916156e-05,
"loss": 0.4617,
"step": 419
},
{
"epoch": 0.8243375858684985,
"grad_norm": 0.2786295258327648,
"learning_rate": 1.811938005715857e-05,
"loss": 0.4791,
"step": 420
},
{
"epoch": 0.8263002944062807,
"grad_norm": 0.23317317430153287,
"learning_rate": 1.81102735513721e-05,
"loss": 0.4434,
"step": 421
},
{
"epoch": 0.8282630029440629,
"grad_norm": 0.2451241340574092,
"learning_rate": 1.8101147350670905e-05,
"loss": 0.4528,
"step": 422
},
{
"epoch": 0.830225711481845,
"grad_norm": 0.22624335497909545,
"learning_rate": 1.8092001477216975e-05,
"loss": 0.4644,
"step": 423
},
{
"epoch": 0.8321884200196271,
"grad_norm": 0.23107219922308564,
"learning_rate": 1.8082835953220055e-05,
"loss": 0.4352,
"step": 424
},
{
"epoch": 0.8341511285574092,
"grad_norm": 0.21335476872866976,
"learning_rate": 1.8073650800937627e-05,
"loss": 0.4515,
"step": 425
},
{
"epoch": 0.8361138370951914,
"grad_norm": 0.2076816094768591,
"learning_rate": 1.806444604267483e-05,
"loss": 0.4544,
"step": 426
},
{
"epoch": 0.8380765456329735,
"grad_norm": 0.19718486185623377,
"learning_rate": 1.805522170078441e-05,
"loss": 0.4428,
"step": 427
},
{
"epoch": 0.8400392541707556,
"grad_norm": 0.227660993052998,
"learning_rate": 1.8045977797666685e-05,
"loss": 0.4443,
"step": 428
},
{
"epoch": 0.8420019627085378,
"grad_norm": 0.21176408168289124,
"learning_rate": 1.803671435576946e-05,
"loss": 0.4617,
"step": 429
},
{
"epoch": 0.8439646712463199,
"grad_norm": 0.16842294987099302,
"learning_rate": 1.8027431397587993e-05,
"loss": 0.462,
"step": 430
},
{
"epoch": 0.845927379784102,
"grad_norm": 0.2398849119706587,
"learning_rate": 1.8018128945664936e-05,
"loss": 0.4408,
"step": 431
},
{
"epoch": 0.8478900883218842,
"grad_norm": 0.493692443958903,
"learning_rate": 1.8008807022590283e-05,
"loss": 0.4592,
"step": 432
},
{
"epoch": 0.8498527968596663,
"grad_norm": 0.544249601841786,
"learning_rate": 1.7999465651001297e-05,
"loss": 0.4476,
"step": 433
},
{
"epoch": 0.8518155053974484,
"grad_norm": 0.13924680999659736,
"learning_rate": 1.7990104853582494e-05,
"loss": 0.4281,
"step": 434
},
{
"epoch": 0.8537782139352306,
"grad_norm": 0.173740825747816,
"learning_rate": 1.7980724653065538e-05,
"loss": 0.4269,
"step": 435
},
{
"epoch": 0.8557409224730128,
"grad_norm": 0.15075725057898212,
"learning_rate": 1.7971325072229227e-05,
"loss": 0.4276,
"step": 436
},
{
"epoch": 0.8577036310107949,
"grad_norm": 0.1448476326691961,
"learning_rate": 1.7961906133899417e-05,
"loss": 0.4259,
"step": 437
},
{
"epoch": 0.8596663395485771,
"grad_norm": 0.15099110760456436,
"learning_rate": 1.7952467860948975e-05,
"loss": 0.4385,
"step": 438
},
{
"epoch": 0.8616290480863592,
"grad_norm": 0.2908525767466193,
"learning_rate": 1.7943010276297717e-05,
"loss": 0.4532,
"step": 439
},
{
"epoch": 0.8635917566241413,
"grad_norm": 0.19591315110166763,
"learning_rate": 1.7933533402912354e-05,
"loss": 0.4443,
"step": 440
},
{
"epoch": 0.8655544651619235,
"grad_norm": 0.1427892817158057,
"learning_rate": 1.792403726380644e-05,
"loss": 0.4497,
"step": 441
},
{
"epoch": 0.8675171736997056,
"grad_norm": 0.1557386790206549,
"learning_rate": 1.791452188204031e-05,
"loss": 0.4455,
"step": 442
},
{
"epoch": 0.8694798822374877,
"grad_norm": 0.13076931277969595,
"learning_rate": 1.7904987280721037e-05,
"loss": 0.4273,
"step": 443
},
{
"epoch": 0.8714425907752699,
"grad_norm": 0.17697976434079088,
"learning_rate": 1.7895433483002356e-05,
"loss": 0.4411,
"step": 444
},
{
"epoch": 0.873405299313052,
"grad_norm": 0.12603127769001743,
"learning_rate": 1.7885860512084622e-05,
"loss": 0.4254,
"step": 445
},
{
"epoch": 0.8753680078508341,
"grad_norm": 0.7782445781966569,
"learning_rate": 1.7876268391214756e-05,
"loss": 0.4558,
"step": 446
},
{
"epoch": 0.8773307163886163,
"grad_norm": 0.12647292734359114,
"learning_rate": 1.786665714368617e-05,
"loss": 0.4407,
"step": 447
},
{
"epoch": 0.8792934249263984,
"grad_norm": 0.20575076444088897,
"learning_rate": 1.785702679283874e-05,
"loss": 0.438,
"step": 448
},
{
"epoch": 0.8812561334641805,
"grad_norm": 0.19417352083348613,
"learning_rate": 1.7847377362058712e-05,
"loss": 0.4484,
"step": 449
},
{
"epoch": 0.8832188420019627,
"grad_norm": 0.12756724963135646,
"learning_rate": 1.7837708874778683e-05,
"loss": 0.4348,
"step": 450
},
{
"epoch": 0.8851815505397449,
"grad_norm": 0.11799761970570792,
"learning_rate": 1.7828021354477515e-05,
"loss": 0.437,
"step": 451
},
{
"epoch": 0.887144259077527,
"grad_norm": 0.1251248189106141,
"learning_rate": 1.78183148246803e-05,
"loss": 0.4302,
"step": 452
},
{
"epoch": 0.8891069676153092,
"grad_norm": 0.11474241608076738,
"learning_rate": 1.7808589308958284e-05,
"loss": 0.4343,
"step": 453
},
{
"epoch": 0.8910696761530913,
"grad_norm": 0.12586354241305017,
"learning_rate": 1.7798844830928818e-05,
"loss": 0.4348,
"step": 454
},
{
"epoch": 0.8930323846908734,
"grad_norm": 0.19899892032690525,
"learning_rate": 1.778908141425531e-05,
"loss": 0.4416,
"step": 455
},
{
"epoch": 0.8949950932286556,
"grad_norm": 0.21889266119451378,
"learning_rate": 1.777929908264715e-05,
"loss": 0.431,
"step": 456
},
{
"epoch": 0.8969578017664377,
"grad_norm": 0.1247721576190928,
"learning_rate": 1.7769497859859664e-05,
"loss": 0.4268,
"step": 457
},
{
"epoch": 0.8989205103042198,
"grad_norm": 0.1848738794336329,
"learning_rate": 1.775967776969405e-05,
"loss": 0.4278,
"step": 458
},
{
"epoch": 0.900883218842002,
"grad_norm": 0.14033082433848945,
"learning_rate": 1.774983883599733e-05,
"loss": 0.4202,
"step": 459
},
{
"epoch": 0.9028459273797841,
"grad_norm": 0.1730390019850834,
"learning_rate": 1.7739981082662275e-05,
"loss": 0.4298,
"step": 460
},
{
"epoch": 0.9048086359175662,
"grad_norm": 0.11693891835416576,
"learning_rate": 1.773010453362737e-05,
"loss": 0.4448,
"step": 461
},
{
"epoch": 0.9067713444553483,
"grad_norm": 0.12196405283755252,
"learning_rate": 1.772020921287674e-05,
"loss": 0.4378,
"step": 462
},
{
"epoch": 0.9087340529931305,
"grad_norm": 0.11586861714009496,
"learning_rate": 1.771029514444008e-05,
"loss": 0.4471,
"step": 463
},
{
"epoch": 0.9106967615309126,
"grad_norm": 0.12574469171427458,
"learning_rate": 1.7700362352392632e-05,
"loss": 0.4385,
"step": 464
},
{
"epoch": 0.9126594700686947,
"grad_norm": 0.12237397310589682,
"learning_rate": 1.7690410860855095e-05,
"loss": 0.4309,
"step": 465
},
{
"epoch": 0.914622178606477,
"grad_norm": 0.11990527876395565,
"learning_rate": 1.7680440693993586e-05,
"loss": 0.4336,
"step": 466
},
{
"epoch": 0.9165848871442591,
"grad_norm": 0.11451672776029655,
"learning_rate": 1.7670451876019562e-05,
"loss": 0.4298,
"step": 467
},
{
"epoch": 0.9185475956820413,
"grad_norm": 0.10316051482791179,
"learning_rate": 1.766044443118978e-05,
"loss": 0.4331,
"step": 468
},
{
"epoch": 0.9205103042198234,
"grad_norm": 0.12496652874488692,
"learning_rate": 1.7650418383806233e-05,
"loss": 0.4354,
"step": 469
},
{
"epoch": 0.9224730127576055,
"grad_norm": 0.11863765454148498,
"learning_rate": 1.7640373758216075e-05,
"loss": 0.4352,
"step": 470
},
{
"epoch": 0.9244357212953876,
"grad_norm": 0.11413588741219476,
"learning_rate": 1.763031057881159e-05,
"loss": 0.4302,
"step": 471
},
{
"epoch": 0.9263984298331698,
"grad_norm": 0.11625749644832431,
"learning_rate": 1.762022887003011e-05,
"loss": 0.4336,
"step": 472
},
{
"epoch": 0.9283611383709519,
"grad_norm": 0.11492501049851209,
"learning_rate": 1.761012865635396e-05,
"loss": 0.422,
"step": 473
},
{
"epoch": 0.930323846908734,
"grad_norm": 0.12563889502937342,
"learning_rate": 1.7600009962310417e-05,
"loss": 0.4348,
"step": 474
},
{
"epoch": 0.9322865554465162,
"grad_norm": 0.12443828665681882,
"learning_rate": 1.758987281247162e-05,
"loss": 0.4242,
"step": 475
},
{
"epoch": 0.9342492639842983,
"grad_norm": 0.11644777576344677,
"learning_rate": 1.757971723145453e-05,
"loss": 0.4194,
"step": 476
},
{
"epoch": 0.9362119725220804,
"grad_norm": 0.10793194810684831,
"learning_rate": 1.7569543243920873e-05,
"loss": 0.4345,
"step": 477
},
{
"epoch": 0.9381746810598626,
"grad_norm": 0.11594061352654564,
"learning_rate": 1.7559350874577066e-05,
"loss": 0.4314,
"step": 478
},
{
"epoch": 0.9401373895976447,
"grad_norm": 0.10332500150437529,
"learning_rate": 1.754914014817416e-05,
"loss": 0.4296,
"step": 479
},
{
"epoch": 0.9421000981354269,
"grad_norm": 0.11644279502007197,
"learning_rate": 1.75389110895078e-05,
"loss": 0.429,
"step": 480
},
{
"epoch": 0.9440628066732091,
"grad_norm": 0.10948538558656842,
"learning_rate": 1.7528663723418137e-05,
"loss": 0.4173,
"step": 481
},
{
"epoch": 0.9460255152109912,
"grad_norm": 0.0994509640223274,
"learning_rate": 1.7518398074789776e-05,
"loss": 0.4205,
"step": 482
},
{
"epoch": 0.9479882237487733,
"grad_norm": 0.9276868697631023,
"learning_rate": 1.750811416855173e-05,
"loss": 0.4673,
"step": 483
},
{
"epoch": 0.9499509322865555,
"grad_norm": 0.1374275322216112,
"learning_rate": 1.7497812029677344e-05,
"loss": 0.4242,
"step": 484
},
{
"epoch": 0.9519136408243376,
"grad_norm": 0.38132564392520824,
"learning_rate": 1.7487491683184236e-05,
"loss": 0.4371,
"step": 485
},
{
"epoch": 0.9538763493621197,
"grad_norm": 1.7151988276134096,
"learning_rate": 1.7477153154134244e-05,
"loss": 0.4803,
"step": 486
},
{
"epoch": 0.9558390578999019,
"grad_norm": 0.32545933630412127,
"learning_rate": 1.7466796467633357e-05,
"loss": 0.4369,
"step": 487
},
{
"epoch": 0.957801766437684,
"grad_norm": 0.1866079133255275,
"learning_rate": 1.7456421648831658e-05,
"loss": 0.4334,
"step": 488
},
{
"epoch": 0.9597644749754661,
"grad_norm": 0.3272314043287591,
"learning_rate": 1.7446028722923266e-05,
"loss": 0.4443,
"step": 489
},
{
"epoch": 0.9617271835132483,
"grad_norm": 0.16996374422654656,
"learning_rate": 1.743561771514626e-05,
"loss": 0.4366,
"step": 490
},
{
"epoch": 0.9636898920510304,
"grad_norm": 0.15827075993181033,
"learning_rate": 1.7425188650782648e-05,
"loss": 0.436,
"step": 491
},
{
"epoch": 0.9656526005888125,
"grad_norm": 0.43165470913413256,
"learning_rate": 1.741474155515827e-05,
"loss": 0.4433,
"step": 492
},
{
"epoch": 0.9676153091265947,
"grad_norm": 1.745564670186826,
"learning_rate": 1.7404276453642755e-05,
"loss": 0.546,
"step": 493
},
{
"epoch": 0.9695780176643768,
"grad_norm": 0.8077780518520157,
"learning_rate": 1.739379337164946e-05,
"loss": 0.4497,
"step": 494
},
{
"epoch": 0.971540726202159,
"grad_norm": 0.474709975296865,
"learning_rate": 1.738329233463542e-05,
"loss": 0.4496,
"step": 495
},
{
"epoch": 0.9735034347399412,
"grad_norm": 0.26095766576170426,
"learning_rate": 1.737277336810124e-05,
"loss": 0.4544,
"step": 496
},
{
"epoch": 0.9754661432777233,
"grad_norm": 0.2912691308833845,
"learning_rate": 1.7362236497591097e-05,
"loss": 0.4423,
"step": 497
},
{
"epoch": 0.9774288518155054,
"grad_norm": 0.24976229543285455,
"learning_rate": 1.7351681748692622e-05,
"loss": 0.4398,
"step": 498
},
{
"epoch": 0.9793915603532876,
"grad_norm": 0.20231287847225168,
"learning_rate": 1.7341109147036873e-05,
"loss": 0.4467,
"step": 499
},
{
"epoch": 0.9813542688910697,
"grad_norm": 0.2412935338809819,
"learning_rate": 1.7330518718298263e-05,
"loss": 0.4357,
"step": 500
},
{
"epoch": 0.9833169774288518,
"grad_norm": 0.18003467387447356,
"learning_rate": 1.7319910488194494e-05,
"loss": 0.4318,
"step": 501
},
{
"epoch": 0.985279685966634,
"grad_norm": 0.20072691034722498,
"learning_rate": 1.7309284482486494e-05,
"loss": 0.4458,
"step": 502
},
{
"epoch": 0.9872423945044161,
"grad_norm": 0.19649513698293491,
"learning_rate": 1.7298640726978357e-05,
"loss": 0.4514,
"step": 503
},
{
"epoch": 0.9892051030421982,
"grad_norm": 0.2071441349062203,
"learning_rate": 1.7287979247517285e-05,
"loss": 0.439,
"step": 504
},
{
"epoch": 0.9911678115799804,
"grad_norm": 0.20373439560246617,
"learning_rate": 1.7277300069993515e-05,
"loss": 0.4205,
"step": 505
},
{
"epoch": 0.9931305201177625,
"grad_norm": 0.18699843473849642,
"learning_rate": 1.7266603220340273e-05,
"loss": 0.4264,
"step": 506
},
{
"epoch": 0.9950932286555446,
"grad_norm": 0.1600200894069467,
"learning_rate": 1.725588872453368e-05,
"loss": 0.4277,
"step": 507
},
{
"epoch": 0.9970559371933267,
"grad_norm": 0.18202529308383952,
"learning_rate": 1.7245156608592727e-05,
"loss": 0.4491,
"step": 508
},
{
"epoch": 0.9990186457311089,
"grad_norm": 0.15887455063463465,
"learning_rate": 1.7234406898579187e-05,
"loss": 0.4558,
"step": 509
},
{
"epoch": 1.0,
"grad_norm": 0.15887455063463465,
"learning_rate": 1.7223639620597556e-05,
"loss": 0.209,
"step": 510
},
{
"epoch": 1.0019627085377822,
"grad_norm": 0.1825187909705365,
"learning_rate": 1.7212854800794998e-05,
"loss": 0.3468,
"step": 511
},
{
"epoch": 1.0039254170755643,
"grad_norm": 0.20957758969625895,
"learning_rate": 1.7202052465361268e-05,
"loss": 0.3492,
"step": 512
},
{
"epoch": 1.0039254170755643,
"eval_loss": 0.4533812999725342,
"eval_runtime": 245.7967,
"eval_samples_per_second": 123.488,
"eval_steps_per_second": 3.861,
"step": 512
},
{
"epoch": 1.0058881256133465,
"grad_norm": 0.28069305660927224,
"learning_rate": 1.719123264052866e-05,
"loss": 0.339,
"step": 513
},
{
"epoch": 1.0078508341511285,
"grad_norm": 0.19132375047173436,
"learning_rate": 1.718039535257194e-05,
"loss": 0.3261,
"step": 514
},
{
"epoch": 1.0098135426889108,
"grad_norm": 0.17431307028109386,
"learning_rate": 1.7169540627808276e-05,
"loss": 0.3514,
"step": 515
},
{
"epoch": 1.0117762512266928,
"grad_norm": 0.5111559402715384,
"learning_rate": 1.7158668492597186e-05,
"loss": 0.3241,
"step": 516
},
{
"epoch": 1.013738959764475,
"grad_norm": 0.16901933614697587,
"learning_rate": 1.7147778973340466e-05,
"loss": 0.3228,
"step": 517
},
{
"epoch": 1.015701668302257,
"grad_norm": 0.15876073778651265,
"learning_rate": 1.7136872096482123e-05,
"loss": 0.4057,
"step": 518
},
{
"epoch": 1.0176643768400393,
"grad_norm": 2.1272381042420556,
"learning_rate": 1.7125947888508322e-05,
"loss": 0.3404,
"step": 519
},
{
"epoch": 1.0196270853778213,
"grad_norm": 0.19402112582649061,
"learning_rate": 1.7115006375947304e-05,
"loss": 0.3396,
"step": 520
},
{
"epoch": 1.0215897939156036,
"grad_norm": 0.184906049478743,
"learning_rate": 1.7104047585369345e-05,
"loss": 0.3335,
"step": 521
},
{
"epoch": 1.0235525024533856,
"grad_norm": 0.2027640255413657,
"learning_rate": 1.7093071543386667e-05,
"loss": 0.3336,
"step": 522
},
{
"epoch": 1.0255152109911678,
"grad_norm": 0.1592888909141134,
"learning_rate": 1.7082078276653392e-05,
"loss": 0.3224,
"step": 523
},
{
"epoch": 1.0274779195289498,
"grad_norm": 0.2560733378030634,
"learning_rate": 1.7071067811865477e-05,
"loss": 0.3356,
"step": 524
},
{
"epoch": 1.029440628066732,
"grad_norm": 0.15861327684708404,
"learning_rate": 1.7060040175760626e-05,
"loss": 0.3237,
"step": 525
},
{
"epoch": 1.0314033366045143,
"grad_norm": 0.135405942028162,
"learning_rate": 1.7048995395118253e-05,
"loss": 0.3417,
"step": 526
},
{
"epoch": 1.0333660451422964,
"grad_norm": 0.15133569201059896,
"learning_rate": 1.7037933496759404e-05,
"loss": 0.3343,
"step": 527
},
{
"epoch": 1.0353287536800786,
"grad_norm": 0.134069507215934,
"learning_rate": 1.7026854507546694e-05,
"loss": 0.3338,
"step": 528
},
{
"epoch": 1.0372914622178606,
"grad_norm": 18.90339450893893,
"learning_rate": 1.7015758454384234e-05,
"loss": 0.346,
"step": 529
},
{
"epoch": 1.0392541707556429,
"grad_norm": 0.19965137004126413,
"learning_rate": 1.7004645364217584e-05,
"loss": 0.3264,
"step": 530
},
{
"epoch": 1.0412168792934249,
"grad_norm": 0.16909173895106233,
"learning_rate": 1.699351526403367e-05,
"loss": 0.3292,
"step": 531
},
{
"epoch": 1.0431795878312071,
"grad_norm": 0.2342868214892783,
"learning_rate": 1.698236818086073e-05,
"loss": 0.3262,
"step": 532
},
{
"epoch": 1.0451422963689891,
"grad_norm": 0.18469564769004154,
"learning_rate": 1.6971204141768235e-05,
"loss": 0.3296,
"step": 533
},
{
"epoch": 1.0471050049067714,
"grad_norm": 0.157241496947566,
"learning_rate": 1.6960023173866834e-05,
"loss": 0.3347,
"step": 534
},
{
"epoch": 1.0490677134445534,
"grad_norm": 0.17900244071577498,
"learning_rate": 1.6948825304308293e-05,
"loss": 0.3309,
"step": 535
},
{
"epoch": 1.0510304219823356,
"grad_norm": 0.14808820544169662,
"learning_rate": 1.693761056028542e-05,
"loss": 0.3302,
"step": 536
},
{
"epoch": 1.0529931305201177,
"grad_norm": 0.13596938535164,
"learning_rate": 1.6926378969031988e-05,
"loss": 0.3311,
"step": 537
},
{
"epoch": 1.0549558390579,
"grad_norm": 0.14432634269560748,
"learning_rate": 1.6915130557822698e-05,
"loss": 0.3305,
"step": 538
},
{
"epoch": 1.056918547595682,
"grad_norm": 0.1499905448155453,
"learning_rate": 1.6903865353973087e-05,
"loss": 0.32,
"step": 539
},
{
"epoch": 1.0588812561334642,
"grad_norm": 0.15652727350645462,
"learning_rate": 1.689258338483947e-05,
"loss": 0.3289,
"step": 540
},
{
"epoch": 1.0608439646712464,
"grad_norm": 0.12161904132806402,
"learning_rate": 1.6881284677818892e-05,
"loss": 0.3167,
"step": 541
},
{
"epoch": 1.0628066732090284,
"grad_norm": 0.12695759397898215,
"learning_rate": 1.686996926034902e-05,
"loss": 0.3227,
"step": 542
},
{
"epoch": 1.0647693817468107,
"grad_norm": 0.12620759014174157,
"learning_rate": 1.685863715990811e-05,
"loss": 0.3273,
"step": 543
},
{
"epoch": 1.0667320902845927,
"grad_norm": 0.12300912173446868,
"learning_rate": 1.6847288404014937e-05,
"loss": 0.3234,
"step": 544
},
{
"epoch": 1.068694798822375,
"grad_norm": 0.13604294154259072,
"learning_rate": 1.6835923020228714e-05,
"loss": 0.3218,
"step": 545
},
{
"epoch": 1.070657507360157,
"grad_norm": 0.12911000723405103,
"learning_rate": 1.682454103614904e-05,
"loss": 0.3199,
"step": 546
},
{
"epoch": 1.0726202158979392,
"grad_norm": 0.12394146843939333,
"learning_rate": 1.6813142479415815e-05,
"loss": 0.318,
"step": 547
},
{
"epoch": 1.0745829244357212,
"grad_norm": 0.11914666168696621,
"learning_rate": 1.6801727377709195e-05,
"loss": 0.3225,
"step": 548
},
{
"epoch": 1.0765456329735035,
"grad_norm": 0.3745213254445056,
"learning_rate": 1.6790295758749512e-05,
"loss": 0.3244,
"step": 549
},
{
"epoch": 1.0785083415112855,
"grad_norm": 0.11387952522253666,
"learning_rate": 1.67788476502972e-05,
"loss": 0.327,
"step": 550
},
{
"epoch": 1.0804710500490677,
"grad_norm": 0.1996105910656074,
"learning_rate": 1.6767383080152744e-05,
"loss": 0.3461,
"step": 551
},
{
"epoch": 1.0824337585868498,
"grad_norm": 0.12175257973633413,
"learning_rate": 1.6755902076156606e-05,
"loss": 0.3153,
"step": 552
},
{
"epoch": 1.084396467124632,
"grad_norm": 0.14527489586682849,
"learning_rate": 1.6744404666189146e-05,
"loss": 0.3265,
"step": 553
},
{
"epoch": 1.0863591756624142,
"grad_norm": 0.15226241276567778,
"learning_rate": 1.6732890878170573e-05,
"loss": 0.3289,
"step": 554
},
{
"epoch": 1.0883218842001963,
"grad_norm": 0.16932705018976396,
"learning_rate": 1.6721360740060864e-05,
"loss": 0.3256,
"step": 555
},
{
"epoch": 1.0902845927379785,
"grad_norm": 0.3186308522830804,
"learning_rate": 1.67098142798597e-05,
"loss": 0.3293,
"step": 556
},
{
"epoch": 1.0922473012757605,
"grad_norm": 0.12034379856239824,
"learning_rate": 1.669825152560641e-05,
"loss": 0.3211,
"step": 557
},
{
"epoch": 1.0942100098135428,
"grad_norm": 0.2213258321448838,
"learning_rate": 1.668667250537987e-05,
"loss": 0.3157,
"step": 558
},
{
"epoch": 1.0961727183513248,
"grad_norm": 0.4014643343300227,
"learning_rate": 1.6675077247298475e-05,
"loss": 0.3307,
"step": 559
},
{
"epoch": 1.098135426889107,
"grad_norm": 0.14678019949020815,
"learning_rate": 1.6663465779520042e-05,
"loss": 0.3341,
"step": 560
},
{
"epoch": 1.100098135426889,
"grad_norm": 0.11691608425931684,
"learning_rate": 1.665183813024175e-05,
"loss": 0.3207,
"step": 561
},
{
"epoch": 1.1020608439646713,
"grad_norm": 0.13749331512464025,
"learning_rate": 1.6640194327700087e-05,
"loss": 0.3212,
"step": 562
},
{
"epoch": 1.1040235525024533,
"grad_norm": 0.11763407718323214,
"learning_rate": 1.6628534400170746e-05,
"loss": 0.3184,
"step": 563
},
{
"epoch": 1.1059862610402356,
"grad_norm": 0.10763049323804355,
"learning_rate": 1.6616858375968596e-05,
"loss": 0.3255,
"step": 564
},
{
"epoch": 1.1079489695780176,
"grad_norm": 0.19415854323907672,
"learning_rate": 1.6605166283447587e-05,
"loss": 0.3265,
"step": 565
},
{
"epoch": 1.1099116781157998,
"grad_norm": 0.11392759854867497,
"learning_rate": 1.659345815100069e-05,
"loss": 0.3207,
"step": 566
},
{
"epoch": 1.1118743866535818,
"grad_norm": 0.11596195257412814,
"learning_rate": 1.658173400705983e-05,
"loss": 0.3359,
"step": 567
},
{
"epoch": 1.113837095191364,
"grad_norm": 0.1447901138867746,
"learning_rate": 1.6569993880095807e-05,
"loss": 0.3104,
"step": 568
},
{
"epoch": 1.1157998037291463,
"grad_norm": 0.11274072609698226,
"learning_rate": 1.6558237798618243e-05,
"loss": 0.3207,
"step": 569
},
{
"epoch": 1.1177625122669284,
"grad_norm": 0.1066641260817179,
"learning_rate": 1.6546465791175498e-05,
"loss": 0.33,
"step": 570
},
{
"epoch": 1.1197252208047106,
"grad_norm": 0.10845080737415526,
"learning_rate": 1.6534677886354605e-05,
"loss": 0.3324,
"step": 571
},
{
"epoch": 1.1216879293424926,
"grad_norm": 0.109524857570649,
"learning_rate": 1.6522874112781213e-05,
"loss": 0.33,
"step": 572
},
{
"epoch": 1.1236506378802749,
"grad_norm": 0.1086194965309956,
"learning_rate": 1.6511054499119493e-05,
"loss": 0.3318,
"step": 573
},
{
"epoch": 1.1256133464180569,
"grad_norm": 0.1097485465672784,
"learning_rate": 1.6499219074072087e-05,
"loss": 0.3201,
"step": 574
},
{
"epoch": 1.1275760549558391,
"grad_norm": 0.12784106789186606,
"learning_rate": 1.6487367866380037e-05,
"loss": 0.3277,
"step": 575
},
{
"epoch": 1.1295387634936211,
"grad_norm": 0.11285311191327178,
"learning_rate": 1.6475500904822707e-05,
"loss": 0.3249,
"step": 576
},
{
"epoch": 1.1315014720314034,
"grad_norm": 0.1024337781066779,
"learning_rate": 1.646361821821772e-05,
"loss": 0.3366,
"step": 577
},
{
"epoch": 1.1334641805691854,
"grad_norm": 0.11906272118674736,
"learning_rate": 1.645171983542088e-05,
"loss": 0.3339,
"step": 578
},
{
"epoch": 1.1354268891069677,
"grad_norm": 0.11321533330226549,
"learning_rate": 1.6439805785326114e-05,
"loss": 0.3276,
"step": 579
},
{
"epoch": 1.1373895976447497,
"grad_norm": 0.10571636728133642,
"learning_rate": 1.6427876096865394e-05,
"loss": 0.3289,
"step": 580
},
{
"epoch": 1.139352306182532,
"grad_norm": 0.22333265835518828,
"learning_rate": 1.6415930799008668e-05,
"loss": 0.325,
"step": 581
},
{
"epoch": 1.141315014720314,
"grad_norm": 0.10810955681560351,
"learning_rate": 1.640396992076379e-05,
"loss": 0.4055,
"step": 582
},
{
"epoch": 1.1432777232580962,
"grad_norm": 0.26109788537078404,
"learning_rate": 1.6391993491176445e-05,
"loss": 0.3162,
"step": 583
},
{
"epoch": 1.1452404317958784,
"grad_norm": 0.11788156500671607,
"learning_rate": 1.6380001539330088e-05,
"loss": 0.3132,
"step": 584
},
{
"epoch": 1.1472031403336604,
"grad_norm": 0.111527885844509,
"learning_rate": 1.6367994094345864e-05,
"loss": 0.3213,
"step": 585
},
{
"epoch": 1.1491658488714427,
"grad_norm": 0.10923834950482783,
"learning_rate": 1.6355971185382547e-05,
"loss": 0.3235,
"step": 586
},
{
"epoch": 1.1511285574092247,
"grad_norm": 0.11352228058272218,
"learning_rate": 1.6343932841636455e-05,
"loss": 0.3242,
"step": 587
},
{
"epoch": 1.153091265947007,
"grad_norm": 2.442797565842492,
"learning_rate": 1.6331879092341402e-05,
"loss": 0.3605,
"step": 588
},
{
"epoch": 1.155053974484789,
"grad_norm": 0.1340497478004765,
"learning_rate": 1.631980996676859e-05,
"loss": 0.3231,
"step": 589
},
{
"epoch": 1.1570166830225712,
"grad_norm": 7.251138564125245,
"learning_rate": 1.6307725494226586e-05,
"loss": 0.464,
"step": 590
},
{
"epoch": 1.1589793915603532,
"grad_norm": 1.3172662093711907,
"learning_rate": 1.6295625704061204e-05,
"loss": 0.3337,
"step": 591
},
{
"epoch": 1.1609421000981355,
"grad_norm": 0.31327939692037343,
"learning_rate": 1.6283510625655474e-05,
"loss": 0.3278,
"step": 592
},
{
"epoch": 1.1629048086359175,
"grad_norm": 4.811713971885485,
"learning_rate": 1.6271380288429535e-05,
"loss": 0.3485,
"step": 593
},
{
"epoch": 1.1648675171736997,
"grad_norm": 1.2548300965536887,
"learning_rate": 1.6259234721840595e-05,
"loss": 0.333,
"step": 594
},
{
"epoch": 1.1668302257114818,
"grad_norm": 0.2330658914469722,
"learning_rate": 1.624707395538283e-05,
"loss": 0.3294,
"step": 595
},
{
"epoch": 1.168792934249264,
"grad_norm": 0.24251502191211774,
"learning_rate": 1.6234898018587336e-05,
"loss": 0.3244,
"step": 596
},
{
"epoch": 1.170755642787046,
"grad_norm": 0.235968588849149,
"learning_rate": 1.6222706941022054e-05,
"loss": 0.3284,
"step": 597
},
{
"epoch": 1.1727183513248283,
"grad_norm": 0.2248234241197579,
"learning_rate": 1.6210500752291682e-05,
"loss": 0.3391,
"step": 598
},
{
"epoch": 1.1746810598626105,
"grad_norm": 0.4542190820206638,
"learning_rate": 1.6198279482037617e-05,
"loss": 0.3306,
"step": 599
},
{
"epoch": 1.1766437684003925,
"grad_norm": 0.19360720341661192,
"learning_rate": 1.6186043159937884e-05,
"loss": 0.3293,
"step": 600
},
{
"epoch": 1.1786064769381748,
"grad_norm": 0.33521960291085745,
"learning_rate": 1.6173791815707053e-05,
"loss": 0.3296,
"step": 601
},
{
"epoch": 1.1805691854759568,
"grad_norm": 0.1676340468101461,
"learning_rate": 1.616152547909618e-05,
"loss": 0.3227,
"step": 602
},
{
"epoch": 1.182531894013739,
"grad_norm": 0.166589926871346,
"learning_rate": 1.614924417989272e-05,
"loss": 0.3124,
"step": 603
},
{
"epoch": 1.184494602551521,
"grad_norm": 0.6062179431946622,
"learning_rate": 1.6136947947920477e-05,
"loss": 0.3542,
"step": 604
},
{
"epoch": 1.1864573110893033,
"grad_norm": 0.3148233399749341,
"learning_rate": 1.6124636813039502e-05,
"loss": 0.3583,
"step": 605
},
{
"epoch": 1.1884200196270853,
"grad_norm": 0.16441596245061335,
"learning_rate": 1.611231080514605e-05,
"loss": 0.3295,
"step": 606
},
{
"epoch": 1.1903827281648676,
"grad_norm": 0.16022712151086638,
"learning_rate": 1.609996995417248e-05,
"loss": 0.3271,
"step": 607
},
{
"epoch": 1.1923454367026496,
"grad_norm": 0.29243754067572564,
"learning_rate": 1.608761429008721e-05,
"loss": 0.3337,
"step": 608
},
{
"epoch": 1.1943081452404318,
"grad_norm": 0.15303231575175122,
"learning_rate": 1.6075243842894614e-05,
"loss": 0.336,
"step": 609
},
{
"epoch": 1.1962708537782138,
"grad_norm": 0.14650148016034228,
"learning_rate": 1.606285864263498e-05,
"loss": 0.329,
"step": 610
},
{
"epoch": 1.198233562315996,
"grad_norm": 0.15261136727699032,
"learning_rate": 1.605045871938441e-05,
"loss": 0.3218,
"step": 611
},
{
"epoch": 1.2001962708537781,
"grad_norm": 0.21026340819790304,
"learning_rate": 1.6038044103254775e-05,
"loss": 0.3292,
"step": 612
},
{
"epoch": 1.2021589793915604,
"grad_norm": 0.13503564973923055,
"learning_rate": 1.6025614824393606e-05,
"loss": 0.326,
"step": 613
},
{
"epoch": 1.2041216879293426,
"grad_norm": 0.1312176723931598,
"learning_rate": 1.601317091298406e-05,
"loss": 0.3258,
"step": 614
},
{
"epoch": 1.2060843964671246,
"grad_norm": 0.14448732114233953,
"learning_rate": 1.6000712399244813e-05,
"loss": 0.3358,
"step": 615
},
{
"epoch": 1.2080471050049069,
"grad_norm": 0.12820019901824906,
"learning_rate": 1.5988239313430004e-05,
"loss": 0.3186,
"step": 616
},
{
"epoch": 1.2100098135426889,
"grad_norm": 0.2613273277747731,
"learning_rate": 1.5975751685829167e-05,
"loss": 0.337,
"step": 617
},
{
"epoch": 1.2119725220804711,
"grad_norm": 0.1134463713652465,
"learning_rate": 1.5963249546767144e-05,
"loss": 0.3217,
"step": 618
},
{
"epoch": 1.2139352306182531,
"grad_norm": 0.12611151264418188,
"learning_rate": 1.5950732926604012e-05,
"loss": 0.3182,
"step": 619
},
{
"epoch": 1.2158979391560354,
"grad_norm": 0.11276212188506385,
"learning_rate": 1.5938201855735017e-05,
"loss": 0.3227,
"step": 620
},
{
"epoch": 1.2178606476938174,
"grad_norm": 0.12586598314647854,
"learning_rate": 1.5925656364590504e-05,
"loss": 0.3228,
"step": 621
},
{
"epoch": 1.2198233562315997,
"grad_norm": 0.23240521031529912,
"learning_rate": 1.5913096483635827e-05,
"loss": 0.3173,
"step": 622
},
{
"epoch": 1.2217860647693817,
"grad_norm": 0.12306625560657121,
"learning_rate": 1.5900522243371283e-05,
"loss": 0.3176,
"step": 623
},
{
"epoch": 1.223748773307164,
"grad_norm": 0.11574270115288804,
"learning_rate": 1.5887933674332048e-05,
"loss": 0.3362,
"step": 624
},
{
"epoch": 1.225711481844946,
"grad_norm": 0.113799124679694,
"learning_rate": 1.587533080708809e-05,
"loss": 0.3279,
"step": 625
},
{
"epoch": 1.2276741903827282,
"grad_norm": 0.11649342842785392,
"learning_rate": 1.5862713672244092e-05,
"loss": 0.3276,
"step": 626
},
{
"epoch": 1.2296368989205102,
"grad_norm": 0.25352594498528047,
"learning_rate": 1.5850082300439395e-05,
"loss": 0.3241,
"step": 627
},
{
"epoch": 1.2315996074582924,
"grad_norm": 0.16227315088252345,
"learning_rate": 1.5837436722347902e-05,
"loss": 0.3303,
"step": 628
},
{
"epoch": 1.2335623159960747,
"grad_norm": 0.1678226345146868,
"learning_rate": 1.5824776968678024e-05,
"loss": 0.3315,
"step": 629
},
{
"epoch": 1.2355250245338567,
"grad_norm": 0.13075124341701522,
"learning_rate": 1.5812103070172592e-05,
"loss": 0.3108,
"step": 630
},
{
"epoch": 1.237487733071639,
"grad_norm": 0.11601688723318539,
"learning_rate": 1.5799415057608785e-05,
"loss": 0.3264,
"step": 631
},
{
"epoch": 1.239450441609421,
"grad_norm": 0.11324740404685361,
"learning_rate": 1.578671296179806e-05,
"loss": 0.3177,
"step": 632
},
{
"epoch": 1.2414131501472032,
"grad_norm": 0.11840350625128808,
"learning_rate": 1.5773996813586067e-05,
"loss": 0.3354,
"step": 633
},
{
"epoch": 1.2433758586849852,
"grad_norm": 0.13875834597920514,
"learning_rate": 1.5761266643852587e-05,
"loss": 0.317,
"step": 634
},
{
"epoch": 1.2453385672227675,
"grad_norm": 0.117375827430945,
"learning_rate": 1.574852248351145e-05,
"loss": 0.3234,
"step": 635
},
{
"epoch": 1.2473012757605495,
"grad_norm": 0.11942850208082942,
"learning_rate": 1.573576436351046e-05,
"loss": 0.3087,
"step": 636
},
{
"epoch": 1.2492639842983317,
"grad_norm": 0.1335866448171688,
"learning_rate": 1.572299231483132e-05,
"loss": 0.3296,
"step": 637
},
{
"epoch": 1.2512266928361138,
"grad_norm": 0.12470555907017847,
"learning_rate": 1.5710206368489555e-05,
"loss": 0.3273,
"step": 638
},
{
"epoch": 1.253189401373896,
"grad_norm": 0.1757334780641794,
"learning_rate": 1.569740655553444e-05,
"loss": 0.3206,
"step": 639
},
{
"epoch": 1.2551521099116782,
"grad_norm": 0.11262712029964689,
"learning_rate": 1.5684592907048925e-05,
"loss": 0.3265,
"step": 640
},
{
"epoch": 1.2551521099116782,
"eval_loss": 0.4366247355937958,
"eval_runtime": 245.4297,
"eval_samples_per_second": 123.673,
"eval_steps_per_second": 3.867,
"step": 640
},
{
"epoch": 1.2571148184494603,
"grad_norm": 0.13469264864574937,
"learning_rate": 1.5671765454149558e-05,
"loss": 0.3286,
"step": 641
},
{
"epoch": 1.2590775269872423,
"grad_norm": 0.12053610787587447,
"learning_rate": 1.5658924227986415e-05,
"loss": 0.3275,
"step": 642
},
{
"epoch": 1.2610402355250245,
"grad_norm": 0.14172736929838212,
"learning_rate": 1.5646069259743007e-05,
"loss": 0.3323,
"step": 643
},
{
"epoch": 1.2630029440628068,
"grad_norm": 0.13230460831867666,
"learning_rate": 1.563320058063622e-05,
"loss": 0.3406,
"step": 644
},
{
"epoch": 1.2649656526005888,
"grad_norm": 0.11391731055963321,
"learning_rate": 1.5620318221916245e-05,
"loss": 0.3321,
"step": 645
},
{
"epoch": 1.2669283611383708,
"grad_norm": 0.20710778775533598,
"learning_rate": 1.560742221486648e-05,
"loss": 0.3233,
"step": 646
},
{
"epoch": 1.268891069676153,
"grad_norm": 0.12086849482365936,
"learning_rate": 1.5594512590803476e-05,
"loss": 0.3227,
"step": 647
},
{
"epoch": 1.2708537782139353,
"grad_norm": 0.12570156067267854,
"learning_rate": 1.5581589381076843e-05,
"loss": 0.3219,
"step": 648
},
{
"epoch": 1.2728164867517173,
"grad_norm": 0.12011193178735968,
"learning_rate": 1.556865261706918e-05,
"loss": 0.3182,
"step": 649
},
{
"epoch": 1.2747791952894996,
"grad_norm": 0.1648783672119669,
"learning_rate": 1.5555702330196024e-05,
"loss": 0.3297,
"step": 650
},
{
"epoch": 1.2767419038272816,
"grad_norm": 0.12110445033914359,
"learning_rate": 1.554273855190572e-05,
"loss": 0.3254,
"step": 651
},
{
"epoch": 1.2787046123650638,
"grad_norm": 0.11020981189703709,
"learning_rate": 1.5529761313679396e-05,
"loss": 0.3235,
"step": 652
},
{
"epoch": 1.2806673209028459,
"grad_norm": 0.15853842771502125,
"learning_rate": 1.551677064703086e-05,
"loss": 0.3264,
"step": 653
},
{
"epoch": 1.282630029440628,
"grad_norm": 0.11133610328097518,
"learning_rate": 1.5503766583506522e-05,
"loss": 0.329,
"step": 654
},
{
"epoch": 1.2845927379784103,
"grad_norm": 0.1193501946144463,
"learning_rate": 1.549074915468534e-05,
"loss": 0.332,
"step": 655
},
{
"epoch": 1.2865554465161924,
"grad_norm": 0.12269380787533944,
"learning_rate": 1.5477718392178716e-05,
"loss": 0.3314,
"step": 656
},
{
"epoch": 1.2885181550539744,
"grad_norm": 0.11313715963097541,
"learning_rate": 1.5464674327630437e-05,
"loss": 0.3222,
"step": 657
},
{
"epoch": 1.2904808635917566,
"grad_norm": 0.12048024103575065,
"learning_rate": 1.545161699271659e-05,
"loss": 0.3294,
"step": 658
},
{
"epoch": 1.2924435721295389,
"grad_norm": 0.11947900701309468,
"learning_rate": 1.543854641914549e-05,
"loss": 0.3235,
"step": 659
},
{
"epoch": 1.2944062806673209,
"grad_norm": 0.11207590380323458,
"learning_rate": 1.5425462638657597e-05,
"loss": 0.3266,
"step": 660
},
{
"epoch": 1.296368989205103,
"grad_norm": 0.12528223605914246,
"learning_rate": 1.5412365683025447e-05,
"loss": 0.3321,
"step": 661
},
{
"epoch": 1.2983316977428851,
"grad_norm": 0.10633033440100298,
"learning_rate": 1.5399255584053568e-05,
"loss": 0.3165,
"step": 662
},
{
"epoch": 1.3002944062806674,
"grad_norm": 0.11406876546510498,
"learning_rate": 1.5386132373578405e-05,
"loss": 0.3247,
"step": 663
},
{
"epoch": 1.3022571148184494,
"grad_norm": 0.11139990576253381,
"learning_rate": 1.5372996083468242e-05,
"loss": 0.3261,
"step": 664
},
{
"epoch": 1.3042198233562317,
"grad_norm": 0.10272448533365049,
"learning_rate": 1.5359846745623128e-05,
"loss": 0.316,
"step": 665
},
{
"epoch": 1.3061825318940137,
"grad_norm": 0.10645469773905245,
"learning_rate": 1.5346684391974792e-05,
"loss": 0.3311,
"step": 666
},
{
"epoch": 1.308145240431796,
"grad_norm": 0.10518169859307575,
"learning_rate": 1.5333509054486583e-05,
"loss": 0.3245,
"step": 667
},
{
"epoch": 1.310107948969578,
"grad_norm": 0.2242263732415979,
"learning_rate": 1.5320320765153367e-05,
"loss": 0.3109,
"step": 668
},
{
"epoch": 1.3120706575073602,
"grad_norm": 0.12309952064417946,
"learning_rate": 1.5307119556001463e-05,
"loss": 0.3196,
"step": 669
},
{
"epoch": 1.3140333660451424,
"grad_norm": 0.1129490904179756,
"learning_rate": 1.529390545908857e-05,
"loss": 0.3222,
"step": 670
},
{
"epoch": 1.3159960745829244,
"grad_norm": 0.1196632767822302,
"learning_rate": 1.528067850650368e-05,
"loss": 0.3182,
"step": 671
},
{
"epoch": 1.3179587831207065,
"grad_norm": 0.11338552151174172,
"learning_rate": 1.526743873036701e-05,
"loss": 0.3419,
"step": 672
},
{
"epoch": 1.3199214916584887,
"grad_norm": 0.17835671344319715,
"learning_rate": 1.5254186162829903e-05,
"loss": 0.3169,
"step": 673
},
{
"epoch": 1.321884200196271,
"grad_norm": 0.10783382011254725,
"learning_rate": 1.5240920836074777e-05,
"loss": 0.3196,
"step": 674
},
{
"epoch": 1.323846908734053,
"grad_norm": 0.10869968464668042,
"learning_rate": 1.5227642782315037e-05,
"loss": 0.3208,
"step": 675
},
{
"epoch": 1.325809617271835,
"grad_norm": 0.10637526513057953,
"learning_rate": 1.5214352033794981e-05,
"loss": 0.3263,
"step": 676
},
{
"epoch": 1.3277723258096172,
"grad_norm": 0.1239954358141019,
"learning_rate": 1.5201048622789747e-05,
"loss": 0.3262,
"step": 677
},
{
"epoch": 1.3297350343473995,
"grad_norm": 0.10993374676490457,
"learning_rate": 1.5187732581605217e-05,
"loss": 0.3367,
"step": 678
},
{
"epoch": 1.3316977428851815,
"grad_norm": 0.12433721542893529,
"learning_rate": 1.5174403942577942e-05,
"loss": 0.3329,
"step": 679
},
{
"epoch": 1.3336604514229637,
"grad_norm": 0.10189759821643117,
"learning_rate": 1.5161062738075068e-05,
"loss": 0.3191,
"step": 680
},
{
"epoch": 1.3356231599607458,
"grad_norm": 0.13313656343105662,
"learning_rate": 1.5147709000494258e-05,
"loss": 0.32,
"step": 681
},
{
"epoch": 1.337585868498528,
"grad_norm": 0.13908249732191438,
"learning_rate": 1.5134342762263606e-05,
"loss": 0.3295,
"step": 682
},
{
"epoch": 1.33954857703631,
"grad_norm": 0.11489030169841424,
"learning_rate": 1.5120964055841563e-05,
"loss": 0.3316,
"step": 683
},
{
"epoch": 1.3415112855740923,
"grad_norm": 0.3799510436085688,
"learning_rate": 1.5107572913716859e-05,
"loss": 0.3162,
"step": 684
},
{
"epoch": 1.3434739941118745,
"grad_norm": 0.11349571041844954,
"learning_rate": 1.509416936840842e-05,
"loss": 0.333,
"step": 685
},
{
"epoch": 1.3454367026496565,
"grad_norm": 0.11005963527146614,
"learning_rate": 1.5080753452465296e-05,
"loss": 0.3272,
"step": 686
},
{
"epoch": 1.3473994111874386,
"grad_norm": 0.11959100787233001,
"learning_rate": 1.5067325198466576e-05,
"loss": 0.3311,
"step": 687
},
{
"epoch": 1.3493621197252208,
"grad_norm": 0.10886477394608883,
"learning_rate": 1.505388463902131e-05,
"loss": 0.3335,
"step": 688
},
{
"epoch": 1.351324828263003,
"grad_norm": 0.11344016969622332,
"learning_rate": 1.504043180676843e-05,
"loss": 0.3207,
"step": 689
},
{
"epoch": 1.353287536800785,
"grad_norm": 0.1102062785109799,
"learning_rate": 1.502696673437667e-05,
"loss": 0.3338,
"step": 690
},
{
"epoch": 1.3552502453385673,
"grad_norm": 0.1134536009046737,
"learning_rate": 1.5013489454544494e-05,
"loss": 0.3212,
"step": 691
},
{
"epoch": 1.3572129538763493,
"grad_norm": 0.10812804193252123,
"learning_rate": 1.5000000000000002e-05,
"loss": 0.3244,
"step": 692
},
{
"epoch": 1.3591756624141316,
"grad_norm": 0.10930042393971806,
"learning_rate": 1.4986498403500864e-05,
"loss": 0.3196,
"step": 693
},
{
"epoch": 1.3611383709519136,
"grad_norm": 0.11325005359482505,
"learning_rate": 1.4972984697834238e-05,
"loss": 0.3283,
"step": 694
},
{
"epoch": 1.3631010794896958,
"grad_norm": 0.11944962015232705,
"learning_rate": 1.4959458915816681e-05,
"loss": 0.3071,
"step": 695
},
{
"epoch": 1.3650637880274779,
"grad_norm": 0.11035654504778693,
"learning_rate": 1.4945921090294076e-05,
"loss": 0.3258,
"step": 696
},
{
"epoch": 1.36702649656526,
"grad_norm": 0.2791207437797803,
"learning_rate": 1.4932371254141562e-05,
"loss": 0.3444,
"step": 697
},
{
"epoch": 1.3689892051030421,
"grad_norm": 0.17253680739822636,
"learning_rate": 1.4918809440263435e-05,
"loss": 0.3263,
"step": 698
},
{
"epoch": 1.3709519136408244,
"grad_norm": 0.11509592444315454,
"learning_rate": 1.4905235681593079e-05,
"loss": 0.3159,
"step": 699
},
{
"epoch": 1.3729146221786066,
"grad_norm": 0.11893850145942958,
"learning_rate": 1.4891650011092896e-05,
"loss": 0.3361,
"step": 700
},
{
"epoch": 1.3748773307163886,
"grad_norm": 0.11595816082844375,
"learning_rate": 1.4878052461754192e-05,
"loss": 0.3172,
"step": 701
},
{
"epoch": 1.3768400392541706,
"grad_norm": 0.18006969131556863,
"learning_rate": 1.486444306659714e-05,
"loss": 0.3262,
"step": 702
},
{
"epoch": 1.3788027477919529,
"grad_norm": 0.11978594076007715,
"learning_rate": 1.4850821858670668e-05,
"loss": 0.3223,
"step": 703
},
{
"epoch": 1.3807654563297351,
"grad_norm": 0.1214666563188013,
"learning_rate": 1.4837188871052399e-05,
"loss": 0.3287,
"step": 704
},
{
"epoch": 1.3827281648675172,
"grad_norm": 0.1087134010196951,
"learning_rate": 1.4823544136848554e-05,
"loss": 0.3211,
"step": 705
},
{
"epoch": 1.3846908734052994,
"grad_norm": 0.11263273643445283,
"learning_rate": 1.4809887689193878e-05,
"loss": 0.321,
"step": 706
},
{
"epoch": 1.3866535819430814,
"grad_norm": 0.10441056093366385,
"learning_rate": 1.4796219561251569e-05,
"loss": 0.3205,
"step": 707
},
{
"epoch": 1.3886162904808637,
"grad_norm": 0.11847864217596303,
"learning_rate": 1.4782539786213184e-05,
"loss": 0.3358,
"step": 708
},
{
"epoch": 1.3905789990186457,
"grad_norm": 0.11231538216944834,
"learning_rate": 1.4768848397298562e-05,
"loss": 0.3265,
"step": 709
},
{
"epoch": 1.392541707556428,
"grad_norm": 0.1033742948772916,
"learning_rate": 1.4755145427755755e-05,
"loss": 0.3295,
"step": 710
},
{
"epoch": 1.39450441609421,
"grad_norm": 0.11293962071828316,
"learning_rate": 1.4741430910860918e-05,
"loss": 0.3284,
"step": 711
},
{
"epoch": 1.3964671246319922,
"grad_norm": 0.10568902630893928,
"learning_rate": 1.4727704879918272e-05,
"loss": 0.324,
"step": 712
},
{
"epoch": 1.3984298331697742,
"grad_norm": 0.10454582574341005,
"learning_rate": 1.4713967368259981e-05,
"loss": 0.3294,
"step": 713
},
{
"epoch": 1.4003925417075564,
"grad_norm": 0.11151842422641142,
"learning_rate": 1.4700218409246087e-05,
"loss": 0.3272,
"step": 714
},
{
"epoch": 1.4023552502453387,
"grad_norm": 0.1918179917026182,
"learning_rate": 1.4686458036264446e-05,
"loss": 0.3157,
"step": 715
},
{
"epoch": 1.4043179587831207,
"grad_norm": 0.11036284447673995,
"learning_rate": 1.4672686282730622e-05,
"loss": 0.312,
"step": 716
},
{
"epoch": 1.4062806673209027,
"grad_norm": 0.13510751112765382,
"learning_rate": 1.4658903182087814e-05,
"loss": 0.3382,
"step": 717
},
{
"epoch": 1.408243375858685,
"grad_norm": 0.10984538370082228,
"learning_rate": 1.4645108767806778e-05,
"loss": 0.3186,
"step": 718
},
{
"epoch": 1.4102060843964672,
"grad_norm": 0.11386113381529725,
"learning_rate": 1.4631303073385745e-05,
"loss": 0.333,
"step": 719
},
{
"epoch": 1.4121687929342492,
"grad_norm": 0.12922683158535153,
"learning_rate": 1.4617486132350343e-05,
"loss": 0.3131,
"step": 720
},
{
"epoch": 1.4141315014720315,
"grad_norm": 0.11429762711313823,
"learning_rate": 1.4603657978253499e-05,
"loss": 0.3356,
"step": 721
},
{
"epoch": 1.4160942100098135,
"grad_norm": 0.11032902828120253,
"learning_rate": 1.4589818644675378e-05,
"loss": 0.324,
"step": 722
},
{
"epoch": 1.4180569185475957,
"grad_norm": 0.10896038916378659,
"learning_rate": 1.4575968165223297e-05,
"loss": 0.3273,
"step": 723
},
{
"epoch": 1.4200196270853778,
"grad_norm": 0.11197390514652242,
"learning_rate": 1.4562106573531632e-05,
"loss": 0.3212,
"step": 724
},
{
"epoch": 1.42198233562316,
"grad_norm": 0.11854344219452063,
"learning_rate": 1.4548233903261746e-05,
"loss": 0.3266,
"step": 725
},
{
"epoch": 1.423945044160942,
"grad_norm": 0.10788963190927026,
"learning_rate": 1.4534350188101905e-05,
"loss": 0.3261,
"step": 726
},
{
"epoch": 1.4259077526987243,
"grad_norm": 0.10409181843438578,
"learning_rate": 1.45204554617672e-05,
"loss": 0.3267,
"step": 727
},
{
"epoch": 1.4278704612365063,
"grad_norm": 0.11457863412214213,
"learning_rate": 1.4506549757999456e-05,
"loss": 0.3322,
"step": 728
},
{
"epoch": 1.4298331697742885,
"grad_norm": 1.3410888637080636,
"learning_rate": 1.4492633110567155e-05,
"loss": 0.3836,
"step": 729
},
{
"epoch": 1.4317958783120708,
"grad_norm": 0.12437392309356543,
"learning_rate": 1.4478705553265363e-05,
"loss": 0.3244,
"step": 730
},
{
"epoch": 1.4337585868498528,
"grad_norm": 0.11654470773611598,
"learning_rate": 1.446476711991563e-05,
"loss": 0.334,
"step": 731
},
{
"epoch": 1.4357212953876348,
"grad_norm": 0.295599491807141,
"learning_rate": 1.4450817844365924e-05,
"loss": 0.3292,
"step": 732
},
{
"epoch": 1.437684003925417,
"grad_norm": 0.23594221943762184,
"learning_rate": 1.4436857760490539e-05,
"loss": 0.3244,
"step": 733
},
{
"epoch": 1.4396467124631993,
"grad_norm": 0.13208951730861257,
"learning_rate": 1.4422886902190014e-05,
"loss": 0.3218,
"step": 734
},
{
"epoch": 1.4416094210009813,
"grad_norm": 0.12203894918097015,
"learning_rate": 1.4408905303391054e-05,
"loss": 0.3184,
"step": 735
},
{
"epoch": 1.4435721295387636,
"grad_norm": 0.12073834849861981,
"learning_rate": 1.4394912998046451e-05,
"loss": 0.3297,
"step": 736
},
{
"epoch": 1.4455348380765456,
"grad_norm": 0.42701652972686566,
"learning_rate": 1.4380910020134988e-05,
"loss": 0.332,
"step": 737
},
{
"epoch": 1.4474975466143278,
"grad_norm": 0.12608912444867507,
"learning_rate": 1.436689640366137e-05,
"loss": 0.3267,
"step": 738
},
{
"epoch": 1.4494602551521099,
"grad_norm": 0.16001847759687152,
"learning_rate": 1.435287218265614e-05,
"loss": 0.3315,
"step": 739
},
{
"epoch": 1.451422963689892,
"grad_norm": 0.12030660319385979,
"learning_rate": 1.4338837391175582e-05,
"loss": 0.3285,
"step": 740
},
{
"epoch": 1.4533856722276741,
"grad_norm": 0.1253610461061891,
"learning_rate": 1.4324792063301662e-05,
"loss": 0.3351,
"step": 741
},
{
"epoch": 1.4553483807654564,
"grad_norm": 0.11052209080423725,
"learning_rate": 1.4310736233141926e-05,
"loss": 0.3289,
"step": 742
},
{
"epoch": 1.4573110893032384,
"grad_norm": 0.128573183549157,
"learning_rate": 1.4296669934829425e-05,
"loss": 0.3281,
"step": 743
},
{
"epoch": 1.4592737978410206,
"grad_norm": 0.11507190335793269,
"learning_rate": 1.4282593202522627e-05,
"loss": 0.331,
"step": 744
},
{
"epoch": 1.4612365063788029,
"grad_norm": 0.11860253438573154,
"learning_rate": 1.4268506070405345e-05,
"loss": 0.3278,
"step": 745
},
{
"epoch": 1.463199214916585,
"grad_norm": 0.10619138540627954,
"learning_rate": 1.4254408572686642e-05,
"loss": 0.3211,
"step": 746
},
{
"epoch": 1.465161923454367,
"grad_norm": 0.2575777832730168,
"learning_rate": 1.424030074360075e-05,
"loss": 0.3263,
"step": 747
},
{
"epoch": 1.4671246319921492,
"grad_norm": 0.11876073393454559,
"learning_rate": 1.4226182617406996e-05,
"loss": 0.3416,
"step": 748
},
{
"epoch": 1.4690873405299314,
"grad_norm": 0.1093129912760756,
"learning_rate": 1.4212054228389712e-05,
"loss": 0.3233,
"step": 749
},
{
"epoch": 1.4710500490677134,
"grad_norm": 0.11230780423472099,
"learning_rate": 1.4197915610858143e-05,
"loss": 0.32,
"step": 750
},
{
"epoch": 1.4730127576054957,
"grad_norm": 0.12038548720872048,
"learning_rate": 1.4183766799146383e-05,
"loss": 0.3246,
"step": 751
},
{
"epoch": 1.4749754661432777,
"grad_norm": 0.15570969251778305,
"learning_rate": 1.4169607827613284e-05,
"loss": 0.3261,
"step": 752
},
{
"epoch": 1.47693817468106,
"grad_norm": 0.3408875294488595,
"learning_rate": 1.4155438730642354e-05,
"loss": 0.3372,
"step": 753
},
{
"epoch": 1.478900883218842,
"grad_norm": 0.10372129141329851,
"learning_rate": 1.4141259542641706e-05,
"loss": 0.3275,
"step": 754
},
{
"epoch": 1.4808635917566242,
"grad_norm": 0.10861393950490046,
"learning_rate": 1.4127070298043949e-05,
"loss": 0.3195,
"step": 755
},
{
"epoch": 1.4828263002944062,
"grad_norm": 0.2287314698384735,
"learning_rate": 1.4112871031306118e-05,
"loss": 0.3392,
"step": 756
},
{
"epoch": 1.4847890088321885,
"grad_norm": 0.14387105066919137,
"learning_rate": 1.4098661776909581e-05,
"loss": 0.3278,
"step": 757
},
{
"epoch": 1.4867517173699705,
"grad_norm": 0.11031577836845029,
"learning_rate": 1.4084442569359964e-05,
"loss": 0.3261,
"step": 758
},
{
"epoch": 1.4887144259077527,
"grad_norm": 0.10372345153909722,
"learning_rate": 1.4070213443187062e-05,
"loss": 0.304,
"step": 759
},
{
"epoch": 1.490677134445535,
"grad_norm": 1.1167059057492386,
"learning_rate": 1.4055974432944753e-05,
"loss": 0.3512,
"step": 760
},
{
"epoch": 1.492639842983317,
"grad_norm": 0.3507372435965062,
"learning_rate": 1.404172557321092e-05,
"loss": 0.3467,
"step": 761
},
{
"epoch": 1.494602551521099,
"grad_norm": 0.13192221407076998,
"learning_rate": 1.4027466898587375e-05,
"loss": 0.3255,
"step": 762
},
{
"epoch": 1.4965652600588812,
"grad_norm": 0.13116992214786896,
"learning_rate": 1.401319844369974e-05,
"loss": 0.3385,
"step": 763
},
{
"epoch": 1.4985279685966635,
"grad_norm": 0.14515008145396116,
"learning_rate": 1.3998920243197408e-05,
"loss": 0.331,
"step": 764
},
{
"epoch": 1.5004906771344455,
"grad_norm": 0.12599311264669502,
"learning_rate": 1.3984632331753436e-05,
"loss": 0.3184,
"step": 765
},
{
"epoch": 1.5024533856722275,
"grad_norm": 0.11722994192309723,
"learning_rate": 1.3970334744064451e-05,
"loss": 0.3186,
"step": 766
},
{
"epoch": 1.5044160942100098,
"grad_norm": 0.1297039712182582,
"learning_rate": 1.395602751485059e-05,
"loss": 0.3239,
"step": 767
},
{
"epoch": 1.506378802747792,
"grad_norm": 0.12096476484978132,
"learning_rate": 1.3941710678855396e-05,
"loss": 0.3299,
"step": 768
},
{
"epoch": 1.506378802747792,
"eval_loss": 0.4330715835094452,
"eval_runtime": 245.5392,
"eval_samples_per_second": 123.618,
"eval_steps_per_second": 3.865,
"step": 768
},
{
"epoch": 1.508341511285574,
"grad_norm": 0.12230386295101764,
"learning_rate": 1.3927384270845744e-05,
"loss": 0.3251,
"step": 769
},
{
"epoch": 1.5103042198233563,
"grad_norm": 0.5983535137475166,
"learning_rate": 1.391304832561175e-05,
"loss": 0.3648,
"step": 770
},
{
"epoch": 1.5122669283611385,
"grad_norm": 0.11674041507665114,
"learning_rate": 1.38987028779667e-05,
"loss": 0.3392,
"step": 771
},
{
"epoch": 1.5142296368989205,
"grad_norm": 0.1729456768990231,
"learning_rate": 1.3884347962746949e-05,
"loss": 0.3225,
"step": 772
},
{
"epoch": 1.5161923454367026,
"grad_norm": 0.1252493496614802,
"learning_rate": 1.3869983614811837e-05,
"loss": 0.3209,
"step": 773
},
{
"epoch": 1.5181550539744848,
"grad_norm": 0.11700364355947357,
"learning_rate": 1.3855609869043618e-05,
"loss": 0.3315,
"step": 774
},
{
"epoch": 1.520117762512267,
"grad_norm": 0.1462155960573189,
"learning_rate": 1.384122676034737e-05,
"loss": 0.3259,
"step": 775
},
{
"epoch": 1.522080471050049,
"grad_norm": 0.13319429573990046,
"learning_rate": 1.3826834323650899e-05,
"loss": 0.3334,
"step": 776
},
{
"epoch": 1.524043179587831,
"grad_norm": 0.12414522959921269,
"learning_rate": 1.381243259390467e-05,
"loss": 0.3316,
"step": 777
},
{
"epoch": 1.5260058881256133,
"grad_norm": 0.12126161126577431,
"learning_rate": 1.3798021606081713e-05,
"loss": 0.3178,
"step": 778
},
{
"epoch": 1.5279685966633956,
"grad_norm": 0.11507460737639888,
"learning_rate": 1.3783601395177537e-05,
"loss": 0.3294,
"step": 779
},
{
"epoch": 1.5299313052011776,
"grad_norm": 0.11575304901544674,
"learning_rate": 1.3769171996210053e-05,
"loss": 0.3223,
"step": 780
},
{
"epoch": 1.5318940137389596,
"grad_norm": 0.131991412651207,
"learning_rate": 1.3754733444219488e-05,
"loss": 0.3302,
"step": 781
},
{
"epoch": 1.5338567222767419,
"grad_norm": 0.190709934290855,
"learning_rate": 1.3740285774268282e-05,
"loss": 0.3371,
"step": 782
},
{
"epoch": 1.535819430814524,
"grad_norm": 0.47614707484422286,
"learning_rate": 1.372582902144103e-05,
"loss": 0.3193,
"step": 783
},
{
"epoch": 1.5377821393523061,
"grad_norm": 0.11311414215315257,
"learning_rate": 1.371136322084438e-05,
"loss": 0.3147,
"step": 784
},
{
"epoch": 1.5397448478900884,
"grad_norm": 0.12393935880179471,
"learning_rate": 1.3696888407606952e-05,
"loss": 0.3294,
"step": 785
},
{
"epoch": 1.5417075564278706,
"grad_norm": 0.12459873866982599,
"learning_rate": 1.3682404616879246e-05,
"loss": 0.3523,
"step": 786
},
{
"epoch": 1.5436702649656526,
"grad_norm": 1.5376937047472998,
"learning_rate": 1.3667911883833573e-05,
"loss": 0.3232,
"step": 787
},
{
"epoch": 1.5456329735034346,
"grad_norm": 0.17863323707586204,
"learning_rate": 1.3653410243663953e-05,
"loss": 0.3294,
"step": 788
},
{
"epoch": 1.547595682041217,
"grad_norm": 0.13291935976093486,
"learning_rate": 1.3638899731586036e-05,
"loss": 0.3343,
"step": 789
},
{
"epoch": 1.5495583905789991,
"grad_norm": 0.19803813677781645,
"learning_rate": 1.3624380382837017e-05,
"loss": 0.3359,
"step": 790
},
{
"epoch": 1.5515210991167812,
"grad_norm": 0.1595795029165099,
"learning_rate": 1.3609852232675558e-05,
"loss": 0.326,
"step": 791
},
{
"epoch": 1.5534838076545632,
"grad_norm": 0.14037586512565553,
"learning_rate": 1.3595315316381676e-05,
"loss": 0.3302,
"step": 792
},
{
"epoch": 1.5554465161923454,
"grad_norm": 0.16013145607494655,
"learning_rate": 1.3580769669256695e-05,
"loss": 0.3383,
"step": 793
},
{
"epoch": 1.5574092247301277,
"grad_norm": 0.13377340237611735,
"learning_rate": 1.3566215326623131e-05,
"loss": 0.3235,
"step": 794
},
{
"epoch": 1.5593719332679097,
"grad_norm": 0.1385245785766373,
"learning_rate": 1.3551652323824617e-05,
"loss": 0.3214,
"step": 795
},
{
"epoch": 1.5613346418056917,
"grad_norm": 0.15040533378760934,
"learning_rate": 1.3537080696225815e-05,
"loss": 0.3396,
"step": 796
},
{
"epoch": 1.563297350343474,
"grad_norm": 0.1329758232018135,
"learning_rate": 1.3522500479212337e-05,
"loss": 0.3358,
"step": 797
},
{
"epoch": 1.5652600588812562,
"grad_norm": 0.12257059717404521,
"learning_rate": 1.3507911708190646e-05,
"loss": 0.3176,
"step": 798
},
{
"epoch": 1.5672227674190382,
"grad_norm": 0.13209253543187308,
"learning_rate": 1.3493314418587982e-05,
"loss": 0.3314,
"step": 799
},
{
"epoch": 1.5691854759568205,
"grad_norm": 0.11457657070098855,
"learning_rate": 1.3478708645852272e-05,
"loss": 0.3306,
"step": 800
},
{
"epoch": 1.5711481844946027,
"grad_norm": 0.19371352224283817,
"learning_rate": 1.3464094425452046e-05,
"loss": 0.3145,
"step": 801
},
{
"epoch": 1.5731108930323847,
"grad_norm": 0.10702419786852241,
"learning_rate": 1.3449471792876333e-05,
"loss": 0.3227,
"step": 802
},
{
"epoch": 1.5750736015701667,
"grad_norm": 0.11647703035094253,
"learning_rate": 1.3434840783634611e-05,
"loss": 0.3246,
"step": 803
},
{
"epoch": 1.577036310107949,
"grad_norm": 0.14359657153339322,
"learning_rate": 1.342020143325669e-05,
"loss": 0.3198,
"step": 804
},
{
"epoch": 1.5789990186457312,
"grad_norm": 0.11430760797920907,
"learning_rate": 1.3405553777292627e-05,
"loss": 0.3174,
"step": 805
},
{
"epoch": 1.5809617271835132,
"grad_norm": 0.12045105701279527,
"learning_rate": 1.3390897851312667e-05,
"loss": 0.3281,
"step": 806
},
{
"epoch": 1.5829244357212953,
"grad_norm": 0.11152977267946707,
"learning_rate": 1.3376233690907126e-05,
"loss": 0.3343,
"step": 807
},
{
"epoch": 1.5848871442590775,
"grad_norm": 0.11046335035659553,
"learning_rate": 1.336156133168631e-05,
"loss": 0.3222,
"step": 808
},
{
"epoch": 1.5868498527968598,
"grad_norm": 0.2805415736645124,
"learning_rate": 1.3346880809280451e-05,
"loss": 0.3285,
"step": 809
},
{
"epoch": 1.5888125613346418,
"grad_norm": 0.1165026997728163,
"learning_rate": 1.3332192159339595e-05,
"loss": 0.3327,
"step": 810
},
{
"epoch": 1.5907752698724238,
"grad_norm": 0.10269266294028667,
"learning_rate": 1.3317495417533523e-05,
"loss": 0.3288,
"step": 811
},
{
"epoch": 1.592737978410206,
"grad_norm": 0.10882094664575565,
"learning_rate": 1.3302790619551673e-05,
"loss": 0.3226,
"step": 812
},
{
"epoch": 1.5947006869479883,
"grad_norm": 0.10402717390991092,
"learning_rate": 1.3288077801103041e-05,
"loss": 0.3242,
"step": 813
},
{
"epoch": 1.5966633954857703,
"grad_norm": 0.11773324479057384,
"learning_rate": 1.3273356997916106e-05,
"loss": 0.3312,
"step": 814
},
{
"epoch": 1.5986261040235525,
"grad_norm": 0.10710334700945738,
"learning_rate": 1.3258628245738726e-05,
"loss": 0.3249,
"step": 815
},
{
"epoch": 1.6005888125613348,
"grad_norm": 0.10399693106990227,
"learning_rate": 1.3243891580338074e-05,
"loss": 0.3272,
"step": 816
},
{
"epoch": 1.6025515210991168,
"grad_norm": 0.107596072668019,
"learning_rate": 1.3229147037500534e-05,
"loss": 0.3334,
"step": 817
},
{
"epoch": 1.6045142296368988,
"grad_norm": 0.1069444594286671,
"learning_rate": 1.3214394653031616e-05,
"loss": 0.318,
"step": 818
},
{
"epoch": 1.606476938174681,
"grad_norm": 0.2003747727263415,
"learning_rate": 1.3199634462755886e-05,
"loss": 0.321,
"step": 819
},
{
"epoch": 1.6084396467124633,
"grad_norm": 0.10473434193181139,
"learning_rate": 1.3184866502516846e-05,
"loss": 0.3289,
"step": 820
},
{
"epoch": 1.6104023552502453,
"grad_norm": 0.11035758045532282,
"learning_rate": 1.3170090808176883e-05,
"loss": 0.3302,
"step": 821
},
{
"epoch": 1.6123650637880274,
"grad_norm": 0.10519561826066223,
"learning_rate": 1.3155307415617156e-05,
"loss": 0.3264,
"step": 822
},
{
"epoch": 1.6143277723258096,
"grad_norm": 0.11015590527571209,
"learning_rate": 1.3140516360737523e-05,
"loss": 0.3173,
"step": 823
},
{
"epoch": 1.6162904808635918,
"grad_norm": 0.10092061607132297,
"learning_rate": 1.3125717679456447e-05,
"loss": 0.331,
"step": 824
},
{
"epoch": 1.6182531894013739,
"grad_norm": 0.1024320679905342,
"learning_rate": 1.3110911407710909e-05,
"loss": 0.3177,
"step": 825
},
{
"epoch": 1.6202158979391559,
"grad_norm": 0.10726261404380205,
"learning_rate": 1.309609758145633e-05,
"loss": 0.33,
"step": 826
},
{
"epoch": 1.6221786064769381,
"grad_norm": 0.30696269389558967,
"learning_rate": 1.308127623666647e-05,
"loss": 0.3217,
"step": 827
},
{
"epoch": 1.6241413150147204,
"grad_norm": 0.11038918441384825,
"learning_rate": 1.3066447409333345e-05,
"loss": 0.3428,
"step": 828
},
{
"epoch": 1.6261040235525024,
"grad_norm": 0.13114032892510316,
"learning_rate": 1.3051611135467145e-05,
"loss": 0.3217,
"step": 829
},
{
"epoch": 1.6280667320902846,
"grad_norm": 0.10209817496354183,
"learning_rate": 1.3036767451096148e-05,
"loss": 0.3169,
"step": 830
},
{
"epoch": 1.6300294406280669,
"grad_norm": 0.11163305235516735,
"learning_rate": 1.3021916392266618e-05,
"loss": 0.3215,
"step": 831
},
{
"epoch": 1.631992149165849,
"grad_norm": 0.10335841003276967,
"learning_rate": 1.300705799504273e-05,
"loss": 0.3179,
"step": 832
},
{
"epoch": 1.633954857703631,
"grad_norm": 0.10298399337006449,
"learning_rate": 1.2992192295506489e-05,
"loss": 0.3342,
"step": 833
},
{
"epoch": 1.6359175662414132,
"grad_norm": 0.10601682383580134,
"learning_rate": 1.2977319329757616e-05,
"loss": 0.3189,
"step": 834
},
{
"epoch": 1.6378802747791954,
"grad_norm": 0.1018229680915178,
"learning_rate": 1.296243913391349e-05,
"loss": 0.3231,
"step": 835
},
{
"epoch": 1.6398429833169774,
"grad_norm": 0.1085023326981021,
"learning_rate": 1.2947551744109044e-05,
"loss": 0.3184,
"step": 836
},
{
"epoch": 1.6418056918547594,
"grad_norm": 0.10661650845272141,
"learning_rate": 1.2932657196496678e-05,
"loss": 0.3167,
"step": 837
},
{
"epoch": 1.6437684003925417,
"grad_norm": 0.11175562549636477,
"learning_rate": 1.2917755527246179e-05,
"loss": 0.3316,
"step": 838
},
{
"epoch": 1.645731108930324,
"grad_norm": 0.10523122873271762,
"learning_rate": 1.2902846772544625e-05,
"loss": 0.3312,
"step": 839
},
{
"epoch": 1.647693817468106,
"grad_norm": 0.13408743582159743,
"learning_rate": 1.28879309685963e-05,
"loss": 0.3279,
"step": 840
},
{
"epoch": 1.649656526005888,
"grad_norm": 0.4305626635372785,
"learning_rate": 1.2873008151622606e-05,
"loss": 0.3143,
"step": 841
},
{
"epoch": 1.6516192345436702,
"grad_norm": 0.10214474856379448,
"learning_rate": 1.2858078357861979e-05,
"loss": 0.3398,
"step": 842
},
{
"epoch": 1.6535819430814525,
"grad_norm": 0.19415909087113795,
"learning_rate": 1.2843141623569792e-05,
"loss": 0.3205,
"step": 843
},
{
"epoch": 1.6555446516192345,
"grad_norm": 0.11011953169021917,
"learning_rate": 1.2828197985018276e-05,
"loss": 0.3336,
"step": 844
},
{
"epoch": 1.6575073601570167,
"grad_norm": 0.1503839751469065,
"learning_rate": 1.2813247478496428e-05,
"loss": 0.3279,
"step": 845
},
{
"epoch": 1.659470068694799,
"grad_norm": 0.11781800325747578,
"learning_rate": 1.2798290140309924e-05,
"loss": 0.3254,
"step": 846
},
{
"epoch": 1.661432777232581,
"grad_norm": 0.10932301796870458,
"learning_rate": 1.2783326006781023e-05,
"loss": 0.3131,
"step": 847
},
{
"epoch": 1.663395485770363,
"grad_norm": 0.11978600321445199,
"learning_rate": 1.2768355114248493e-05,
"loss": 0.3277,
"step": 848
},
{
"epoch": 1.6653581943081452,
"grad_norm": 0.11039786376284047,
"learning_rate": 1.2753377499067522e-05,
"loss": 0.3522,
"step": 849
},
{
"epoch": 1.6673209028459275,
"grad_norm": 0.2497098566113429,
"learning_rate": 1.2738393197609602e-05,
"loss": 0.3297,
"step": 850
},
{
"epoch": 1.6692836113837095,
"grad_norm": 0.11865444749257295,
"learning_rate": 1.2723402246262484e-05,
"loss": 0.3271,
"step": 851
},
{
"epoch": 1.6712463199214915,
"grad_norm": 0.11095090763334178,
"learning_rate": 1.2708404681430054e-05,
"loss": 0.3243,
"step": 852
},
{
"epoch": 1.6732090284592738,
"grad_norm": 0.11867828722172012,
"learning_rate": 1.2693400539532263e-05,
"loss": 0.3141,
"step": 853
},
{
"epoch": 1.675171736997056,
"grad_norm": 0.1131423717161425,
"learning_rate": 1.2678389857005033e-05,
"loss": 0.3256,
"step": 854
},
{
"epoch": 1.677134445534838,
"grad_norm": 0.8893669616224853,
"learning_rate": 1.266337267030017e-05,
"loss": 0.3338,
"step": 855
},
{
"epoch": 1.67909715407262,
"grad_norm": 0.1621276184736285,
"learning_rate": 1.2648349015885272e-05,
"loss": 0.331,
"step": 856
},
{
"epoch": 1.6810598626104023,
"grad_norm": 0.12724899204009446,
"learning_rate": 1.2633318930243647e-05,
"loss": 0.3168,
"step": 857
},
{
"epoch": 1.6830225711481845,
"grad_norm": 0.15162789739847926,
"learning_rate": 1.2618282449874221e-05,
"loss": 0.3147,
"step": 858
},
{
"epoch": 1.6849852796859666,
"grad_norm": 0.14035812455971758,
"learning_rate": 1.2603239611291445e-05,
"loss": 0.3215,
"step": 859
},
{
"epoch": 1.6869479882237488,
"grad_norm": 0.46008864002942473,
"learning_rate": 1.2588190451025209e-05,
"loss": 0.325,
"step": 860
},
{
"epoch": 1.688910696761531,
"grad_norm": 0.15188638903617155,
"learning_rate": 1.2573135005620757e-05,
"loss": 0.3091,
"step": 861
},
{
"epoch": 1.690873405299313,
"grad_norm": 0.1372788892477886,
"learning_rate": 1.2558073311638604e-05,
"loss": 0.3313,
"step": 862
},
{
"epoch": 1.692836113837095,
"grad_norm": 0.1317279830573809,
"learning_rate": 1.2543005405654418e-05,
"loss": 0.3169,
"step": 863
},
{
"epoch": 1.6947988223748773,
"grad_norm": 0.13349361942014895,
"learning_rate": 1.2527931324258975e-05,
"loss": 0.3203,
"step": 864
},
{
"epoch": 1.6967615309126596,
"grad_norm": 0.12909951263188774,
"learning_rate": 1.2512851104058038e-05,
"loss": 0.3202,
"step": 865
},
{
"epoch": 1.6987242394504416,
"grad_norm": 0.2566617860841747,
"learning_rate": 1.249776478167227e-05,
"loss": 0.3295,
"step": 866
},
{
"epoch": 1.7006869479882236,
"grad_norm": 0.11969281942134942,
"learning_rate": 1.2482672393737164e-05,
"loss": 0.3342,
"step": 867
},
{
"epoch": 1.7026496565260059,
"grad_norm": 0.12518924207179494,
"learning_rate": 1.2467573976902936e-05,
"loss": 0.3333,
"step": 868
},
{
"epoch": 1.704612365063788,
"grad_norm": 0.21335666453466184,
"learning_rate": 1.2452469567834449e-05,
"loss": 0.3198,
"step": 869
},
{
"epoch": 1.7065750736015701,
"grad_norm": 0.12157902223714591,
"learning_rate": 1.2437359203211109e-05,
"loss": 0.308,
"step": 870
},
{
"epoch": 1.7085377821393521,
"grad_norm": 0.13207634498097479,
"learning_rate": 1.2422242919726786e-05,
"loss": 0.3338,
"step": 871
},
{
"epoch": 1.7105004906771346,
"grad_norm": 0.10700849261671908,
"learning_rate": 1.2407120754089733e-05,
"loss": 0.3221,
"step": 872
},
{
"epoch": 1.7124631992149166,
"grad_norm": 0.12604995554733597,
"learning_rate": 1.2391992743022472e-05,
"loss": 0.3165,
"step": 873
},
{
"epoch": 1.7144259077526987,
"grad_norm": 0.1613701947168364,
"learning_rate": 1.2376858923261732e-05,
"loss": 0.3217,
"step": 874
},
{
"epoch": 1.716388616290481,
"grad_norm": 0.1328873058860125,
"learning_rate": 1.2361719331558346e-05,
"loss": 0.3262,
"step": 875
},
{
"epoch": 1.7183513248282631,
"grad_norm": 0.11726688731965726,
"learning_rate": 1.2346574004677154e-05,
"loss": 0.3298,
"step": 876
},
{
"epoch": 1.7203140333660452,
"grad_norm": 0.1214011987184142,
"learning_rate": 1.2331422979396936e-05,
"loss": 0.337,
"step": 877
},
{
"epoch": 1.7222767419038272,
"grad_norm": 0.12820530343064107,
"learning_rate": 1.2316266292510305e-05,
"loss": 0.327,
"step": 878
},
{
"epoch": 1.7242394504416094,
"grad_norm": 0.11451384636314305,
"learning_rate": 1.2301103980823619e-05,
"loss": 0.3235,
"step": 879
},
{
"epoch": 1.7262021589793917,
"grad_norm": 0.10873652825630871,
"learning_rate": 1.2285936081156897e-05,
"loss": 0.311,
"step": 880
},
{
"epoch": 1.7281648675171737,
"grad_norm": 0.10624002895012581,
"learning_rate": 1.2270762630343734e-05,
"loss": 0.3245,
"step": 881
},
{
"epoch": 1.7301275760549557,
"grad_norm": 0.10425222382859203,
"learning_rate": 1.2255583665231196e-05,
"loss": 0.3211,
"step": 882
},
{
"epoch": 1.732090284592738,
"grad_norm": 0.10943316735201042,
"learning_rate": 1.2240399222679747e-05,
"loss": 0.3166,
"step": 883
},
{
"epoch": 1.7340529931305202,
"grad_norm": 0.108007053530337,
"learning_rate": 1.2225209339563144e-05,
"loss": 0.3271,
"step": 884
},
{
"epoch": 1.7360157016683022,
"grad_norm": 0.1806278564221946,
"learning_rate": 1.221001405276837e-05,
"loss": 0.3342,
"step": 885
},
{
"epoch": 1.7379784102060842,
"grad_norm": 0.11217999203542753,
"learning_rate": 1.2194813399195518e-05,
"loss": 0.3174,
"step": 886
},
{
"epoch": 1.7399411187438667,
"grad_norm": 0.11446924843967385,
"learning_rate": 1.217960741575771e-05,
"loss": 0.3375,
"step": 887
},
{
"epoch": 1.7419038272816487,
"grad_norm": 0.11052403929142486,
"learning_rate": 1.2164396139381029e-05,
"loss": 0.3207,
"step": 888
},
{
"epoch": 1.7438665358194307,
"grad_norm": 0.12041833183157077,
"learning_rate": 1.2149179607004396e-05,
"loss": 0.3215,
"step": 889
},
{
"epoch": 1.745829244357213,
"grad_norm": 0.11164358188191847,
"learning_rate": 1.2133957855579501e-05,
"loss": 0.3124,
"step": 890
},
{
"epoch": 1.7477919528949952,
"grad_norm": 0.10293795446688632,
"learning_rate": 1.2118730922070707e-05,
"loss": 0.323,
"step": 891
},
{
"epoch": 1.7497546614327772,
"grad_norm": 0.10970003809084786,
"learning_rate": 1.210349884345496e-05,
"loss": 0.3267,
"step": 892
},
{
"epoch": 1.7517173699705593,
"grad_norm": 0.10828051591336894,
"learning_rate": 1.20882616567217e-05,
"loss": 0.329,
"step": 893
},
{
"epoch": 1.7536800785083415,
"grad_norm": 0.11148013332448067,
"learning_rate": 1.2073019398872778e-05,
"loss": 0.3201,
"step": 894
},
{
"epoch": 1.7556427870461238,
"grad_norm": 0.1154913650725476,
"learning_rate": 1.205777210692235e-05,
"loss": 0.3318,
"step": 895
},
{
"epoch": 1.7576054955839058,
"grad_norm": 0.10882465073502968,
"learning_rate": 1.2042519817896805e-05,
"loss": 0.3311,
"step": 896
},
{
"epoch": 1.7576054955839058,
"eval_loss": 0.4246857762336731,
"eval_runtime": 245.6728,
"eval_samples_per_second": 123.551,
"eval_steps_per_second": 3.863,
"step": 896
},
{
"epoch": 1.7595682041216878,
"grad_norm": 0.1164747664318698,
"learning_rate": 1.202726256883466e-05,
"loss": 0.317,
"step": 897
},
{
"epoch": 1.76153091265947,
"grad_norm": 0.11482257791759276,
"learning_rate": 1.2012000396786485e-05,
"loss": 0.3361,
"step": 898
},
{
"epoch": 1.7634936211972523,
"grad_norm": 0.10646048917975168,
"learning_rate": 1.1996733338814795e-05,
"loss": 0.3241,
"step": 899
},
{
"epoch": 1.7654563297350343,
"grad_norm": 0.10984226703740471,
"learning_rate": 1.1981461431993978e-05,
"loss": 0.3133,
"step": 900
},
{
"epoch": 1.7674190382728163,
"grad_norm": 0.1047308434934182,
"learning_rate": 1.1966184713410192e-05,
"loss": 0.3335,
"step": 901
},
{
"epoch": 1.7693817468105988,
"grad_norm": 0.10425069083553949,
"learning_rate": 1.1950903220161286e-05,
"loss": 0.3226,
"step": 902
},
{
"epoch": 1.7713444553483808,
"grad_norm": 0.10353476391750117,
"learning_rate": 1.1935616989356693e-05,
"loss": 0.3341,
"step": 903
},
{
"epoch": 1.7733071638861628,
"grad_norm": 0.13560691295515553,
"learning_rate": 1.1920326058117364e-05,
"loss": 0.3233,
"step": 904
},
{
"epoch": 1.775269872423945,
"grad_norm": 0.10178800675902581,
"learning_rate": 1.190503046357565e-05,
"loss": 0.3167,
"step": 905
},
{
"epoch": 1.7772325809617273,
"grad_norm": 0.21889157131254638,
"learning_rate": 1.1889730242875243e-05,
"loss": 0.3422,
"step": 906
},
{
"epoch": 1.7791952894995093,
"grad_norm": 0.10614581871424822,
"learning_rate": 1.1874425433171055e-05,
"loss": 0.309,
"step": 907
},
{
"epoch": 1.7811579980372914,
"grad_norm": 0.10365379424748211,
"learning_rate": 1.1859116071629148e-05,
"loss": 0.3137,
"step": 908
},
{
"epoch": 1.7831207065750736,
"grad_norm": 0.10348421216023944,
"learning_rate": 1.1843802195426634e-05,
"loss": 0.3225,
"step": 909
},
{
"epoch": 1.7850834151128558,
"grad_norm": 0.1060682052762049,
"learning_rate": 1.1828483841751597e-05,
"loss": 0.3139,
"step": 910
},
{
"epoch": 1.7870461236506379,
"grad_norm": 0.11042414275132055,
"learning_rate": 1.1813161047802986e-05,
"loss": 0.35,
"step": 911
},
{
"epoch": 1.7890088321884199,
"grad_norm": 0.11161988036910164,
"learning_rate": 1.1797833850790527e-05,
"loss": 0.3313,
"step": 912
},
{
"epoch": 1.7909715407262021,
"grad_norm": 0.11073431733692736,
"learning_rate": 1.1782502287934659e-05,
"loss": 0.3326,
"step": 913
},
{
"epoch": 1.7929342492639844,
"grad_norm": 0.10914796174284935,
"learning_rate": 1.1767166396466404e-05,
"loss": 0.3177,
"step": 914
},
{
"epoch": 1.7948969578017664,
"grad_norm": 0.10465459851626065,
"learning_rate": 1.1751826213627297e-05,
"loss": 0.3266,
"step": 915
},
{
"epoch": 1.7968596663395484,
"grad_norm": 0.10041864035051187,
"learning_rate": 1.1736481776669307e-05,
"loss": 0.3213,
"step": 916
},
{
"epoch": 1.7988223748773309,
"grad_norm": 0.11485845865027179,
"learning_rate": 1.172113312285472e-05,
"loss": 0.3265,
"step": 917
},
{
"epoch": 1.800785083415113,
"grad_norm": 0.12033881568016211,
"learning_rate": 1.1705780289456069e-05,
"loss": 0.3197,
"step": 918
},
{
"epoch": 1.802747791952895,
"grad_norm": 0.11332046772222625,
"learning_rate": 1.1690423313756037e-05,
"loss": 0.3139,
"step": 919
},
{
"epoch": 1.8047105004906772,
"grad_norm": 0.10473876200254037,
"learning_rate": 1.1675062233047365e-05,
"loss": 0.3281,
"step": 920
},
{
"epoch": 1.8066732090284594,
"grad_norm": 0.1184964198201036,
"learning_rate": 1.165969708463276e-05,
"loss": 0.3223,
"step": 921
},
{
"epoch": 1.8086359175662414,
"grad_norm": 0.12474305158961706,
"learning_rate": 1.1644327905824808e-05,
"loss": 0.3214,
"step": 922
},
{
"epoch": 1.8105986261040234,
"grad_norm": 0.09933741044605185,
"learning_rate": 1.162895473394589e-05,
"loss": 0.33,
"step": 923
},
{
"epoch": 1.8125613346418057,
"grad_norm": 0.25775919840623956,
"learning_rate": 1.1613577606328068e-05,
"loss": 0.3202,
"step": 924
},
{
"epoch": 1.814524043179588,
"grad_norm": 0.10094921766955992,
"learning_rate": 1.1598196560313024e-05,
"loss": 0.3092,
"step": 925
},
{
"epoch": 1.81648675171737,
"grad_norm": 0.11514280556001413,
"learning_rate": 1.1582811633251949e-05,
"loss": 0.3109,
"step": 926
},
{
"epoch": 1.818449460255152,
"grad_norm": 0.13680270431298738,
"learning_rate": 1.1567422862505465e-05,
"loss": 0.3275,
"step": 927
},
{
"epoch": 1.8204121687929342,
"grad_norm": 0.10821477293958764,
"learning_rate": 1.1552030285443516e-05,
"loss": 0.3185,
"step": 928
},
{
"epoch": 1.8223748773307165,
"grad_norm": 0.1020483099705299,
"learning_rate": 1.1536633939445302e-05,
"loss": 0.3216,
"step": 929
},
{
"epoch": 1.8243375858684985,
"grad_norm": 0.11310324846120763,
"learning_rate": 1.1521233861899168e-05,
"loss": 0.312,
"step": 930
},
{
"epoch": 1.8263002944062807,
"grad_norm": 0.10125352819285569,
"learning_rate": 1.1505830090202524e-05,
"loss": 0.3202,
"step": 931
},
{
"epoch": 1.828263002944063,
"grad_norm": 0.10518388367230969,
"learning_rate": 1.1490422661761744e-05,
"loss": 0.3161,
"step": 932
},
{
"epoch": 1.830225711481845,
"grad_norm": 0.10596564037766877,
"learning_rate": 1.1475011613992097e-05,
"loss": 0.3208,
"step": 933
},
{
"epoch": 1.832188420019627,
"grad_norm": 0.10832549491208261,
"learning_rate": 1.1459596984317622e-05,
"loss": 0.3198,
"step": 934
},
{
"epoch": 1.8341511285574092,
"grad_norm": 0.10340633509154877,
"learning_rate": 1.1444178810171074e-05,
"loss": 0.3347,
"step": 935
},
{
"epoch": 1.8361138370951915,
"grad_norm": 0.11187292516624683,
"learning_rate": 1.1428757128993801e-05,
"loss": 0.3159,
"step": 936
},
{
"epoch": 1.8380765456329735,
"grad_norm": 0.10201292596555332,
"learning_rate": 1.1413331978235677e-05,
"loss": 0.3166,
"step": 937
},
{
"epoch": 1.8400392541707555,
"grad_norm": 0.10804358781472774,
"learning_rate": 1.1397903395354996e-05,
"loss": 0.323,
"step": 938
},
{
"epoch": 1.8420019627085378,
"grad_norm": 0.10468656385490047,
"learning_rate": 1.138247141781839e-05,
"loss": 0.3185,
"step": 939
},
{
"epoch": 1.84396467124632,
"grad_norm": 0.21138099918352082,
"learning_rate": 1.1367036083100735e-05,
"loss": 0.3321,
"step": 940
},
{
"epoch": 1.845927379784102,
"grad_norm": 0.1007451545158691,
"learning_rate": 1.1351597428685055e-05,
"loss": 0.3201,
"step": 941
},
{
"epoch": 1.847890088321884,
"grad_norm": 0.10166102006184587,
"learning_rate": 1.1336155492062439e-05,
"loss": 0.3217,
"step": 942
},
{
"epoch": 1.8498527968596663,
"grad_norm": 0.10712467451663077,
"learning_rate": 1.132071031073195e-05,
"loss": 0.3207,
"step": 943
},
{
"epoch": 1.8518155053974485,
"grad_norm": 0.11530938712137909,
"learning_rate": 1.130526192220052e-05,
"loss": 0.3265,
"step": 944
},
{
"epoch": 1.8537782139352306,
"grad_norm": 0.102223465331792,
"learning_rate": 1.1289810363982875e-05,
"loss": 0.3109,
"step": 945
},
{
"epoch": 1.8557409224730128,
"grad_norm": 0.10435635169915954,
"learning_rate": 1.1274355673601446e-05,
"loss": 0.3215,
"step": 946
},
{
"epoch": 1.857703631010795,
"grad_norm": 0.10323000509883015,
"learning_rate": 1.1258897888586256e-05,
"loss": 0.3245,
"step": 947
},
{
"epoch": 1.859666339548577,
"grad_norm": 0.2356211384064762,
"learning_rate": 1.1243437046474854e-05,
"loss": 0.325,
"step": 948
},
{
"epoch": 1.861629048086359,
"grad_norm": 0.09975701738920462,
"learning_rate": 1.1227973184812207e-05,
"loss": 0.3125,
"step": 949
},
{
"epoch": 1.8635917566241413,
"grad_norm": 0.10475942757691821,
"learning_rate": 1.1212506341150615e-05,
"loss": 0.3168,
"step": 950
},
{
"epoch": 1.8655544651619236,
"grad_norm": 0.4497324410418576,
"learning_rate": 1.1197036553049626e-05,
"loss": 0.3325,
"step": 951
},
{
"epoch": 1.8675171736997056,
"grad_norm": 0.10885422432097863,
"learning_rate": 1.118156385807593e-05,
"loss": 0.3134,
"step": 952
},
{
"epoch": 1.8694798822374876,
"grad_norm": 0.9319115703562274,
"learning_rate": 1.1166088293803276e-05,
"loss": 0.3329,
"step": 953
},
{
"epoch": 1.8714425907752699,
"grad_norm": 0.1123785787687025,
"learning_rate": 1.1150609897812387e-05,
"loss": 0.311,
"step": 954
},
{
"epoch": 1.873405299313052,
"grad_norm": 0.10319256076853457,
"learning_rate": 1.1135128707690862e-05,
"loss": 0.3156,
"step": 955
},
{
"epoch": 1.8753680078508341,
"grad_norm": 0.7095695854915799,
"learning_rate": 1.1119644761033079e-05,
"loss": 0.3579,
"step": 956
},
{
"epoch": 1.8773307163886161,
"grad_norm": 0.167595727369702,
"learning_rate": 1.1104158095440115e-05,
"loss": 0.3277,
"step": 957
},
{
"epoch": 1.8792934249263984,
"grad_norm": 0.270060302607285,
"learning_rate": 1.1088668748519646e-05,
"loss": 0.3242,
"step": 958
},
{
"epoch": 1.8812561334641806,
"grad_norm": 0.15698959445501914,
"learning_rate": 1.1073176757885866e-05,
"loss": 0.361,
"step": 959
},
{
"epoch": 1.8832188420019627,
"grad_norm": 0.3221946038337756,
"learning_rate": 1.105768216115938e-05,
"loss": 0.3188,
"step": 960
},
{
"epoch": 1.885181550539745,
"grad_norm": 0.1406486088615483,
"learning_rate": 1.1042184995967127e-05,
"loss": 0.3312,
"step": 961
},
{
"epoch": 1.8871442590775271,
"grad_norm": 0.13397810433694357,
"learning_rate": 1.1026685299942286e-05,
"loss": 0.3261,
"step": 962
},
{
"epoch": 1.8891069676153092,
"grad_norm": 1.4841344998774781,
"learning_rate": 1.1011183110724173e-05,
"loss": 0.3272,
"step": 963
},
{
"epoch": 1.8910696761530912,
"grad_norm": 0.1659981257856718,
"learning_rate": 1.0995678465958168e-05,
"loss": 0.3288,
"step": 964
},
{
"epoch": 1.8930323846908734,
"grad_norm": 0.16778967117736607,
"learning_rate": 1.098017140329561e-05,
"loss": 0.3232,
"step": 965
},
{
"epoch": 1.8949950932286557,
"grad_norm": 0.21080374151960482,
"learning_rate": 1.0964661960393703e-05,
"loss": 0.3274,
"step": 966
},
{
"epoch": 1.8969578017664377,
"grad_norm": 0.16463715782689323,
"learning_rate": 1.0949150174915441e-05,
"loss": 0.3352,
"step": 967
},
{
"epoch": 1.8989205103042197,
"grad_norm": 0.2267681805465417,
"learning_rate": 1.0933636084529507e-05,
"loss": 0.3388,
"step": 968
},
{
"epoch": 1.900883218842002,
"grad_norm": 0.14488446998635424,
"learning_rate": 1.0918119726910175e-05,
"loss": 0.3202,
"step": 969
},
{
"epoch": 1.9028459273797842,
"grad_norm": 0.18140470030881337,
"learning_rate": 1.0902601139737225e-05,
"loss": 0.3329,
"step": 970
},
{
"epoch": 1.9048086359175662,
"grad_norm": 0.1546698941699075,
"learning_rate": 1.0887080360695855e-05,
"loss": 0.3306,
"step": 971
},
{
"epoch": 1.9067713444553482,
"grad_norm": 1.2221913272883775,
"learning_rate": 1.0871557427476585e-05,
"loss": 0.3399,
"step": 972
},
{
"epoch": 1.9087340529931305,
"grad_norm": 0.15426601384648553,
"learning_rate": 1.0856032377775161e-05,
"loss": 0.3358,
"step": 973
},
{
"epoch": 1.9106967615309127,
"grad_norm": 1.1114760751159305,
"learning_rate": 1.0840505249292477e-05,
"loss": 0.3356,
"step": 974
},
{
"epoch": 1.9126594700686947,
"grad_norm": 0.22900624172519007,
"learning_rate": 1.0824976079734472e-05,
"loss": 0.3134,
"step": 975
},
{
"epoch": 1.914622178606477,
"grad_norm": 0.9459775335785819,
"learning_rate": 1.0809444906812034e-05,
"loss": 0.3235,
"step": 976
},
{
"epoch": 1.9165848871442592,
"grad_norm": 0.17887149210801348,
"learning_rate": 1.079391176824093e-05,
"loss": 0.3304,
"step": 977
},
{
"epoch": 1.9185475956820413,
"grad_norm": 0.17484413383294636,
"learning_rate": 1.0778376701741688e-05,
"loss": 0.3298,
"step": 978
},
{
"epoch": 1.9205103042198233,
"grad_norm": 0.15868896830006096,
"learning_rate": 1.0762839745039526e-05,
"loss": 0.325,
"step": 979
},
{
"epoch": 1.9224730127576055,
"grad_norm": 0.13952635278363976,
"learning_rate": 1.0747300935864245e-05,
"loss": 0.3246,
"step": 980
},
{
"epoch": 1.9244357212953878,
"grad_norm": 0.14676838005524964,
"learning_rate": 1.073176031195015e-05,
"loss": 0.3367,
"step": 981
},
{
"epoch": 1.9263984298331698,
"grad_norm": 0.1393769904490578,
"learning_rate": 1.0716217911035952e-05,
"loss": 0.331,
"step": 982
},
{
"epoch": 1.9283611383709518,
"grad_norm": 0.12073516486793993,
"learning_rate": 1.0700673770864673e-05,
"loss": 0.3215,
"step": 983
},
{
"epoch": 1.930323846908734,
"grad_norm": 0.17261593509801376,
"learning_rate": 1.0685127929183567e-05,
"loss": 0.3377,
"step": 984
},
{
"epoch": 1.9322865554465163,
"grad_norm": 0.1288861939566597,
"learning_rate": 1.0669580423744014e-05,
"loss": 0.3268,
"step": 985
},
{
"epoch": 1.9342492639842983,
"grad_norm": 0.12242177876640145,
"learning_rate": 1.0654031292301432e-05,
"loss": 0.3331,
"step": 986
},
{
"epoch": 1.9362119725220803,
"grad_norm": 0.135438680122695,
"learning_rate": 1.063848057261519e-05,
"loss": 0.3276,
"step": 987
},
{
"epoch": 1.9381746810598626,
"grad_norm": 0.12113837642836334,
"learning_rate": 1.0622928302448523e-05,
"loss": 0.3196,
"step": 988
},
{
"epoch": 1.9401373895976448,
"grad_norm": 0.117144755644724,
"learning_rate": 1.0607374519568412e-05,
"loss": 0.3247,
"step": 989
},
{
"epoch": 1.9421000981354268,
"grad_norm": 0.11697508630127905,
"learning_rate": 1.0591819261745528e-05,
"loss": 0.3223,
"step": 990
},
{
"epoch": 1.944062806673209,
"grad_norm": 0.11238405496608947,
"learning_rate": 1.0576262566754121e-05,
"loss": 0.3212,
"step": 991
},
{
"epoch": 1.9460255152109913,
"grad_norm": 0.1139647888181511,
"learning_rate": 1.0560704472371919e-05,
"loss": 0.3253,
"step": 992
},
{
"epoch": 1.9479882237487733,
"grad_norm": 0.11466955150117668,
"learning_rate": 1.0545145016380065e-05,
"loss": 0.3273,
"step": 993
},
{
"epoch": 1.9499509322865554,
"grad_norm": 0.10970406400371188,
"learning_rate": 1.0529584236562995e-05,
"loss": 0.3254,
"step": 994
},
{
"epoch": 1.9519136408243376,
"grad_norm": 0.11114557334217089,
"learning_rate": 1.0514022170708374e-05,
"loss": 0.3246,
"step": 995
},
{
"epoch": 1.9538763493621198,
"grad_norm": 0.11801065397328833,
"learning_rate": 1.0498458856606972e-05,
"loss": 0.3294,
"step": 996
},
{
"epoch": 1.9558390578999019,
"grad_norm": 0.10405757613653188,
"learning_rate": 1.0482894332052607e-05,
"loss": 0.3261,
"step": 997
},
{
"epoch": 1.9578017664376839,
"grad_norm": 0.1103740005321899,
"learning_rate": 1.0467328634842024e-05,
"loss": 0.3289,
"step": 998
},
{
"epoch": 1.9597644749754661,
"grad_norm": 0.09987852489854114,
"learning_rate": 1.0451761802774824e-05,
"loss": 0.3172,
"step": 999
},
{
"epoch": 1.9617271835132484,
"grad_norm": 0.10893453828326174,
"learning_rate": 1.0436193873653362e-05,
"loss": 0.3228,
"step": 1000
},
{
"epoch": 1.9636898920510304,
"grad_norm": 0.10174600038711792,
"learning_rate": 1.0420624885282653e-05,
"loss": 0.3225,
"step": 1001
},
{
"epoch": 1.9656526005888124,
"grad_norm": 0.11640082237562227,
"learning_rate": 1.0405054875470287e-05,
"loss": 0.3193,
"step": 1002
},
{
"epoch": 1.9676153091265947,
"grad_norm": 0.15766357183117252,
"learning_rate": 1.0389483882026334e-05,
"loss": 0.3216,
"step": 1003
},
{
"epoch": 1.969578017664377,
"grad_norm": 0.10997358802993172,
"learning_rate": 1.037391194276326e-05,
"loss": 0.3142,
"step": 1004
},
{
"epoch": 1.971540726202159,
"grad_norm": 0.10152791127501247,
"learning_rate": 1.0358339095495811e-05,
"loss": 0.3169,
"step": 1005
},
{
"epoch": 1.9735034347399412,
"grad_norm": 0.20219624773435688,
"learning_rate": 1.0342765378040953e-05,
"loss": 0.324,
"step": 1006
},
{
"epoch": 1.9754661432777234,
"grad_norm": 0.10297719458762343,
"learning_rate": 1.0327190828217763e-05,
"loss": 0.3181,
"step": 1007
},
{
"epoch": 1.9774288518155054,
"grad_norm": 0.11052971601307274,
"learning_rate": 1.0311615483847333e-05,
"loss": 0.3245,
"step": 1008
},
{
"epoch": 1.9793915603532874,
"grad_norm": 0.09914734230876245,
"learning_rate": 1.0296039382752687e-05,
"loss": 0.324,
"step": 1009
},
{
"epoch": 1.9813542688910697,
"grad_norm": 0.10159497435434399,
"learning_rate": 1.028046256275869e-05,
"loss": 0.3119,
"step": 1010
},
{
"epoch": 1.983316977428852,
"grad_norm": 0.11056760688752829,
"learning_rate": 1.0264885061691954e-05,
"loss": 0.323,
"step": 1011
},
{
"epoch": 1.985279685966634,
"grad_norm": 0.10804474565070168,
"learning_rate": 1.0249306917380731e-05,
"loss": 0.3113,
"step": 1012
},
{
"epoch": 1.987242394504416,
"grad_norm": 0.13940608144653047,
"learning_rate": 1.023372816765485e-05,
"loss": 0.3123,
"step": 1013
},
{
"epoch": 1.9892051030421982,
"grad_norm": 0.10238052023308335,
"learning_rate": 1.0218148850345613e-05,
"loss": 0.3235,
"step": 1014
},
{
"epoch": 1.9911678115799805,
"grad_norm": 0.12527000090641732,
"learning_rate": 1.0202569003285683e-05,
"loss": 0.3337,
"step": 1015
},
{
"epoch": 1.9931305201177625,
"grad_norm": 0.11514202583090338,
"learning_rate": 1.0186988664309023e-05,
"loss": 0.333,
"step": 1016
},
{
"epoch": 1.9950932286555445,
"grad_norm": 0.11172112347247365,
"learning_rate": 1.017140787125079e-05,
"loss": 0.3227,
"step": 1017
},
{
"epoch": 1.9970559371933267,
"grad_norm": 0.15649181628225953,
"learning_rate": 1.0155826661947232e-05,
"loss": 0.3208,
"step": 1018
}
],
"logging_steps": 1,
"max_steps": 2036,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 509,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 6.450464370500567e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}