{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9796616612811253, "eval_steps": 500, "global_step": 1314, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001520623455616803, "grad_norm": 3.80404000128578, "learning_rate": 9.033333333333334e-07, "loss": 1.3234, "step": 1 }, { "epoch": 0.003041246911233606, "grad_norm": 3.7593589398056593, "learning_rate": 1.8066666666666668e-06, "loss": 1.2935, "step": 2 }, { "epoch": 0.0045618703668504085, "grad_norm": 3.5532664758635017, "learning_rate": 2.7100000000000003e-06, "loss": 1.2481, "step": 3 }, { "epoch": 0.006082493822467212, "grad_norm": 3.48298105333948, "learning_rate": 3.6133333333333336e-06, "loss": 1.2786, "step": 4 }, { "epoch": 0.0076031172780840145, "grad_norm": 2.7761816462967768, "learning_rate": 4.516666666666667e-06, "loss": 1.233, "step": 5 }, { "epoch": 0.009123740733700817, "grad_norm": 1.7543427633433306, "learning_rate": 5.420000000000001e-06, "loss": 1.2459, "step": 6 }, { "epoch": 0.01064436418931762, "grad_norm": 1.6976304839039194, "learning_rate": 6.3233333333333335e-06, "loss": 1.1905, "step": 7 }, { "epoch": 0.012164987644934424, "grad_norm": 3.06752449258415, "learning_rate": 7.226666666666667e-06, "loss": 1.1417, "step": 8 }, { "epoch": 0.013685611100551226, "grad_norm": 3.445950798936361, "learning_rate": 8.13e-06, "loss": 1.1746, "step": 9 }, { "epoch": 0.015206234556168029, "grad_norm": 3.088218157818855, "learning_rate": 9.033333333333334e-06, "loss": 1.1209, "step": 10 }, { "epoch": 0.01672685801178483, "grad_norm": 3.3669650804349973, "learning_rate": 9.936666666666666e-06, "loss": 1.138, "step": 11 }, { "epoch": 0.018247481467401634, "grad_norm": 3.1041727231133174, "learning_rate": 1.0840000000000001e-05, "loss": 1.1188, "step": 12 }, { "epoch": 0.019768104923018438, "grad_norm": 2.21205722327462, "learning_rate": 1.1743333333333335e-05, "loss": 1.0873, "step": 13 }, { "epoch": 0.02128872837863524, "grad_norm": 1.32813912161683, "learning_rate": 1.2646666666666667e-05, "loss": 1.1163, "step": 14 }, { "epoch": 0.022809351834252044, "grad_norm": 2.3535108175852097, "learning_rate": 1.355e-05, "loss": 1.1189, "step": 15 }, { "epoch": 0.024329975289868848, "grad_norm": 2.8032523790378274, "learning_rate": 1.4453333333333334e-05, "loss": 1.0621, "step": 16 }, { "epoch": 0.025850598745485648, "grad_norm": 1.8636564282922823, "learning_rate": 1.5356666666666668e-05, "loss": 1.0059, "step": 17 }, { "epoch": 0.02737122220110245, "grad_norm": 1.198430471743954, "learning_rate": 1.626e-05, "loss": 1.0572, "step": 18 }, { "epoch": 0.028891845656719255, "grad_norm": 1.3874306913597059, "learning_rate": 1.7163333333333332e-05, "loss": 1.024, "step": 19 }, { "epoch": 0.030412469112336058, "grad_norm": 1.1392396226185426, "learning_rate": 1.8066666666666668e-05, "loss": 1.0426, "step": 20 }, { "epoch": 0.03193309256795286, "grad_norm": 0.7881813145609246, "learning_rate": 1.897e-05, "loss": 1.0382, "step": 21 }, { "epoch": 0.03345371602356966, "grad_norm": 1.0470837153138448, "learning_rate": 1.987333333333333e-05, "loss": 0.9916, "step": 22 }, { "epoch": 0.03497433947918647, "grad_norm": 0.9265643204523712, "learning_rate": 2.077666666666667e-05, "loss": 1.0348, "step": 23 }, { "epoch": 0.03649496293480327, "grad_norm": 0.8620032399568974, "learning_rate": 2.1680000000000002e-05, "loss": 0.9893, "step": 24 }, { "epoch": 0.038015586390420075, "grad_norm": 0.963710426135712, "learning_rate": 2.2583333333333335e-05, "loss": 0.9931, "step": 25 }, { "epoch": 0.039536209846036875, "grad_norm": 0.7104087725888792, "learning_rate": 2.348666666666667e-05, "loss": 0.9723, "step": 26 }, { "epoch": 0.041056833301653675, "grad_norm": 0.8206318828097199, "learning_rate": 2.4390000000000002e-05, "loss": 1.0048, "step": 27 }, { "epoch": 0.04257745675727048, "grad_norm": 0.88460359130528, "learning_rate": 2.5293333333333334e-05, "loss": 0.9686, "step": 28 }, { "epoch": 0.04409808021288728, "grad_norm": 0.6842646410238233, "learning_rate": 2.619666666666667e-05, "loss": 0.9942, "step": 29 }, { "epoch": 0.04561870366850409, "grad_norm": 0.9420397284384021, "learning_rate": 2.71e-05, "loss": 0.9619, "step": 30 }, { "epoch": 0.04713932712412089, "grad_norm": 0.7311688429092981, "learning_rate": 2.7099959441789884e-05, "loss": 0.9722, "step": 31 }, { "epoch": 0.048659950579737696, "grad_norm": 0.8897905291941917, "learning_rate": 2.7099837767402332e-05, "loss": 0.9767, "step": 32 }, { "epoch": 0.050180574035354496, "grad_norm": 0.7198270553694668, "learning_rate": 2.709963497756574e-05, "loss": 0.9209, "step": 33 }, { "epoch": 0.051701197490971296, "grad_norm": 0.6437197076975016, "learning_rate": 2.7099351073494104e-05, "loss": 0.954, "step": 34 }, { "epoch": 0.0532218209465881, "grad_norm": 0.6539693647049317, "learning_rate": 2.7098986056886998e-05, "loss": 0.9281, "step": 35 }, { "epoch": 0.0547424444022049, "grad_norm": 0.7438647990016182, "learning_rate": 2.7098539929929583e-05, "loss": 0.9198, "step": 36 }, { "epoch": 0.05626306785782171, "grad_norm": 0.6710026252513194, "learning_rate": 2.7098012695292568e-05, "loss": 0.9535, "step": 37 }, { "epoch": 0.05778369131343851, "grad_norm": 0.7336846915639782, "learning_rate": 2.709740435613222e-05, "loss": 0.9188, "step": 38 }, { "epoch": 0.05930431476905531, "grad_norm": 0.7107571686807039, "learning_rate": 2.709671491609034e-05, "loss": 0.9464, "step": 39 }, { "epoch": 0.060824938224672116, "grad_norm": 0.6815590773111158, "learning_rate": 2.7095944379294215e-05, "loss": 0.9652, "step": 40 }, { "epoch": 0.062345561680288916, "grad_norm": 0.7639644922873411, "learning_rate": 2.709509275035663e-05, "loss": 0.9305, "step": 41 }, { "epoch": 0.06386618513590572, "grad_norm": 0.6137196824616831, "learning_rate": 2.709416003437583e-05, "loss": 0.8947, "step": 42 }, { "epoch": 0.06538680859152253, "grad_norm": 0.730883182696482, "learning_rate": 2.709314623693546e-05, "loss": 0.9284, "step": 43 }, { "epoch": 0.06690743204713932, "grad_norm": 0.6378775889130142, "learning_rate": 2.7092051364104584e-05, "loss": 0.9319, "step": 44 }, { "epoch": 0.06842805550275613, "grad_norm": 0.6881392029525518, "learning_rate": 2.709087542243759e-05, "loss": 0.9561, "step": 45 }, { "epoch": 0.06994867895837294, "grad_norm": 0.7653476020607418, "learning_rate": 2.708961841897421e-05, "loss": 0.9147, "step": 46 }, { "epoch": 0.07146930241398973, "grad_norm": 0.6329992439002645, "learning_rate": 2.7088280361239425e-05, "loss": 0.9097, "step": 47 }, { "epoch": 0.07298992586960654, "grad_norm": 0.7176283996505511, "learning_rate": 2.7086861257243455e-05, "loss": 0.923, "step": 48 }, { "epoch": 0.07451054932522334, "grad_norm": 0.5868244553340308, "learning_rate": 2.7085361115481697e-05, "loss": 0.9099, "step": 49 }, { "epoch": 0.07603117278084015, "grad_norm": 0.6598413586849197, "learning_rate": 2.7083779944934685e-05, "loss": 0.9198, "step": 50 }, { "epoch": 0.07755179623645694, "grad_norm": 0.5578106414522218, "learning_rate": 2.7082117755068008e-05, "loss": 0.8798, "step": 51 }, { "epoch": 0.07907241969207375, "grad_norm": 0.5382560015493899, "learning_rate": 2.708037455583229e-05, "loss": 0.9423, "step": 52 }, { "epoch": 0.08059304314769056, "grad_norm": 0.7104523737465288, "learning_rate": 2.7078550357663116e-05, "loss": 0.8525, "step": 53 }, { "epoch": 0.08211366660330735, "grad_norm": 0.6283073524726329, "learning_rate": 2.7076645171480954e-05, "loss": 0.8691, "step": 54 }, { "epoch": 0.08363429005892416, "grad_norm": 0.6650416502799579, "learning_rate": 2.7074659008691105e-05, "loss": 0.867, "step": 55 }, { "epoch": 0.08515491351454096, "grad_norm": 0.7744652466782505, "learning_rate": 2.707259188118364e-05, "loss": 0.9067, "step": 56 }, { "epoch": 0.08667553697015777, "grad_norm": 0.8615866504032589, "learning_rate": 2.7070443801333323e-05, "loss": 0.888, "step": 57 }, { "epoch": 0.08819616042577456, "grad_norm": 0.7552923225662669, "learning_rate": 2.706821478199952e-05, "loss": 0.9394, "step": 58 }, { "epoch": 0.08971678388139137, "grad_norm": 0.8827458802779125, "learning_rate": 2.706590483652616e-05, "loss": 0.9333, "step": 59 }, { "epoch": 0.09123740733700818, "grad_norm": 0.5735989763886916, "learning_rate": 2.7063513978741612e-05, "loss": 0.9045, "step": 60 }, { "epoch": 0.09275803079262497, "grad_norm": 0.935081852925912, "learning_rate": 2.706104222295863e-05, "loss": 0.9103, "step": 61 }, { "epoch": 0.09427865424824178, "grad_norm": 0.974887379533721, "learning_rate": 2.7058489583974263e-05, "loss": 0.8588, "step": 62 }, { "epoch": 0.09579927770385858, "grad_norm": 0.8802808107178965, "learning_rate": 2.7055856077069762e-05, "loss": 0.9196, "step": 63 }, { "epoch": 0.09731990115947539, "grad_norm": 0.7021620798712934, "learning_rate": 2.7053141718010486e-05, "loss": 0.891, "step": 64 }, { "epoch": 0.09884052461509218, "grad_norm": 1.088284062839966, "learning_rate": 2.7050346523045816e-05, "loss": 0.8944, "step": 65 }, { "epoch": 0.10036114807070899, "grad_norm": 0.6749292391517889, "learning_rate": 2.7047470508909053e-05, "loss": 0.8815, "step": 66 }, { "epoch": 0.1018817715263258, "grad_norm": 0.8723921836119483, "learning_rate": 2.704451369281731e-05, "loss": 0.859, "step": 67 }, { "epoch": 0.10340239498194259, "grad_norm": 0.7760246702094213, "learning_rate": 2.7041476092471437e-05, "loss": 0.9184, "step": 68 }, { "epoch": 0.1049230184375594, "grad_norm": 0.7139493353356939, "learning_rate": 2.7038357726055864e-05, "loss": 0.8692, "step": 69 }, { "epoch": 0.1064436418931762, "grad_norm": 0.6739017082503604, "learning_rate": 2.7035158612238555e-05, "loss": 0.8609, "step": 70 }, { "epoch": 0.10796426534879301, "grad_norm": 0.7433509963897333, "learning_rate": 2.7031878770170844e-05, "loss": 0.9063, "step": 71 }, { "epoch": 0.1094848888044098, "grad_norm": 0.6089113038384066, "learning_rate": 2.7028518219487355e-05, "loss": 0.8703, "step": 72 }, { "epoch": 0.11100551226002661, "grad_norm": 0.7618672475143956, "learning_rate": 2.7025076980305847e-05, "loss": 0.882, "step": 73 }, { "epoch": 0.11252613571564342, "grad_norm": 0.5112910612081544, "learning_rate": 2.7021555073227146e-05, "loss": 0.8726, "step": 74 }, { "epoch": 0.11404675917126021, "grad_norm": 0.6949396608818647, "learning_rate": 2.701795251933497e-05, "loss": 0.8671, "step": 75 }, { "epoch": 0.11556738262687702, "grad_norm": 0.741408465947176, "learning_rate": 2.7014269340195837e-05, "loss": 0.8816, "step": 76 }, { "epoch": 0.11708800608249383, "grad_norm": 0.6798649172316688, "learning_rate": 2.7010505557858927e-05, "loss": 0.87, "step": 77 }, { "epoch": 0.11860862953811062, "grad_norm": 0.7927523181373072, "learning_rate": 2.7006661194855928e-05, "loss": 0.8749, "step": 78 }, { "epoch": 0.12012925299372743, "grad_norm": 0.5635726177412718, "learning_rate": 2.7002736274200943e-05, "loss": 0.8472, "step": 79 }, { "epoch": 0.12164987644934423, "grad_norm": 0.6854198399728353, "learning_rate": 2.699873081939032e-05, "loss": 0.8673, "step": 80 }, { "epoch": 0.12317049990496104, "grad_norm": 0.692082771269964, "learning_rate": 2.6994644854402514e-05, "loss": 0.887, "step": 81 }, { "epoch": 0.12469112336057783, "grad_norm": 0.5839628354594621, "learning_rate": 2.6990478403697964e-05, "loss": 0.8566, "step": 82 }, { "epoch": 0.12621174681619465, "grad_norm": 0.5365274637626306, "learning_rate": 2.698623149221892e-05, "loss": 0.8246, "step": 83 }, { "epoch": 0.12773237027181145, "grad_norm": 0.597903169885847, "learning_rate": 2.6981904145389317e-05, "loss": 0.8331, "step": 84 }, { "epoch": 0.12925299372742824, "grad_norm": 0.5825252240476892, "learning_rate": 2.697749638911461e-05, "loss": 0.8496, "step": 85 }, { "epoch": 0.13077361718304506, "grad_norm": 0.7057959507531225, "learning_rate": 2.697300824978161e-05, "loss": 0.9079, "step": 86 }, { "epoch": 0.13229424063866185, "grad_norm": 0.5509650257453511, "learning_rate": 2.6968439754258348e-05, "loss": 0.8447, "step": 87 }, { "epoch": 0.13381486409427865, "grad_norm": 0.6392433098525124, "learning_rate": 2.6963790929893908e-05, "loss": 0.8563, "step": 88 }, { "epoch": 0.13533548754989547, "grad_norm": 0.6150296701472108, "learning_rate": 2.695906180451825e-05, "loss": 0.8746, "step": 89 }, { "epoch": 0.13685611100551226, "grad_norm": 0.5243071440543693, "learning_rate": 2.6954252406442054e-05, "loss": 0.8462, "step": 90 }, { "epoch": 0.13837673446112905, "grad_norm": 0.5930951445222234, "learning_rate": 2.6949362764456548e-05, "loss": 0.8579, "step": 91 }, { "epoch": 0.13989735791674587, "grad_norm": 0.5384200127653291, "learning_rate": 2.694439290783334e-05, "loss": 0.8442, "step": 92 }, { "epoch": 0.14141798137236267, "grad_norm": 0.4526509726759922, "learning_rate": 2.693934286632423e-05, "loss": 0.84, "step": 93 }, { "epoch": 0.14293860482797946, "grad_norm": 0.5941168775185707, "learning_rate": 2.6934212670161057e-05, "loss": 0.8715, "step": 94 }, { "epoch": 0.14445922828359628, "grad_norm": 0.5044642393405714, "learning_rate": 2.6929002350055486e-05, "loss": 0.8815, "step": 95 }, { "epoch": 0.14597985173921307, "grad_norm": 0.5805569476986123, "learning_rate": 2.6923711937198847e-05, "loss": 0.8678, "step": 96 }, { "epoch": 0.14750047519482987, "grad_norm": 0.4790647294772535, "learning_rate": 2.6918341463261945e-05, "loss": 0.8557, "step": 97 }, { "epoch": 0.1490210986504467, "grad_norm": 0.5484601209066985, "learning_rate": 2.691289096039486e-05, "loss": 0.8565, "step": 98 }, { "epoch": 0.15054172210606348, "grad_norm": 0.5095445371601812, "learning_rate": 2.6907360461226763e-05, "loss": 0.8785, "step": 99 }, { "epoch": 0.1520623455616803, "grad_norm": 0.5604825268976987, "learning_rate": 2.6901749998865718e-05, "loss": 0.891, "step": 100 }, { "epoch": 0.1535829690172971, "grad_norm": 0.49509934406592176, "learning_rate": 2.6896059606898493e-05, "loss": 0.8456, "step": 101 }, { "epoch": 0.1551035924729139, "grad_norm": 0.6476921749987207, "learning_rate": 2.6890289319390343e-05, "loss": 0.8217, "step": 102 }, { "epoch": 0.1566242159285307, "grad_norm": 0.4780896251811207, "learning_rate": 2.688443917088481e-05, "loss": 0.8711, "step": 103 }, { "epoch": 0.1581448393841475, "grad_norm": 0.6545666658477076, "learning_rate": 2.687850919640353e-05, "loss": 0.8444, "step": 104 }, { "epoch": 0.1596654628397643, "grad_norm": 0.61000177257977, "learning_rate": 2.687249943144601e-05, "loss": 0.8433, "step": 105 }, { "epoch": 0.16118608629538111, "grad_norm": 0.5126707787512214, "learning_rate": 2.6866409911989412e-05, "loss": 0.839, "step": 106 }, { "epoch": 0.1627067097509979, "grad_norm": 0.4882732798086082, "learning_rate": 2.6860240674488355e-05, "loss": 0.8098, "step": 107 }, { "epoch": 0.1642273332066147, "grad_norm": 0.5554067699143529, "learning_rate": 2.685399175587468e-05, "loss": 0.8406, "step": 108 }, { "epoch": 0.16574795666223152, "grad_norm": 0.530718125213625, "learning_rate": 2.6847663193557236e-05, "loss": 0.8281, "step": 109 }, { "epoch": 0.16726858011784831, "grad_norm": 0.4949823541094598, "learning_rate": 2.684125502542165e-05, "loss": 0.8457, "step": 110 }, { "epoch": 0.1687892035734651, "grad_norm": 0.556274418077457, "learning_rate": 2.683476728983012e-05, "loss": 0.8543, "step": 111 }, { "epoch": 0.17030982702908193, "grad_norm": 0.45222659620728906, "learning_rate": 2.682820002562116e-05, "loss": 0.8348, "step": 112 }, { "epoch": 0.17183045048469872, "grad_norm": 0.4911581618988163, "learning_rate": 2.6821553272109377e-05, "loss": 0.8276, "step": 113 }, { "epoch": 0.17335107394031554, "grad_norm": 0.5259633420118487, "learning_rate": 2.6814827069085237e-05, "loss": 0.8319, "step": 114 }, { "epoch": 0.17487169739593234, "grad_norm": 0.49901119248436054, "learning_rate": 2.6808021456814832e-05, "loss": 0.8608, "step": 115 }, { "epoch": 0.17639232085154913, "grad_norm": 0.44652937902498396, "learning_rate": 2.6801136476039637e-05, "loss": 0.8363, "step": 116 }, { "epoch": 0.17791294430716595, "grad_norm": 0.4814399942165516, "learning_rate": 2.6794172167976247e-05, "loss": 0.8291, "step": 117 }, { "epoch": 0.17943356776278274, "grad_norm": 0.4877382938090122, "learning_rate": 2.6787128574316158e-05, "loss": 0.8547, "step": 118 }, { "epoch": 0.18095419121839953, "grad_norm": 0.6191110354751792, "learning_rate": 2.6780005737225512e-05, "loss": 0.81, "step": 119 }, { "epoch": 0.18247481467401636, "grad_norm": 0.5437167197085974, "learning_rate": 2.677280369934482e-05, "loss": 0.8393, "step": 120 }, { "epoch": 0.18399543812963315, "grad_norm": 0.5297828079815319, "learning_rate": 2.676552250378873e-05, "loss": 0.7954, "step": 121 }, { "epoch": 0.18551606158524994, "grad_norm": 0.5514706250971693, "learning_rate": 2.6758162194145783e-05, "loss": 0.815, "step": 122 }, { "epoch": 0.18703668504086676, "grad_norm": 0.5843944412164658, "learning_rate": 2.6750722814478098e-05, "loss": 0.8276, "step": 123 }, { "epoch": 0.18855730849648356, "grad_norm": 0.7942439470942368, "learning_rate": 2.6743204409321177e-05, "loss": 0.8254, "step": 124 }, { "epoch": 0.19007793195210035, "grad_norm": 0.6174070370639455, "learning_rate": 2.673560702368358e-05, "loss": 0.8549, "step": 125 }, { "epoch": 0.19159855540771717, "grad_norm": 0.49426597944430567, "learning_rate": 2.6727930703046695e-05, "loss": 0.8447, "step": 126 }, { "epoch": 0.19311917886333396, "grad_norm": 0.4989733325059912, "learning_rate": 2.6720175493364437e-05, "loss": 0.8499, "step": 127 }, { "epoch": 0.19463980231895078, "grad_norm": 0.5741663400348602, "learning_rate": 2.6712341441063006e-05, "loss": 0.8675, "step": 128 }, { "epoch": 0.19616042577456758, "grad_norm": 0.532503832446581, "learning_rate": 2.6704428593040568e-05, "loss": 0.8053, "step": 129 }, { "epoch": 0.19768104923018437, "grad_norm": 0.47559928839451404, "learning_rate": 2.6696436996667005e-05, "loss": 0.8075, "step": 130 }, { "epoch": 0.1992016726858012, "grad_norm": 0.5372739975864717, "learning_rate": 2.6688366699783625e-05, "loss": 0.8565, "step": 131 }, { "epoch": 0.20072229614141798, "grad_norm": 0.5575627816877122, "learning_rate": 2.6680217750702874e-05, "loss": 0.8082, "step": 132 }, { "epoch": 0.20224291959703478, "grad_norm": 0.5868207344499796, "learning_rate": 2.6671990198208038e-05, "loss": 0.781, "step": 133 }, { "epoch": 0.2037635430526516, "grad_norm": 0.4362747143153947, "learning_rate": 2.6663684091552962e-05, "loss": 0.817, "step": 134 }, { "epoch": 0.2052841665082684, "grad_norm": 0.5793319639019973, "learning_rate": 2.6655299480461753e-05, "loss": 0.8414, "step": 135 }, { "epoch": 0.20680478996388518, "grad_norm": 0.5002998693635706, "learning_rate": 2.6646836415128478e-05, "loss": 0.8399, "step": 136 }, { "epoch": 0.208325413419502, "grad_norm": 0.5183039332160528, "learning_rate": 2.6638294946216876e-05, "loss": 0.8388, "step": 137 }, { "epoch": 0.2098460368751188, "grad_norm": 0.5333906191333849, "learning_rate": 2.6629675124860034e-05, "loss": 0.8579, "step": 138 }, { "epoch": 0.2113666603307356, "grad_norm": 0.5678231978891924, "learning_rate": 2.662097700266009e-05, "loss": 0.7973, "step": 139 }, { "epoch": 0.2128872837863524, "grad_norm": 0.6779338312838604, "learning_rate": 2.6612200631687935e-05, "loss": 0.7991, "step": 140 }, { "epoch": 0.2144079072419692, "grad_norm": 0.4549445363589815, "learning_rate": 2.6603346064482896e-05, "loss": 0.8342, "step": 141 }, { "epoch": 0.21592853069758602, "grad_norm": 0.5534663745815431, "learning_rate": 2.6594413354052406e-05, "loss": 0.8486, "step": 142 }, { "epoch": 0.21744915415320282, "grad_norm": 0.5263202593856873, "learning_rate": 2.6585402553871707e-05, "loss": 0.8162, "step": 143 }, { "epoch": 0.2189697776088196, "grad_norm": 0.5005224606123282, "learning_rate": 2.6576313717883517e-05, "loss": 0.8157, "step": 144 }, { "epoch": 0.22049040106443643, "grad_norm": 0.4408661056131368, "learning_rate": 2.6567146900497715e-05, "loss": 0.8073, "step": 145 }, { "epoch": 0.22201102452005322, "grad_norm": 0.5697511997765187, "learning_rate": 2.655790215659101e-05, "loss": 0.814, "step": 146 }, { "epoch": 0.22353164797567002, "grad_norm": 0.5227769257193289, "learning_rate": 2.654857954150661e-05, "loss": 0.8437, "step": 147 }, { "epoch": 0.22505227143128684, "grad_norm": 0.5335084066857467, "learning_rate": 2.6539179111053904e-05, "loss": 0.8122, "step": 148 }, { "epoch": 0.22657289488690363, "grad_norm": 0.6845924526755846, "learning_rate": 2.6529700921508117e-05, "loss": 0.8109, "step": 149 }, { "epoch": 0.22809351834252042, "grad_norm": 0.5497202515145223, "learning_rate": 2.652014502960997e-05, "loss": 0.8482, "step": 150 }, { "epoch": 0.22961414179813724, "grad_norm": 0.5925506572004898, "learning_rate": 2.651051149256535e-05, "loss": 0.8182, "step": 151 }, { "epoch": 0.23113476525375404, "grad_norm": 0.4748040759396778, "learning_rate": 2.6500800368044956e-05, "loss": 0.8236, "step": 152 }, { "epoch": 0.23265538870937083, "grad_norm": 0.4949711715480288, "learning_rate": 2.6491011714183972e-05, "loss": 0.8044, "step": 153 }, { "epoch": 0.23417601216498765, "grad_norm": 0.4942784715771247, "learning_rate": 2.6481145589581697e-05, "loss": 0.8288, "step": 154 }, { "epoch": 0.23569663562060444, "grad_norm": 0.5323652293681927, "learning_rate": 2.647120205330121e-05, "loss": 0.7997, "step": 155 }, { "epoch": 0.23721725907622124, "grad_norm": 0.6476575452897515, "learning_rate": 2.646118116486901e-05, "loss": 0.8077, "step": 156 }, { "epoch": 0.23873788253183806, "grad_norm": 0.47184034820772325, "learning_rate": 2.6451082984274666e-05, "loss": 0.8292, "step": 157 }, { "epoch": 0.24025850598745485, "grad_norm": 0.6205069464172376, "learning_rate": 2.6440907571970438e-05, "loss": 0.8393, "step": 158 }, { "epoch": 0.24177912944307167, "grad_norm": 0.5209484434921768, "learning_rate": 2.6430654988870954e-05, "loss": 0.8245, "step": 159 }, { "epoch": 0.24329975289868846, "grad_norm": 0.5933368833938595, "learning_rate": 2.6420325296352796e-05, "loss": 0.8197, "step": 160 }, { "epoch": 0.24482037635430526, "grad_norm": 0.4788998696836293, "learning_rate": 2.6409918556254172e-05, "loss": 0.7937, "step": 161 }, { "epoch": 0.24634099980992208, "grad_norm": 0.567878399230218, "learning_rate": 2.639943483087453e-05, "loss": 0.8109, "step": 162 }, { "epoch": 0.24786162326553887, "grad_norm": 0.507261963638979, "learning_rate": 2.6388874182974187e-05, "loss": 0.8218, "step": 163 }, { "epoch": 0.24938224672115566, "grad_norm": 0.5864902720090787, "learning_rate": 2.637823667577395e-05, "loss": 0.8176, "step": 164 }, { "epoch": 0.2509028701767725, "grad_norm": 0.5918504624886162, "learning_rate": 2.636752237295474e-05, "loss": 0.83, "step": 165 }, { "epoch": 0.2524234936323893, "grad_norm": 0.43858309995443784, "learning_rate": 2.6356731338657212e-05, "loss": 0.8123, "step": 166 }, { "epoch": 0.25394411708800607, "grad_norm": 0.49034964682512816, "learning_rate": 2.6345863637481374e-05, "loss": 0.8174, "step": 167 }, { "epoch": 0.2554647405436229, "grad_norm": 0.5837987779382616, "learning_rate": 2.6334919334486195e-05, "loss": 0.8298, "step": 168 }, { "epoch": 0.2569853639992397, "grad_norm": 0.5281077280878048, "learning_rate": 2.6323898495189215e-05, "loss": 0.8175, "step": 169 }, { "epoch": 0.2585059874548565, "grad_norm": 0.5014520173664685, "learning_rate": 2.6312801185566158e-05, "loss": 0.799, "step": 170 }, { "epoch": 0.2600266109104733, "grad_norm": 0.6803320146493841, "learning_rate": 2.6301627472050526e-05, "loss": 0.7996, "step": 171 }, { "epoch": 0.2615472343660901, "grad_norm": 0.7395949415978587, "learning_rate": 2.629037742153322e-05, "loss": 0.8107, "step": 172 }, { "epoch": 0.2630678578217069, "grad_norm": 0.4513389629082933, "learning_rate": 2.6279051101362122e-05, "loss": 0.7952, "step": 173 }, { "epoch": 0.2645884812773237, "grad_norm": 0.6486009072322451, "learning_rate": 2.62676485793417e-05, "loss": 0.7874, "step": 174 }, { "epoch": 0.2661091047329405, "grad_norm": 0.5502319884395829, "learning_rate": 2.62561699237326e-05, "loss": 0.7963, "step": 175 }, { "epoch": 0.2676297281885573, "grad_norm": 0.5997728345326634, "learning_rate": 2.624461520325124e-05, "loss": 0.7972, "step": 176 }, { "epoch": 0.2691503516441741, "grad_norm": 0.7905728427904212, "learning_rate": 2.62329844870694e-05, "loss": 0.8029, "step": 177 }, { "epoch": 0.27067097509979093, "grad_norm": 0.6068096518092964, "learning_rate": 2.62212778448138e-05, "loss": 0.8434, "step": 178 }, { "epoch": 0.2721915985554077, "grad_norm": 0.8084200216220121, "learning_rate": 2.620949534656568e-05, "loss": 0.8005, "step": 179 }, { "epoch": 0.2737122220110245, "grad_norm": 0.4950225377554484, "learning_rate": 2.61976370628604e-05, "loss": 0.8261, "step": 180 }, { "epoch": 0.27523284546664134, "grad_norm": 0.5184301676475981, "learning_rate": 2.618570306468701e-05, "loss": 0.8239, "step": 181 }, { "epoch": 0.2767534689222581, "grad_norm": 0.601710109353413, "learning_rate": 2.61736934234878e-05, "loss": 0.7888, "step": 182 }, { "epoch": 0.2782740923778749, "grad_norm": 0.5736633456413043, "learning_rate": 2.616160821115792e-05, "loss": 0.8098, "step": 183 }, { "epoch": 0.27979471583349175, "grad_norm": 0.5690928538172545, "learning_rate": 2.6149447500044904e-05, "loss": 0.8294, "step": 184 }, { "epoch": 0.2813153392891085, "grad_norm": 0.6275365250577597, "learning_rate": 2.6137211362948256e-05, "loss": 0.8386, "step": 185 }, { "epoch": 0.28283596274472533, "grad_norm": 0.6701221466806971, "learning_rate": 2.612489987311903e-05, "loss": 0.791, "step": 186 }, { "epoch": 0.28435658620034215, "grad_norm": 0.7648465329508422, "learning_rate": 2.6112513104259354e-05, "loss": 0.8125, "step": 187 }, { "epoch": 0.2858772096559589, "grad_norm": 0.44806704127100555, "learning_rate": 2.6100051130522028e-05, "loss": 0.7849, "step": 188 }, { "epoch": 0.28739783311157574, "grad_norm": 0.7123885835323628, "learning_rate": 2.608751402651006e-05, "loss": 0.7607, "step": 189 }, { "epoch": 0.28891845656719256, "grad_norm": 0.4934771024884215, "learning_rate": 2.6074901867276204e-05, "loss": 0.8035, "step": 190 }, { "epoch": 0.2904390800228093, "grad_norm": 0.8095025862539972, "learning_rate": 2.6062214728322555e-05, "loss": 0.7829, "step": 191 }, { "epoch": 0.29195970347842615, "grad_norm": 0.5050611453741563, "learning_rate": 2.6049452685600052e-05, "loss": 0.7977, "step": 192 }, { "epoch": 0.29348032693404297, "grad_norm": 0.7210248910137731, "learning_rate": 2.6036615815508043e-05, "loss": 0.7956, "step": 193 }, { "epoch": 0.29500095038965973, "grad_norm": 0.5038427749270625, "learning_rate": 2.6023704194893835e-05, "loss": 0.7938, "step": 194 }, { "epoch": 0.29652157384527655, "grad_norm": 0.6126219303001782, "learning_rate": 2.6010717901052224e-05, "loss": 0.7922, "step": 195 }, { "epoch": 0.2980421973008934, "grad_norm": 0.5362006682961364, "learning_rate": 2.5997657011725022e-05, "loss": 0.8688, "step": 196 }, { "epoch": 0.2995628207565102, "grad_norm": 0.5017573346892746, "learning_rate": 2.598452160510062e-05, "loss": 0.7833, "step": 197 }, { "epoch": 0.30108344421212696, "grad_norm": 0.5086171325358533, "learning_rate": 2.597131175981349e-05, "loss": 0.8173, "step": 198 }, { "epoch": 0.3026040676677438, "grad_norm": 0.46242832959563607, "learning_rate": 2.5958027554943737e-05, "loss": 0.8228, "step": 199 }, { "epoch": 0.3041246911233606, "grad_norm": 0.5365620033072362, "learning_rate": 2.594466907001661e-05, "loss": 0.8468, "step": 200 }, { "epoch": 0.30564531457897737, "grad_norm": 0.4536900282460799, "learning_rate": 2.5931236385002023e-05, "loss": 0.7659, "step": 201 }, { "epoch": 0.3071659380345942, "grad_norm": 0.5349495426115096, "learning_rate": 2.591772958031411e-05, "loss": 0.8031, "step": 202 }, { "epoch": 0.308686561490211, "grad_norm": 0.5171572447957907, "learning_rate": 2.5904148736810704e-05, "loss": 0.8192, "step": 203 }, { "epoch": 0.3102071849458278, "grad_norm": 0.543619906524382, "learning_rate": 2.5890493935792855e-05, "loss": 0.8088, "step": 204 }, { "epoch": 0.3117278084014446, "grad_norm": 0.5471779246235068, "learning_rate": 2.5876765259004386e-05, "loss": 0.775, "step": 205 }, { "epoch": 0.3132484318570614, "grad_norm": 0.5530162050934202, "learning_rate": 2.5862962788631346e-05, "loss": 0.7886, "step": 206 }, { "epoch": 0.3147690553126782, "grad_norm": 0.5564374353682834, "learning_rate": 2.5849086607301564e-05, "loss": 0.7822, "step": 207 }, { "epoch": 0.316289678768295, "grad_norm": 0.50158890267917, "learning_rate": 2.583513679808412e-05, "loss": 0.7857, "step": 208 }, { "epoch": 0.3178103022239118, "grad_norm": 0.5514584077868472, "learning_rate": 2.582111344448888e-05, "loss": 0.7834, "step": 209 }, { "epoch": 0.3193309256795286, "grad_norm": 0.4984807668723581, "learning_rate": 2.580701663046597e-05, "loss": 0.7979, "step": 210 }, { "epoch": 0.3208515491351454, "grad_norm": 0.5779043543552743, "learning_rate": 2.579284644040529e-05, "loss": 0.8365, "step": 211 }, { "epoch": 0.32237217259076223, "grad_norm": 0.4914952548115715, "learning_rate": 2.5778602959135996e-05, "loss": 0.8082, "step": 212 }, { "epoch": 0.323892796046379, "grad_norm": 0.45463188187952797, "learning_rate": 2.5764286271925997e-05, "loss": 0.8002, "step": 213 }, { "epoch": 0.3254134195019958, "grad_norm": 0.5405224256712096, "learning_rate": 2.5749896464481448e-05, "loss": 0.8177, "step": 214 }, { "epoch": 0.32693404295761264, "grad_norm": 0.4945710149798116, "learning_rate": 2.5735433622946242e-05, "loss": 0.8202, "step": 215 }, { "epoch": 0.3284546664132294, "grad_norm": 0.5193387642723397, "learning_rate": 2.572089783390148e-05, "loss": 0.7878, "step": 216 }, { "epoch": 0.3299752898688462, "grad_norm": 0.43644483818152885, "learning_rate": 2.5706289184364962e-05, "loss": 0.8465, "step": 217 }, { "epoch": 0.33149591332446304, "grad_norm": 0.6429429755905944, "learning_rate": 2.569160776179065e-05, "loss": 0.793, "step": 218 }, { "epoch": 0.3330165367800798, "grad_norm": 0.4881784207711263, "learning_rate": 2.567685365406819e-05, "loss": 0.7787, "step": 219 }, { "epoch": 0.33453716023569663, "grad_norm": 0.5903797222010695, "learning_rate": 2.566202694952232e-05, "loss": 0.8071, "step": 220 }, { "epoch": 0.33605778369131345, "grad_norm": 0.4935742650802179, "learning_rate": 2.5647127736912397e-05, "loss": 0.8128, "step": 221 }, { "epoch": 0.3375784071469302, "grad_norm": 0.6206075855689506, "learning_rate": 2.5632156105431836e-05, "loss": 0.8134, "step": 222 }, { "epoch": 0.33909903060254704, "grad_norm": 0.4693335100552506, "learning_rate": 2.561711214470759e-05, "loss": 0.7877, "step": 223 }, { "epoch": 0.34061965405816386, "grad_norm": 0.5541352157938085, "learning_rate": 2.560199594479959e-05, "loss": 0.793, "step": 224 }, { "epoch": 0.3421402775137807, "grad_norm": 0.5119311192678664, "learning_rate": 2.5586807596200247e-05, "loss": 0.8076, "step": 225 }, { "epoch": 0.34366090096939744, "grad_norm": 0.44821567228976084, "learning_rate": 2.5571547189833876e-05, "loss": 0.7676, "step": 226 }, { "epoch": 0.34518152442501426, "grad_norm": 0.5108021271320223, "learning_rate": 2.555621481705616e-05, "loss": 0.7761, "step": 227 }, { "epoch": 0.3467021478806311, "grad_norm": 0.520419930384087, "learning_rate": 2.5540810569653606e-05, "loss": 0.8113, "step": 228 }, { "epoch": 0.34822277133624785, "grad_norm": 0.49172520245402473, "learning_rate": 2.5525334539843e-05, "loss": 0.8049, "step": 229 }, { "epoch": 0.34974339479186467, "grad_norm": 0.5350433416575631, "learning_rate": 2.550978682027084e-05, "loss": 0.7918, "step": 230 }, { "epoch": 0.3512640182474815, "grad_norm": 0.5521639416583181, "learning_rate": 2.5494167504012807e-05, "loss": 0.7705, "step": 231 }, { "epoch": 0.35278464170309826, "grad_norm": 0.4924879256498397, "learning_rate": 2.547847668457318e-05, "loss": 0.8116, "step": 232 }, { "epoch": 0.3543052651587151, "grad_norm": 0.5466037905256877, "learning_rate": 2.5462714455884287e-05, "loss": 0.7808, "step": 233 }, { "epoch": 0.3558258886143319, "grad_norm": 0.5030091896084465, "learning_rate": 2.5446880912305954e-05, "loss": 0.7968, "step": 234 }, { "epoch": 0.35734651206994866, "grad_norm": 0.47695683680901585, "learning_rate": 2.543097614862492e-05, "loss": 0.7777, "step": 235 }, { "epoch": 0.3588671355255655, "grad_norm": 0.49323354270252745, "learning_rate": 2.5415000260054287e-05, "loss": 0.8114, "step": 236 }, { "epoch": 0.3603877589811823, "grad_norm": 0.5547918130480031, "learning_rate": 2.5398953342232942e-05, "loss": 0.7959, "step": 237 }, { "epoch": 0.36190838243679907, "grad_norm": 0.42592099953025825, "learning_rate": 2.5382835491224977e-05, "loss": 0.8067, "step": 238 }, { "epoch": 0.3634290058924159, "grad_norm": 0.5473160997053151, "learning_rate": 2.536664680351913e-05, "loss": 0.8149, "step": 239 }, { "epoch": 0.3649496293480327, "grad_norm": 0.5161750461424985, "learning_rate": 2.5350387376028207e-05, "loss": 0.7456, "step": 240 }, { "epoch": 0.3664702528036495, "grad_norm": 0.562540015047346, "learning_rate": 2.5334057306088472e-05, "loss": 0.8161, "step": 241 }, { "epoch": 0.3679908762592663, "grad_norm": 0.5331918164095757, "learning_rate": 2.5317656691459103e-05, "loss": 0.7834, "step": 242 }, { "epoch": 0.3695114997148831, "grad_norm": 0.5704962131477223, "learning_rate": 2.530118563032159e-05, "loss": 0.7881, "step": 243 }, { "epoch": 0.3710321231704999, "grad_norm": 0.5208634783184293, "learning_rate": 2.5284644221279147e-05, "loss": 0.8138, "step": 244 }, { "epoch": 0.3725527466261167, "grad_norm": 0.5211185784219639, "learning_rate": 2.526803256335611e-05, "loss": 0.831, "step": 245 }, { "epoch": 0.3740733700817335, "grad_norm": 0.5045487224701021, "learning_rate": 2.525135075599738e-05, "loss": 0.8156, "step": 246 }, { "epoch": 0.3755939935373503, "grad_norm": 0.5069878820838861, "learning_rate": 2.523459889906778e-05, "loss": 0.7955, "step": 247 }, { "epoch": 0.3771146169929671, "grad_norm": 0.4949818081031886, "learning_rate": 2.5217777092851496e-05, "loss": 0.7941, "step": 248 }, { "epoch": 0.37863524044858393, "grad_norm": 0.47385420044302173, "learning_rate": 2.5200885438051454e-05, "loss": 0.7897, "step": 249 }, { "epoch": 0.3801558639042007, "grad_norm": 0.4584812154097583, "learning_rate": 2.5183924035788738e-05, "loss": 0.8149, "step": 250 }, { "epoch": 0.3816764873598175, "grad_norm": 0.40766128893251813, "learning_rate": 2.5166892987601954e-05, "loss": 0.7965, "step": 251 }, { "epoch": 0.38319711081543434, "grad_norm": 0.4464504741734316, "learning_rate": 2.5149792395446655e-05, "loss": 0.7793, "step": 252 }, { "epoch": 0.3847177342710511, "grad_norm": 0.42738515223990925, "learning_rate": 2.5132622361694704e-05, "loss": 0.7957, "step": 253 }, { "epoch": 0.3862383577266679, "grad_norm": 0.4980456405593658, "learning_rate": 2.5115382989133686e-05, "loss": 0.8032, "step": 254 }, { "epoch": 0.38775898118228475, "grad_norm": 0.5114356446133503, "learning_rate": 2.5098074380966263e-05, "loss": 0.781, "step": 255 }, { "epoch": 0.38927960463790157, "grad_norm": 0.4141552486101403, "learning_rate": 2.508069664080959e-05, "loss": 0.7593, "step": 256 }, { "epoch": 0.39080022809351833, "grad_norm": 0.43458143129055027, "learning_rate": 2.5063249872694662e-05, "loss": 0.7879, "step": 257 }, { "epoch": 0.39232085154913515, "grad_norm": 0.45941178211274825, "learning_rate": 2.504573418106572e-05, "loss": 0.7947, "step": 258 }, { "epoch": 0.393841475004752, "grad_norm": 0.49731812944358156, "learning_rate": 2.5028149670779597e-05, "loss": 0.7796, "step": 259 }, { "epoch": 0.39536209846036874, "grad_norm": 0.4021875079520183, "learning_rate": 2.5010496447105118e-05, "loss": 0.7844, "step": 260 }, { "epoch": 0.39688272191598556, "grad_norm": 0.5734524869531662, "learning_rate": 2.4992774615722457e-05, "loss": 0.7739, "step": 261 }, { "epoch": 0.3984033453716024, "grad_norm": 0.4767545528986454, "learning_rate": 2.4974984282722498e-05, "loss": 0.7787, "step": 262 }, { "epoch": 0.39992396882721915, "grad_norm": 0.45569173851803857, "learning_rate": 2.4957125554606207e-05, "loss": 0.7957, "step": 263 }, { "epoch": 0.40144459228283597, "grad_norm": 0.4583616284261655, "learning_rate": 2.4939198538284008e-05, "loss": 0.7903, "step": 264 }, { "epoch": 0.4029652157384528, "grad_norm": 0.44137521284735437, "learning_rate": 2.4921203341075102e-05, "loss": 0.8021, "step": 265 }, { "epoch": 0.40448583919406955, "grad_norm": 0.4648974770143655, "learning_rate": 2.4903140070706876e-05, "loss": 0.7739, "step": 266 }, { "epoch": 0.4060064626496864, "grad_norm": 0.521001369014178, "learning_rate": 2.4885008835314228e-05, "loss": 0.7837, "step": 267 }, { "epoch": 0.4075270861053032, "grad_norm": 0.4601570641740828, "learning_rate": 2.4866809743438915e-05, "loss": 0.7891, "step": 268 }, { "epoch": 0.40904770956091996, "grad_norm": 0.42883010607831595, "learning_rate": 2.484854290402893e-05, "loss": 0.8002, "step": 269 }, { "epoch": 0.4105683330165368, "grad_norm": 0.5048746687096498, "learning_rate": 2.4830208426437816e-05, "loss": 0.8067, "step": 270 }, { "epoch": 0.4120889564721536, "grad_norm": 0.4373843602878167, "learning_rate": 2.4811806420424043e-05, "loss": 0.7661, "step": 271 }, { "epoch": 0.41360957992777037, "grad_norm": 0.4924768994796526, "learning_rate": 2.479333699615033e-05, "loss": 0.8032, "step": 272 }, { "epoch": 0.4151302033833872, "grad_norm": 0.5495367191198085, "learning_rate": 2.477480026418299e-05, "loss": 0.7823, "step": 273 }, { "epoch": 0.416650826839004, "grad_norm": 0.4439519855217945, "learning_rate": 2.4756196335491273e-05, "loss": 0.7655, "step": 274 }, { "epoch": 0.4181714502946208, "grad_norm": 0.5010590403262122, "learning_rate": 2.47375253214467e-05, "loss": 0.7547, "step": 275 }, { "epoch": 0.4196920737502376, "grad_norm": 0.4482088714543724, "learning_rate": 2.4718787333822395e-05, "loss": 0.7649, "step": 276 }, { "epoch": 0.4212126972058544, "grad_norm": 0.40410880639654234, "learning_rate": 2.469998248479242e-05, "loss": 0.7669, "step": 277 }, { "epoch": 0.4227333206614712, "grad_norm": 0.4569274893231673, "learning_rate": 2.468111088693109e-05, "loss": 0.7526, "step": 278 }, { "epoch": 0.424253944117088, "grad_norm": 0.42296277869189824, "learning_rate": 2.4662172653212313e-05, "loss": 0.7727, "step": 279 }, { "epoch": 0.4257745675727048, "grad_norm": 0.4835330143189326, "learning_rate": 2.464316789700891e-05, "loss": 0.7633, "step": 280 }, { "epoch": 0.4272951910283216, "grad_norm": 0.4287243627654868, "learning_rate": 2.462409673209194e-05, "loss": 0.7836, "step": 281 }, { "epoch": 0.4288158144839384, "grad_norm": 0.5472776086861091, "learning_rate": 2.4604959272630002e-05, "loss": 0.778, "step": 282 }, { "epoch": 0.4303364379395552, "grad_norm": 0.43700006561477783, "learning_rate": 2.4585755633188585e-05, "loss": 0.7657, "step": 283 }, { "epoch": 0.43185706139517205, "grad_norm": 0.5233246056259679, "learning_rate": 2.4566485928729338e-05, "loss": 0.7875, "step": 284 }, { "epoch": 0.4333776848507888, "grad_norm": 0.4739436376744406, "learning_rate": 2.454715027460942e-05, "loss": 0.7769, "step": 285 }, { "epoch": 0.43489830830640563, "grad_norm": 0.466587612255068, "learning_rate": 2.4527748786580785e-05, "loss": 0.7763, "step": 286 }, { "epoch": 0.43641893176202246, "grad_norm": 0.5917227864064121, "learning_rate": 2.4508281580789508e-05, "loss": 0.7643, "step": 287 }, { "epoch": 0.4379395552176392, "grad_norm": 0.47768374233206695, "learning_rate": 2.448874877377508e-05, "loss": 0.7776, "step": 288 }, { "epoch": 0.43946017867325604, "grad_norm": 0.527002493093892, "learning_rate": 2.4469150482469695e-05, "loss": 0.7796, "step": 289 }, { "epoch": 0.44098080212887286, "grad_norm": 0.4662349412142309, "learning_rate": 2.4449486824197585e-05, "loss": 0.7385, "step": 290 }, { "epoch": 0.4425014255844896, "grad_norm": 0.4732365292282727, "learning_rate": 2.4429757916674292e-05, "loss": 0.7967, "step": 291 }, { "epoch": 0.44402204904010645, "grad_norm": 0.47222270479478673, "learning_rate": 2.440996387800596e-05, "loss": 0.7682, "step": 292 }, { "epoch": 0.44554267249572327, "grad_norm": 0.47686692962034305, "learning_rate": 2.439010482668865e-05, "loss": 0.7611, "step": 293 }, { "epoch": 0.44706329595134003, "grad_norm": 0.4008859391749161, "learning_rate": 2.4370180881607614e-05, "loss": 0.7566, "step": 294 }, { "epoch": 0.44858391940695685, "grad_norm": 0.48140611283685464, "learning_rate": 2.4350192162036584e-05, "loss": 0.8217, "step": 295 }, { "epoch": 0.4501045428625737, "grad_norm": 0.4560651975700714, "learning_rate": 2.4330138787637062e-05, "loss": 0.7976, "step": 296 }, { "epoch": 0.45162516631819044, "grad_norm": 0.4034722549460792, "learning_rate": 2.431002087845761e-05, "loss": 0.7679, "step": 297 }, { "epoch": 0.45314578977380726, "grad_norm": 0.4807251922703316, "learning_rate": 2.4289838554933113e-05, "loss": 0.7984, "step": 298 }, { "epoch": 0.4546664132294241, "grad_norm": 0.4746736626277042, "learning_rate": 2.4269591937884083e-05, "loss": 0.7794, "step": 299 }, { "epoch": 0.45618703668504085, "grad_norm": 0.47543886824696496, "learning_rate": 2.4249281148515904e-05, "loss": 0.7865, "step": 300 }, { "epoch": 0.45770766014065767, "grad_norm": 0.4982520457747873, "learning_rate": 2.4228906308418148e-05, "loss": 0.7653, "step": 301 }, { "epoch": 0.4592282835962745, "grad_norm": 0.4541992280820245, "learning_rate": 2.42084675395638e-05, "loss": 0.8036, "step": 302 }, { "epoch": 0.46074890705189125, "grad_norm": 0.4933234210123189, "learning_rate": 2.4187964964308566e-05, "loss": 0.7764, "step": 303 }, { "epoch": 0.4622695305075081, "grad_norm": 0.46059558913855275, "learning_rate": 2.4167398705390122e-05, "loss": 0.7736, "step": 304 }, { "epoch": 0.4637901539631249, "grad_norm": 0.5168855999894718, "learning_rate": 2.414676888592739e-05, "loss": 0.8157, "step": 305 }, { "epoch": 0.46531077741874166, "grad_norm": 0.46112787242313547, "learning_rate": 2.4126075629419772e-05, "loss": 0.767, "step": 306 }, { "epoch": 0.4668314008743585, "grad_norm": 0.4302292629253772, "learning_rate": 2.4105319059746465e-05, "loss": 0.7775, "step": 307 }, { "epoch": 0.4683520243299753, "grad_norm": 0.4152046813472176, "learning_rate": 2.4084499301165655e-05, "loss": 0.8137, "step": 308 }, { "epoch": 0.46987264778559207, "grad_norm": 0.4876150328408346, "learning_rate": 2.4063616478313835e-05, "loss": 0.7809, "step": 309 }, { "epoch": 0.4713932712412089, "grad_norm": 0.47894882147037643, "learning_rate": 2.4042670716205003e-05, "loss": 0.8071, "step": 310 }, { "epoch": 0.4729138946968257, "grad_norm": 0.45314893137296913, "learning_rate": 2.4021662140229957e-05, "loss": 0.7734, "step": 311 }, { "epoch": 0.4744345181524425, "grad_norm": 0.4693692990401656, "learning_rate": 2.4000590876155523e-05, "loss": 0.8049, "step": 312 }, { "epoch": 0.4759551416080593, "grad_norm": 0.48467692671175333, "learning_rate": 2.3979457050123804e-05, "loss": 0.7605, "step": 313 }, { "epoch": 0.4774757650636761, "grad_norm": 0.4222969005662164, "learning_rate": 2.3958260788651426e-05, "loss": 0.7788, "step": 314 }, { "epoch": 0.47899638851929294, "grad_norm": 0.5542291365632019, "learning_rate": 2.3937002218628792e-05, "loss": 0.7928, "step": 315 }, { "epoch": 0.4805170119749097, "grad_norm": 0.392436349975769, "learning_rate": 2.3915681467319292e-05, "loss": 0.7607, "step": 316 }, { "epoch": 0.4820376354305265, "grad_norm": 0.49739152408387094, "learning_rate": 2.3894298662358583e-05, "loss": 0.7614, "step": 317 }, { "epoch": 0.48355825888614334, "grad_norm": 0.433372030563949, "learning_rate": 2.3872853931753787e-05, "loss": 0.7652, "step": 318 }, { "epoch": 0.4850788823417601, "grad_norm": 0.46102099141447744, "learning_rate": 2.385134740388275e-05, "loss": 0.774, "step": 319 }, { "epoch": 0.48659950579737693, "grad_norm": 0.4638491872936722, "learning_rate": 2.382977920749327e-05, "loss": 0.7739, "step": 320 }, { "epoch": 0.48812012925299375, "grad_norm": 0.5020784828944407, "learning_rate": 2.3808149471702295e-05, "loss": 0.7631, "step": 321 }, { "epoch": 0.4896407527086105, "grad_norm": 0.42204949679328635, "learning_rate": 2.3786458325995214e-05, "loss": 0.7823, "step": 322 }, { "epoch": 0.49116137616422734, "grad_norm": 0.5055396085814678, "learning_rate": 2.3764705900225004e-05, "loss": 0.7578, "step": 323 }, { "epoch": 0.49268199961984416, "grad_norm": 0.47833233111711587, "learning_rate": 2.374289232461152e-05, "loss": 0.7545, "step": 324 }, { "epoch": 0.4942026230754609, "grad_norm": 0.3848734118646103, "learning_rate": 2.372101772974068e-05, "loss": 0.7141, "step": 325 }, { "epoch": 0.49572324653107774, "grad_norm": 0.5367848465213906, "learning_rate": 2.3699082246563687e-05, "loss": 0.7789, "step": 326 }, { "epoch": 0.49724386998669456, "grad_norm": 0.5461740930935474, "learning_rate": 2.367708600639625e-05, "loss": 0.7457, "step": 327 }, { "epoch": 0.49876449344231133, "grad_norm": 0.4247642096240754, "learning_rate": 2.3655029140917794e-05, "loss": 0.76, "step": 328 }, { "epoch": 0.5002851168979282, "grad_norm": 0.44998842461109745, "learning_rate": 2.3632911782170683e-05, "loss": 0.7793, "step": 329 }, { "epoch": 0.501805740353545, "grad_norm": 0.47053242929737366, "learning_rate": 2.3610734062559418e-05, "loss": 0.7858, "step": 330 }, { "epoch": 0.5033263638091617, "grad_norm": 0.43924849365849383, "learning_rate": 2.3588496114849843e-05, "loss": 0.8032, "step": 331 }, { "epoch": 0.5048469872647786, "grad_norm": 0.4355576929843276, "learning_rate": 2.3566198072168355e-05, "loss": 0.7691, "step": 332 }, { "epoch": 0.5063676107203954, "grad_norm": 0.4498593288508342, "learning_rate": 2.354384006800112e-05, "loss": 0.7538, "step": 333 }, { "epoch": 0.5078882341760121, "grad_norm": 0.42469702355418737, "learning_rate": 2.352142223619323e-05, "loss": 0.7594, "step": 334 }, { "epoch": 0.509408857631629, "grad_norm": 0.5132890157597293, "learning_rate": 2.3498944710947973e-05, "loss": 0.7693, "step": 335 }, { "epoch": 0.5109294810872458, "grad_norm": 0.43927719097563, "learning_rate": 2.3476407626825953e-05, "loss": 0.7783, "step": 336 }, { "epoch": 0.5124501045428626, "grad_norm": 0.4416642559264683, "learning_rate": 2.3453811118744345e-05, "loss": 0.7683, "step": 337 }, { "epoch": 0.5139707279984794, "grad_norm": 0.44247651582101677, "learning_rate": 2.3431155321976045e-05, "loss": 0.746, "step": 338 }, { "epoch": 0.5154913514540962, "grad_norm": 0.4243771777807823, "learning_rate": 2.340844037214889e-05, "loss": 0.7735, "step": 339 }, { "epoch": 0.517011974909713, "grad_norm": 0.45520920364804773, "learning_rate": 2.3385666405244823e-05, "loss": 0.7605, "step": 340 }, { "epoch": 0.5185325983653298, "grad_norm": 0.41908969886233616, "learning_rate": 2.3362833557599105e-05, "loss": 0.7767, "step": 341 }, { "epoch": 0.5200532218209466, "grad_norm": 0.41422212342426945, "learning_rate": 2.333994196589947e-05, "loss": 0.7858, "step": 342 }, { "epoch": 0.5215738452765634, "grad_norm": 0.45155134128630714, "learning_rate": 2.331699176718533e-05, "loss": 0.7436, "step": 343 }, { "epoch": 0.5230944687321802, "grad_norm": 0.4034165723189431, "learning_rate": 2.3293983098846937e-05, "loss": 0.7615, "step": 344 }, { "epoch": 0.524615092187797, "grad_norm": 0.41147189894163816, "learning_rate": 2.3270916098624567e-05, "loss": 0.7498, "step": 345 }, { "epoch": 0.5261357156434138, "grad_norm": 0.4234882343930261, "learning_rate": 2.3247790904607715e-05, "loss": 0.7078, "step": 346 }, { "epoch": 0.5276563390990306, "grad_norm": 0.4238992343706461, "learning_rate": 2.322460765523423e-05, "loss": 0.769, "step": 347 }, { "epoch": 0.5291769625546474, "grad_norm": 0.4491430678713249, "learning_rate": 2.3201366489289512e-05, "loss": 0.7742, "step": 348 }, { "epoch": 0.5306975860102642, "grad_norm": 0.4445057795157137, "learning_rate": 2.317806754590568e-05, "loss": 0.7798, "step": 349 }, { "epoch": 0.532218209465881, "grad_norm": 0.45149251584132893, "learning_rate": 2.3154710964560735e-05, "loss": 0.7705, "step": 350 }, { "epoch": 0.5337388329214978, "grad_norm": 0.4379914224804398, "learning_rate": 2.313129688507772e-05, "loss": 0.7465, "step": 351 }, { "epoch": 0.5352594563771146, "grad_norm": 0.5240644813586705, "learning_rate": 2.3107825447623893e-05, "loss": 0.7792, "step": 352 }, { "epoch": 0.5367800798327315, "grad_norm": 0.43614968464292103, "learning_rate": 2.3084296792709876e-05, "loss": 0.8012, "step": 353 }, { "epoch": 0.5383007032883482, "grad_norm": 0.5238927538180428, "learning_rate": 2.3060711061188828e-05, "loss": 0.7121, "step": 354 }, { "epoch": 0.539821326743965, "grad_norm": 0.44343055771039847, "learning_rate": 2.303706839425559e-05, "loss": 0.7597, "step": 355 }, { "epoch": 0.5413419501995819, "grad_norm": 0.4760425163011581, "learning_rate": 2.301336893344584e-05, "loss": 0.7686, "step": 356 }, { "epoch": 0.5428625736551986, "grad_norm": 0.3864387482026591, "learning_rate": 2.2989612820635268e-05, "loss": 0.7634, "step": 357 }, { "epoch": 0.5443831971108154, "grad_norm": 0.4454248548842092, "learning_rate": 2.2965800198038684e-05, "loss": 0.7462, "step": 358 }, { "epoch": 0.5459038205664323, "grad_norm": 0.4045116377861445, "learning_rate": 2.2941931208209206e-05, "loss": 0.7538, "step": 359 }, { "epoch": 0.547424444022049, "grad_norm": 0.45148513549461083, "learning_rate": 2.291800599403739e-05, "loss": 0.7576, "step": 360 }, { "epoch": 0.5489450674776658, "grad_norm": 0.4144212534792813, "learning_rate": 2.289402469875037e-05, "loss": 0.7679, "step": 361 }, { "epoch": 0.5504656909332827, "grad_norm": 0.3949666631774562, "learning_rate": 2.2869987465911015e-05, "loss": 0.7334, "step": 362 }, { "epoch": 0.5519863143888994, "grad_norm": 0.592549458633791, "learning_rate": 2.284589443941706e-05, "loss": 0.766, "step": 363 }, { "epoch": 0.5535069378445162, "grad_norm": 0.3625351248308574, "learning_rate": 2.2821745763500237e-05, "loss": 0.7502, "step": 364 }, { "epoch": 0.5550275613001331, "grad_norm": 0.55689868686449, "learning_rate": 2.279754158272543e-05, "loss": 0.741, "step": 365 }, { "epoch": 0.5565481847557499, "grad_norm": 0.4232933895876237, "learning_rate": 2.2773282041989803e-05, "loss": 0.7203, "step": 366 }, { "epoch": 0.5580688082113666, "grad_norm": 0.47439658622550684, "learning_rate": 2.274896728652192e-05, "loss": 0.7413, "step": 367 }, { "epoch": 0.5595894316669835, "grad_norm": 0.4229854780360866, "learning_rate": 2.2724597461880887e-05, "loss": 0.7919, "step": 368 }, { "epoch": 0.5611100551226003, "grad_norm": 0.449535321441458, "learning_rate": 2.2700172713955486e-05, "loss": 0.7871, "step": 369 }, { "epoch": 0.562630678578217, "grad_norm": 0.42277942967988674, "learning_rate": 2.267569318896329e-05, "loss": 0.7492, "step": 370 }, { "epoch": 0.5641513020338339, "grad_norm": 0.4008266318587021, "learning_rate": 2.2651159033449787e-05, "loss": 0.7599, "step": 371 }, { "epoch": 0.5656719254894507, "grad_norm": 0.46433746216794397, "learning_rate": 2.2626570394287524e-05, "loss": 0.769, "step": 372 }, { "epoch": 0.5671925489450674, "grad_norm": 0.4052331130445469, "learning_rate": 2.260192741867519e-05, "loss": 0.7584, "step": 373 }, { "epoch": 0.5687131724006843, "grad_norm": 0.4069059419780127, "learning_rate": 2.2577230254136776e-05, "loss": 0.7605, "step": 374 }, { "epoch": 0.5702337958563011, "grad_norm": 0.45966108042874787, "learning_rate": 2.2552479048520657e-05, "loss": 0.771, "step": 375 }, { "epoch": 0.5717544193119178, "grad_norm": 0.4230352997839189, "learning_rate": 2.2527673949998737e-05, "loss": 0.7859, "step": 376 }, { "epoch": 0.5732750427675347, "grad_norm": 0.42063368235914506, "learning_rate": 2.250281510706553e-05, "loss": 0.7962, "step": 377 }, { "epoch": 0.5747956662231515, "grad_norm": 0.41245979398339844, "learning_rate": 2.2477902668537307e-05, "loss": 0.7692, "step": 378 }, { "epoch": 0.5763162896787682, "grad_norm": 0.42587671502291896, "learning_rate": 2.2452936783551174e-05, "loss": 0.7659, "step": 379 }, { "epoch": 0.5778369131343851, "grad_norm": 0.39899767427234584, "learning_rate": 2.2427917601564194e-05, "loss": 0.7576, "step": 380 }, { "epoch": 0.5793575365900019, "grad_norm": 0.4504820473523934, "learning_rate": 2.2402845272352495e-05, "loss": 0.7657, "step": 381 }, { "epoch": 0.5808781600456187, "grad_norm": 0.41018097177091434, "learning_rate": 2.237771994601036e-05, "loss": 0.7477, "step": 382 }, { "epoch": 0.5823987835012355, "grad_norm": 0.38305001213351025, "learning_rate": 2.235254177294934e-05, "loss": 0.7232, "step": 383 }, { "epoch": 0.5839194069568523, "grad_norm": 0.47095376013594553, "learning_rate": 2.232731090389736e-05, "loss": 0.7741, "step": 384 }, { "epoch": 0.5854400304124691, "grad_norm": 0.4636124674958436, "learning_rate": 2.2302027489897786e-05, "loss": 0.7517, "step": 385 }, { "epoch": 0.5869606538680859, "grad_norm": 0.44034211588944, "learning_rate": 2.2276691682308565e-05, "loss": 0.7418, "step": 386 }, { "epoch": 0.5884812773237027, "grad_norm": 0.4181155110302704, "learning_rate": 2.2251303632801276e-05, "loss": 0.7565, "step": 387 }, { "epoch": 0.5900019007793195, "grad_norm": 0.44045436172545194, "learning_rate": 2.2225863493360255e-05, "loss": 0.743, "step": 388 }, { "epoch": 0.5915225242349363, "grad_norm": 0.3847833966152017, "learning_rate": 2.220037141628167e-05, "loss": 0.7338, "step": 389 }, { "epoch": 0.5930431476905531, "grad_norm": 0.42280183460806886, "learning_rate": 2.2174827554172604e-05, "loss": 0.7704, "step": 390 }, { "epoch": 0.59456377114617, "grad_norm": 0.4636797659898542, "learning_rate": 2.2149232059950155e-05, "loss": 0.8102, "step": 391 }, { "epoch": 0.5960843946017867, "grad_norm": 0.41898988989505653, "learning_rate": 2.2123585086840513e-05, "loss": 0.7743, "step": 392 }, { "epoch": 0.5976050180574035, "grad_norm": 0.4669187287489928, "learning_rate": 2.2097886788378042e-05, "loss": 0.7809, "step": 393 }, { "epoch": 0.5991256415130204, "grad_norm": 0.3896079438766509, "learning_rate": 2.2072137318404363e-05, "loss": 0.7482, "step": 394 }, { "epoch": 0.6006462649686372, "grad_norm": 0.49028722719405726, "learning_rate": 2.2046336831067438e-05, "loss": 0.7412, "step": 395 }, { "epoch": 0.6021668884242539, "grad_norm": 0.40246044806361353, "learning_rate": 2.2020485480820634e-05, "loss": 0.736, "step": 396 }, { "epoch": 0.6036875118798708, "grad_norm": 0.39677182146473855, "learning_rate": 2.199458342242181e-05, "loss": 0.7436, "step": 397 }, { "epoch": 0.6052081353354876, "grad_norm": 0.42836586963405954, "learning_rate": 2.196863081093239e-05, "loss": 0.7649, "step": 398 }, { "epoch": 0.6067287587911043, "grad_norm": 0.4469293995373356, "learning_rate": 2.1942627801716423e-05, "loss": 0.7442, "step": 399 }, { "epoch": 0.6082493822467212, "grad_norm": 0.4004274410012105, "learning_rate": 2.1916574550439677e-05, "loss": 0.7694, "step": 400 }, { "epoch": 0.609770005702338, "grad_norm": 0.4128648841658959, "learning_rate": 2.189047121306867e-05, "loss": 0.7628, "step": 401 }, { "epoch": 0.6112906291579547, "grad_norm": 0.397595260719658, "learning_rate": 2.186431794586978e-05, "loss": 0.7335, "step": 402 }, { "epoch": 0.6128112526135716, "grad_norm": 0.39655951051855476, "learning_rate": 2.1838114905408268e-05, "loss": 0.7488, "step": 403 }, { "epoch": 0.6143318760691884, "grad_norm": 0.3954068854911475, "learning_rate": 2.1811862248547378e-05, "loss": 0.7732, "step": 404 }, { "epoch": 0.6158524995248051, "grad_norm": 0.4029750154489033, "learning_rate": 2.1785560132447365e-05, "loss": 0.7635, "step": 405 }, { "epoch": 0.617373122980422, "grad_norm": 0.4988589982990038, "learning_rate": 2.1759208714564575e-05, "loss": 0.7482, "step": 406 }, { "epoch": 0.6188937464360388, "grad_norm": 0.3830796517420403, "learning_rate": 2.1732808152650507e-05, "loss": 0.7059, "step": 407 }, { "epoch": 0.6204143698916555, "grad_norm": 0.4928324719853511, "learning_rate": 2.1706358604750843e-05, "loss": 0.75, "step": 408 }, { "epoch": 0.6219349933472724, "grad_norm": 0.41872118978417633, "learning_rate": 2.1679860229204526e-05, "loss": 0.7799, "step": 409 }, { "epoch": 0.6234556168028892, "grad_norm": 0.44112153018658945, "learning_rate": 2.1653313184642798e-05, "loss": 0.7554, "step": 410 }, { "epoch": 0.624976240258506, "grad_norm": 0.44207424430007414, "learning_rate": 2.162671762998826e-05, "loss": 0.7303, "step": 411 }, { "epoch": 0.6264968637141228, "grad_norm": 0.4287487064749124, "learning_rate": 2.1600073724453913e-05, "loss": 0.7311, "step": 412 }, { "epoch": 0.6280174871697396, "grad_norm": 0.4795008086486292, "learning_rate": 2.157338162754221e-05, "loss": 0.7538, "step": 413 }, { "epoch": 0.6295381106253564, "grad_norm": 0.4116045721062874, "learning_rate": 2.1546641499044105e-05, "loss": 0.7526, "step": 414 }, { "epoch": 0.6310587340809732, "grad_norm": 0.4601608034979367, "learning_rate": 2.1519853499038075e-05, "loss": 0.7891, "step": 415 }, { "epoch": 0.63257935753659, "grad_norm": 0.43000044965092804, "learning_rate": 2.14930177878892e-05, "loss": 0.7347, "step": 416 }, { "epoch": 0.6340999809922068, "grad_norm": 0.43136823831520454, "learning_rate": 2.1466134526248154e-05, "loss": 0.778, "step": 417 }, { "epoch": 0.6356206044478236, "grad_norm": 0.40082789389914825, "learning_rate": 2.1439203875050295e-05, "loss": 0.7435, "step": 418 }, { "epoch": 0.6371412279034404, "grad_norm": 0.46281961581743375, "learning_rate": 2.1412225995514658e-05, "loss": 0.7644, "step": 419 }, { "epoch": 0.6386618513590572, "grad_norm": 0.39180679623310466, "learning_rate": 2.1385201049143025e-05, "loss": 0.7401, "step": 420 }, { "epoch": 0.640182474814674, "grad_norm": 0.44410256733578407, "learning_rate": 2.135812919771892e-05, "loss": 0.749, "step": 421 }, { "epoch": 0.6417030982702908, "grad_norm": 0.4332510457584124, "learning_rate": 2.133101060330668e-05, "loss": 0.724, "step": 422 }, { "epoch": 0.6432237217259076, "grad_norm": 0.434836178784241, "learning_rate": 2.1303845428250453e-05, "loss": 0.7905, "step": 423 }, { "epoch": 0.6447443451815245, "grad_norm": 0.5203348960832226, "learning_rate": 2.127663383517326e-05, "loss": 0.7775, "step": 424 }, { "epoch": 0.6462649686371412, "grad_norm": 0.4239061363149275, "learning_rate": 2.1249375986975975e-05, "loss": 0.7547, "step": 425 }, { "epoch": 0.647785592092758, "grad_norm": 0.5531850395502146, "learning_rate": 2.1222072046836397e-05, "loss": 0.7562, "step": 426 }, { "epoch": 0.6493062155483749, "grad_norm": 0.44407831554002236, "learning_rate": 2.119472217820824e-05, "loss": 0.7563, "step": 427 }, { "epoch": 0.6508268390039916, "grad_norm": 0.46682309596014615, "learning_rate": 2.1167326544820175e-05, "loss": 0.757, "step": 428 }, { "epoch": 0.6523474624596084, "grad_norm": 0.4356058035250766, "learning_rate": 2.1139885310674832e-05, "loss": 0.7598, "step": 429 }, { "epoch": 0.6538680859152253, "grad_norm": 0.4201883171891972, "learning_rate": 2.1112398640047842e-05, "loss": 0.7347, "step": 430 }, { "epoch": 0.655388709370842, "grad_norm": 0.42395223035483165, "learning_rate": 2.1084866697486813e-05, "loss": 0.7414, "step": 431 }, { "epoch": 0.6569093328264588, "grad_norm": 0.39048036764414096, "learning_rate": 2.1057289647810407e-05, "loss": 0.7338, "step": 432 }, { "epoch": 0.6584299562820757, "grad_norm": 0.44299976473368746, "learning_rate": 2.1029667656107295e-05, "loss": 0.7791, "step": 433 }, { "epoch": 0.6599505797376924, "grad_norm": 0.500893872138014, "learning_rate": 2.100200088773519e-05, "loss": 0.7543, "step": 434 }, { "epoch": 0.6614712031933092, "grad_norm": 0.49372215416444076, "learning_rate": 2.0974289508319867e-05, "loss": 0.7676, "step": 435 }, { "epoch": 0.6629918266489261, "grad_norm": 0.46705701804279404, "learning_rate": 2.094653368375416e-05, "loss": 0.7631, "step": 436 }, { "epoch": 0.6645124501045429, "grad_norm": 0.5152291637276656, "learning_rate": 2.0918733580196976e-05, "loss": 0.778, "step": 437 }, { "epoch": 0.6660330735601596, "grad_norm": 0.4597895523784997, "learning_rate": 2.089088936407228e-05, "loss": 0.7411, "step": 438 }, { "epoch": 0.6675536970157765, "grad_norm": 0.43780405947061785, "learning_rate": 2.0863001202068135e-05, "loss": 0.7234, "step": 439 }, { "epoch": 0.6690743204713933, "grad_norm": 0.48252738543172174, "learning_rate": 2.083506926113568e-05, "loss": 0.7537, "step": 440 }, { "epoch": 0.67059494392701, "grad_norm": 0.386483976594262, "learning_rate": 2.080709370848812e-05, "loss": 0.7385, "step": 441 }, { "epoch": 0.6721155673826269, "grad_norm": 0.4545211695758658, "learning_rate": 2.077907471159975e-05, "loss": 0.7669, "step": 442 }, { "epoch": 0.6736361908382437, "grad_norm": 0.4105864152211242, "learning_rate": 2.075101243820494e-05, "loss": 0.748, "step": 443 }, { "epoch": 0.6751568142938604, "grad_norm": 0.4504033111976225, "learning_rate": 2.0722907056297144e-05, "loss": 0.7364, "step": 444 }, { "epoch": 0.6766774377494773, "grad_norm": 0.43800556294426285, "learning_rate": 2.0694758734127862e-05, "loss": 0.7816, "step": 445 }, { "epoch": 0.6781980612050941, "grad_norm": 0.4525088971872918, "learning_rate": 2.066656764020567e-05, "loss": 0.7795, "step": 446 }, { "epoch": 0.6797186846607108, "grad_norm": 0.48236560996254046, "learning_rate": 2.0638333943295188e-05, "loss": 0.8176, "step": 447 }, { "epoch": 0.6812393081163277, "grad_norm": 0.41583212690425186, "learning_rate": 2.0610057812416076e-05, "loss": 0.7233, "step": 448 }, { "epoch": 0.6827599315719445, "grad_norm": 0.45116571513530795, "learning_rate": 2.0581739416842023e-05, "loss": 0.7537, "step": 449 }, { "epoch": 0.6842805550275614, "grad_norm": 0.3988713943698696, "learning_rate": 2.0553378926099744e-05, "loss": 0.7815, "step": 450 }, { "epoch": 0.6858011784831781, "grad_norm": 0.4119407206884566, "learning_rate": 2.0524976509967945e-05, "loss": 0.7884, "step": 451 }, { "epoch": 0.6873218019387949, "grad_norm": 0.4826174722158836, "learning_rate": 2.0496532338476314e-05, "loss": 0.7514, "step": 452 }, { "epoch": 0.6888424253944118, "grad_norm": 0.4442915093907275, "learning_rate": 2.046804658190451e-05, "loss": 0.7724, "step": 453 }, { "epoch": 0.6903630488500285, "grad_norm": 0.4078747197054875, "learning_rate": 2.0439519410781135e-05, "loss": 0.7422, "step": 454 }, { "epoch": 0.6918836723056453, "grad_norm": 0.40352123173925075, "learning_rate": 2.0410950995882723e-05, "loss": 0.7473, "step": 455 }, { "epoch": 0.6934042957612622, "grad_norm": 0.37505514036015986, "learning_rate": 2.0382341508232713e-05, "loss": 0.7315, "step": 456 }, { "epoch": 0.6949249192168789, "grad_norm": 0.4001378862906619, "learning_rate": 2.0353691119100415e-05, "loss": 0.7503, "step": 457 }, { "epoch": 0.6964455426724957, "grad_norm": 0.4227756951676847, "learning_rate": 2.0325e-05, "loss": 0.7604, "step": 458 }, { "epoch": 0.6979661661281126, "grad_norm": 0.3771305192233914, "learning_rate": 2.029626832268947e-05, "loss": 0.7336, "step": 459 }, { "epoch": 0.6994867895837293, "grad_norm": 0.36251000512548065, "learning_rate": 2.0267496259169623e-05, "loss": 0.7139, "step": 460 }, { "epoch": 0.7010074130393461, "grad_norm": 0.3633369727991047, "learning_rate": 2.023868398168302e-05, "loss": 0.7254, "step": 461 }, { "epoch": 0.702528036494963, "grad_norm": 0.3722032410661908, "learning_rate": 2.0209831662712973e-05, "loss": 0.7291, "step": 462 }, { "epoch": 0.7040486599505797, "grad_norm": 0.39237895415677215, "learning_rate": 2.0180939474982496e-05, "loss": 0.7807, "step": 463 }, { "epoch": 0.7055692834061965, "grad_norm": 0.38128737748687397, "learning_rate": 2.015200759145327e-05, "loss": 0.7525, "step": 464 }, { "epoch": 0.7070899068618134, "grad_norm": 0.39388751320764215, "learning_rate": 2.0123036185324624e-05, "loss": 0.765, "step": 465 }, { "epoch": 0.7086105303174302, "grad_norm": 0.4161404920966044, "learning_rate": 2.0094025430032477e-05, "loss": 0.7746, "step": 466 }, { "epoch": 0.7101311537730469, "grad_norm": 0.38567602060145806, "learning_rate": 2.006497549924831e-05, "loss": 0.7525, "step": 467 }, { "epoch": 0.7116517772286638, "grad_norm": 0.3878678822202346, "learning_rate": 2.0035886566878142e-05, "loss": 0.7363, "step": 468 }, { "epoch": 0.7131724006842806, "grad_norm": 0.4189479849973627, "learning_rate": 2.0006758807061447e-05, "loss": 0.7202, "step": 469 }, { "epoch": 0.7146930241398973, "grad_norm": 0.35897122699509787, "learning_rate": 1.997759239417016e-05, "loss": 0.711, "step": 470 }, { "epoch": 0.7162136475955142, "grad_norm": 0.3719817122737225, "learning_rate": 1.9948387502807602e-05, "loss": 0.7329, "step": 471 }, { "epoch": 0.717734271051131, "grad_norm": 0.38587499281174487, "learning_rate": 1.991914430780744e-05, "loss": 0.7414, "step": 472 }, { "epoch": 0.7192548945067477, "grad_norm": 0.40031974274913, "learning_rate": 1.988986298423265e-05, "loss": 0.7578, "step": 473 }, { "epoch": 0.7207755179623646, "grad_norm": 0.46518346734697724, "learning_rate": 1.986054370737446e-05, "loss": 0.7399, "step": 474 }, { "epoch": 0.7222961414179814, "grad_norm": 0.34848519411774526, "learning_rate": 1.983118665275131e-05, "loss": 0.7102, "step": 475 }, { "epoch": 0.7238167648735981, "grad_norm": 0.3891532280462324, "learning_rate": 1.9801791996107777e-05, "loss": 0.7307, "step": 476 }, { "epoch": 0.725337388329215, "grad_norm": 0.4460610364235613, "learning_rate": 1.9772359913413563e-05, "loss": 0.7933, "step": 477 }, { "epoch": 0.7268580117848318, "grad_norm": 0.3931330019630469, "learning_rate": 1.9742890580862405e-05, "loss": 0.7626, "step": 478 }, { "epoch": 0.7283786352404485, "grad_norm": 0.3869696043446492, "learning_rate": 1.9713384174871036e-05, "loss": 0.7652, "step": 479 }, { "epoch": 0.7298992586960654, "grad_norm": 0.4142673906264668, "learning_rate": 1.9683840872078135e-05, "loss": 0.7683, "step": 480 }, { "epoch": 0.7314198821516822, "grad_norm": 0.3758360055485875, "learning_rate": 1.965426084934325e-05, "loss": 0.7352, "step": 481 }, { "epoch": 0.732940505607299, "grad_norm": 0.3957928284148903, "learning_rate": 1.9624644283745763e-05, "loss": 0.7709, "step": 482 }, { "epoch": 0.7344611290629158, "grad_norm": 0.4166452839891987, "learning_rate": 1.959499135258381e-05, "loss": 0.7428, "step": 483 }, { "epoch": 0.7359817525185326, "grad_norm": 0.3930427010522597, "learning_rate": 1.9565302233373233e-05, "loss": 0.7239, "step": 484 }, { "epoch": 0.7375023759741494, "grad_norm": 0.3601178498038616, "learning_rate": 1.9535577103846507e-05, "loss": 0.7099, "step": 485 }, { "epoch": 0.7390229994297662, "grad_norm": 0.4753582216812286, "learning_rate": 1.950581614195169e-05, "loss": 0.749, "step": 486 }, { "epoch": 0.740543622885383, "grad_norm": 0.3946913425771556, "learning_rate": 1.9476019525851328e-05, "loss": 0.7461, "step": 487 }, { "epoch": 0.7420642463409998, "grad_norm": 0.408093102930642, "learning_rate": 1.9446187433921438e-05, "loss": 0.7433, "step": 488 }, { "epoch": 0.7435848697966166, "grad_norm": 0.4409320643561924, "learning_rate": 1.941632004475039e-05, "loss": 0.7419, "step": 489 }, { "epoch": 0.7451054932522334, "grad_norm": 0.4107498993063142, "learning_rate": 1.938641753713787e-05, "loss": 0.7663, "step": 490 }, { "epoch": 0.7466261167078502, "grad_norm": 0.3992120160222559, "learning_rate": 1.9356480090093787e-05, "loss": 0.7549, "step": 491 }, { "epoch": 0.748146740163467, "grad_norm": 0.4323109678553807, "learning_rate": 1.9326507882837228e-05, "loss": 0.7578, "step": 492 }, { "epoch": 0.7496673636190838, "grad_norm": 0.4188967295796775, "learning_rate": 1.929650109479536e-05, "loss": 0.7623, "step": 493 }, { "epoch": 0.7511879870747006, "grad_norm": 0.4016307344915274, "learning_rate": 1.9266459905602376e-05, "loss": 0.7504, "step": 494 }, { "epoch": 0.7527086105303175, "grad_norm": 0.45431926716757953, "learning_rate": 1.9236384495098395e-05, "loss": 0.7569, "step": 495 }, { "epoch": 0.7542292339859342, "grad_norm": 0.37461201818259915, "learning_rate": 1.9206275043328413e-05, "loss": 0.7228, "step": 496 }, { "epoch": 0.755749857441551, "grad_norm": 0.44795894634744293, "learning_rate": 1.9176131730541206e-05, "loss": 0.7204, "step": 497 }, { "epoch": 0.7572704808971679, "grad_norm": 0.3739926769832927, "learning_rate": 1.9145954737188265e-05, "loss": 0.7123, "step": 498 }, { "epoch": 0.7587911043527846, "grad_norm": 0.39711092135289616, "learning_rate": 1.911574424392269e-05, "loss": 0.7532, "step": 499 }, { "epoch": 0.7603117278084014, "grad_norm": 0.3754721725582843, "learning_rate": 1.9085500431598156e-05, "loss": 0.6882, "step": 500 }, { "epoch": 0.7618323512640183, "grad_norm": 0.39959794535350796, "learning_rate": 1.9055223481267776e-05, "loss": 0.7479, "step": 501 }, { "epoch": 0.763352974719635, "grad_norm": 0.35365524692767664, "learning_rate": 1.9024913574183046e-05, "loss": 0.7487, "step": 502 }, { "epoch": 0.7648735981752518, "grad_norm": 0.4166202806824158, "learning_rate": 1.899457089179276e-05, "loss": 0.7574, "step": 503 }, { "epoch": 0.7663942216308687, "grad_norm": 0.3915634338240503, "learning_rate": 1.896419561574193e-05, "loss": 0.7325, "step": 504 }, { "epoch": 0.7679148450864854, "grad_norm": 0.48602884460303647, "learning_rate": 1.8933787927870668e-05, "loss": 0.7286, "step": 505 }, { "epoch": 0.7694354685421022, "grad_norm": 0.36368946637812705, "learning_rate": 1.890334801021313e-05, "loss": 0.7562, "step": 506 }, { "epoch": 0.7709560919977191, "grad_norm": 0.45475514908964265, "learning_rate": 1.8872876044996417e-05, "loss": 0.7414, "step": 507 }, { "epoch": 0.7724767154533358, "grad_norm": 0.4086589491103268, "learning_rate": 1.8842372214639475e-05, "loss": 0.7661, "step": 508 }, { "epoch": 0.7739973389089527, "grad_norm": 0.4038130105419086, "learning_rate": 1.8811836701752004e-05, "loss": 0.733, "step": 509 }, { "epoch": 0.7755179623645695, "grad_norm": 0.4234636977161692, "learning_rate": 1.8781269689133386e-05, "loss": 0.8073, "step": 510 }, { "epoch": 0.7770385858201863, "grad_norm": 0.395363248052582, "learning_rate": 1.8750671359771563e-05, "loss": 0.7298, "step": 511 }, { "epoch": 0.7785592092758031, "grad_norm": 0.49129541089929685, "learning_rate": 1.8720041896841943e-05, "loss": 0.7118, "step": 512 }, { "epoch": 0.7800798327314199, "grad_norm": 0.41442318450837484, "learning_rate": 1.868938148370634e-05, "loss": 0.7451, "step": 513 }, { "epoch": 0.7816004561870367, "grad_norm": 0.4332314377021711, "learning_rate": 1.8658690303911834e-05, "loss": 0.7456, "step": 514 }, { "epoch": 0.7831210796426535, "grad_norm": 0.49756006836526884, "learning_rate": 1.8627968541189676e-05, "loss": 0.7449, "step": 515 }, { "epoch": 0.7846417030982703, "grad_norm": 0.39053028376926463, "learning_rate": 1.8597216379454236e-05, "loss": 0.7353, "step": 516 }, { "epoch": 0.7861623265538871, "grad_norm": 0.42683269902895027, "learning_rate": 1.8566434002801828e-05, "loss": 0.73, "step": 517 }, { "epoch": 0.787682950009504, "grad_norm": 0.4414935000692991, "learning_rate": 1.8535621595509665e-05, "loss": 0.7222, "step": 518 }, { "epoch": 0.7892035734651207, "grad_norm": 0.36090118649581926, "learning_rate": 1.8504779342034735e-05, "loss": 0.7206, "step": 519 }, { "epoch": 0.7907241969207375, "grad_norm": 0.44956102255887004, "learning_rate": 1.8473907427012702e-05, "loss": 0.7809, "step": 520 }, { "epoch": 0.7922448203763544, "grad_norm": 0.400049613455352, "learning_rate": 1.8443006035256794e-05, "loss": 0.7165, "step": 521 }, { "epoch": 0.7937654438319711, "grad_norm": 0.4096379673462462, "learning_rate": 1.8412075351756694e-05, "loss": 0.7188, "step": 522 }, { "epoch": 0.7952860672875879, "grad_norm": 0.3813329658819688, "learning_rate": 1.8381115561677448e-05, "loss": 0.7434, "step": 523 }, { "epoch": 0.7968066907432048, "grad_norm": 0.3883504678303308, "learning_rate": 1.835012685035834e-05, "loss": 0.7538, "step": 524 }, { "epoch": 0.7983273141988215, "grad_norm": 0.4810224536969676, "learning_rate": 1.831910940331179e-05, "loss": 0.732, "step": 525 }, { "epoch": 0.7998479376544383, "grad_norm": 0.4274875955939722, "learning_rate": 1.828806340622226e-05, "loss": 0.7654, "step": 526 }, { "epoch": 0.8013685611100552, "grad_norm": 0.6167789867453826, "learning_rate": 1.8256989044945094e-05, "loss": 0.723, "step": 527 }, { "epoch": 0.8028891845656719, "grad_norm": 0.38651635512015386, "learning_rate": 1.8225886505505468e-05, "loss": 0.7301, "step": 528 }, { "epoch": 0.8044098080212887, "grad_norm": 0.5105848984644641, "learning_rate": 1.8194755974097218e-05, "loss": 0.7379, "step": 529 }, { "epoch": 0.8059304314769056, "grad_norm": 0.4337444604279403, "learning_rate": 1.8163597637081783e-05, "loss": 0.7438, "step": 530 }, { "epoch": 0.8074510549325223, "grad_norm": 0.38311547081568653, "learning_rate": 1.8132411680987025e-05, "loss": 0.7166, "step": 531 }, { "epoch": 0.8089716783881391, "grad_norm": 0.40940875061446746, "learning_rate": 1.8101198292506176e-05, "loss": 0.7377, "step": 532 }, { "epoch": 0.810492301843756, "grad_norm": 0.39069303091376745, "learning_rate": 1.8069957658496668e-05, "loss": 0.7362, "step": 533 }, { "epoch": 0.8120129252993727, "grad_norm": 0.4034863997233365, "learning_rate": 1.8038689965979057e-05, "loss": 0.7845, "step": 534 }, { "epoch": 0.8135335487549895, "grad_norm": 0.3743182233611382, "learning_rate": 1.8007395402135864e-05, "loss": 0.7249, "step": 535 }, { "epoch": 0.8150541722106064, "grad_norm": 0.4151612507277004, "learning_rate": 1.7976074154310488e-05, "loss": 0.7501, "step": 536 }, { "epoch": 0.8165747956662232, "grad_norm": 0.4052560608313417, "learning_rate": 1.7944726410006063e-05, "loss": 0.7397, "step": 537 }, { "epoch": 0.8180954191218399, "grad_norm": 0.4204385753908119, "learning_rate": 1.7913352356884348e-05, "loss": 0.7269, "step": 538 }, { "epoch": 0.8196160425774568, "grad_norm": 0.36912065852511794, "learning_rate": 1.7881952182764595e-05, "loss": 0.7216, "step": 539 }, { "epoch": 0.8211366660330736, "grad_norm": 0.4091528689771071, "learning_rate": 1.785052607562243e-05, "loss": 0.741, "step": 540 }, { "epoch": 0.8226572894886903, "grad_norm": 0.3712296997315177, "learning_rate": 1.7819074223588718e-05, "loss": 0.7361, "step": 541 }, { "epoch": 0.8241779129443072, "grad_norm": 0.41163387712501454, "learning_rate": 1.7787596814948465e-05, "loss": 0.7153, "step": 542 }, { "epoch": 0.825698536399924, "grad_norm": 0.3881722683934304, "learning_rate": 1.7756094038139645e-05, "loss": 0.7354, "step": 543 }, { "epoch": 0.8272191598555407, "grad_norm": 0.37426692455200733, "learning_rate": 1.7724566081752113e-05, "loss": 0.7132, "step": 544 }, { "epoch": 0.8287397833111576, "grad_norm": 0.36537784686372354, "learning_rate": 1.7693013134526457e-05, "loss": 0.6791, "step": 545 }, { "epoch": 0.8302604067667744, "grad_norm": 0.3789388855291124, "learning_rate": 1.7661435385352875e-05, "loss": 0.7158, "step": 546 }, { "epoch": 0.8317810302223911, "grad_norm": 0.41030137404819805, "learning_rate": 1.762983302327003e-05, "loss": 0.6648, "step": 547 }, { "epoch": 0.833301653678008, "grad_norm": 0.3866957065876247, "learning_rate": 1.7598206237463935e-05, "loss": 0.7383, "step": 548 }, { "epoch": 0.8348222771336248, "grad_norm": 0.42152702106156975, "learning_rate": 1.7566555217266816e-05, "loss": 0.7486, "step": 549 }, { "epoch": 0.8363429005892415, "grad_norm": 0.3837180537567676, "learning_rate": 1.753488015215598e-05, "loss": 0.7139, "step": 550 }, { "epoch": 0.8378635240448584, "grad_norm": 0.45359893466752715, "learning_rate": 1.750318123175267e-05, "loss": 0.7415, "step": 551 }, { "epoch": 0.8393841475004752, "grad_norm": 0.37167124021677417, "learning_rate": 1.7471458645820944e-05, "loss": 0.7341, "step": 552 }, { "epoch": 0.840904770956092, "grad_norm": 0.386261837626807, "learning_rate": 1.7439712584266523e-05, "loss": 0.7753, "step": 553 }, { "epoch": 0.8424253944117088, "grad_norm": 0.3607815357285951, "learning_rate": 1.740794323713568e-05, "loss": 0.7295, "step": 554 }, { "epoch": 0.8439460178673256, "grad_norm": 0.3805870265080064, "learning_rate": 1.7376150794614067e-05, "loss": 0.742, "step": 555 }, { "epoch": 0.8454666413229424, "grad_norm": 0.3634857699838318, "learning_rate": 1.7344335447025617e-05, "loss": 0.7405, "step": 556 }, { "epoch": 0.8469872647785592, "grad_norm": 0.3830792827233582, "learning_rate": 1.731249738483136e-05, "loss": 0.7593, "step": 557 }, { "epoch": 0.848507888234176, "grad_norm": 0.37680323864124654, "learning_rate": 1.728063679862832e-05, "loss": 0.7374, "step": 558 }, { "epoch": 0.8500285116897928, "grad_norm": 0.3812378150946191, "learning_rate": 1.724875387914836e-05, "loss": 0.7317, "step": 559 }, { "epoch": 0.8515491351454096, "grad_norm": 0.3955303602601571, "learning_rate": 1.7216848817257034e-05, "loss": 0.7558, "step": 560 }, { "epoch": 0.8530697586010264, "grad_norm": 0.36089368326825794, "learning_rate": 1.718492180395245e-05, "loss": 0.7124, "step": 561 }, { "epoch": 0.8545903820566432, "grad_norm": 0.39264324951073387, "learning_rate": 1.715297303036414e-05, "loss": 0.6976, "step": 562 }, { "epoch": 0.85611100551226, "grad_norm": 0.3583167479634998, "learning_rate": 1.712100268775188e-05, "loss": 0.7459, "step": 563 }, { "epoch": 0.8576316289678768, "grad_norm": 0.3652537723370093, "learning_rate": 1.708901096750459e-05, "loss": 0.752, "step": 564 }, { "epoch": 0.8591522524234936, "grad_norm": 0.35829414089305417, "learning_rate": 1.7056998061139157e-05, "loss": 0.7293, "step": 565 }, { "epoch": 0.8606728758791105, "grad_norm": 0.41471406367802255, "learning_rate": 1.70249641602993e-05, "loss": 0.7134, "step": 566 }, { "epoch": 0.8621934993347272, "grad_norm": 0.40497506621492396, "learning_rate": 1.6992909456754408e-05, "loss": 0.7266, "step": 567 }, { "epoch": 0.8637141227903441, "grad_norm": 0.5118453814476593, "learning_rate": 1.6960834142398424e-05, "loss": 0.7524, "step": 568 }, { "epoch": 0.8652347462459609, "grad_norm": 0.4027525684201806, "learning_rate": 1.6928738409248668e-05, "loss": 0.7209, "step": 569 }, { "epoch": 0.8667553697015776, "grad_norm": 0.39420637097855954, "learning_rate": 1.689662244944469e-05, "loss": 0.7445, "step": 570 }, { "epoch": 0.8682759931571945, "grad_norm": 0.4106148870965668, "learning_rate": 1.686448645524714e-05, "loss": 0.7655, "step": 571 }, { "epoch": 0.8697966166128113, "grad_norm": 0.3579948523808814, "learning_rate": 1.6832330619036584e-05, "loss": 0.7165, "step": 572 }, { "epoch": 0.871317240068428, "grad_norm": 0.3751959854555261, "learning_rate": 1.6800155133312387e-05, "loss": 0.7708, "step": 573 }, { "epoch": 0.8728378635240449, "grad_norm": 0.39336551188405533, "learning_rate": 1.6767960190691538e-05, "loss": 0.7161, "step": 574 }, { "epoch": 0.8743584869796617, "grad_norm": 0.38785678050667954, "learning_rate": 1.6735745983907504e-05, "loss": 0.719, "step": 575 }, { "epoch": 0.8758791104352784, "grad_norm": 0.3696861272192392, "learning_rate": 1.6703512705809085e-05, "loss": 0.7434, "step": 576 }, { "epoch": 0.8773997338908953, "grad_norm": 0.3807058540190109, "learning_rate": 1.667126054935923e-05, "loss": 0.7336, "step": 577 }, { "epoch": 0.8789203573465121, "grad_norm": 0.3907235180627425, "learning_rate": 1.6638989707633927e-05, "loss": 0.723, "step": 578 }, { "epoch": 0.8804409808021288, "grad_norm": 0.39300541368827063, "learning_rate": 1.6606700373821005e-05, "loss": 0.7313, "step": 579 }, { "epoch": 0.8819616042577457, "grad_norm": 0.3994037757594598, "learning_rate": 1.6574392741219003e-05, "loss": 0.7366, "step": 580 }, { "epoch": 0.8834822277133625, "grad_norm": 0.35949981685128496, "learning_rate": 1.6542067003236e-05, "loss": 0.7288, "step": 581 }, { "epoch": 0.8850028511689793, "grad_norm": 0.3987872705670089, "learning_rate": 1.6509723353388463e-05, "loss": 0.6909, "step": 582 }, { "epoch": 0.8865234746245961, "grad_norm": 0.3829862590465747, "learning_rate": 1.6477361985300094e-05, "loss": 0.6989, "step": 583 }, { "epoch": 0.8880440980802129, "grad_norm": 0.39080192305445455, "learning_rate": 1.6444983092700657e-05, "loss": 0.7357, "step": 584 }, { "epoch": 0.8895647215358297, "grad_norm": 0.39244835712820425, "learning_rate": 1.6412586869424823e-05, "loss": 0.6906, "step": 585 }, { "epoch": 0.8910853449914465, "grad_norm": 0.38987649746021225, "learning_rate": 1.6380173509411024e-05, "loss": 0.7655, "step": 586 }, { "epoch": 0.8926059684470633, "grad_norm": 0.3792483881544404, "learning_rate": 1.634774320670027e-05, "loss": 0.7414, "step": 587 }, { "epoch": 0.8941265919026801, "grad_norm": 0.37786695217123767, "learning_rate": 1.6315296155435003e-05, "loss": 0.7523, "step": 588 }, { "epoch": 0.8956472153582969, "grad_norm": 0.40039846555450576, "learning_rate": 1.6282832549857926e-05, "loss": 0.7358, "step": 589 }, { "epoch": 0.8971678388139137, "grad_norm": 0.3928470737436745, "learning_rate": 1.6250352584310847e-05, "loss": 0.755, "step": 590 }, { "epoch": 0.8986884622695305, "grad_norm": 0.3903264232496443, "learning_rate": 1.6217856453233505e-05, "loss": 0.7334, "step": 591 }, { "epoch": 0.9002090857251474, "grad_norm": 0.3685085038233517, "learning_rate": 1.6185344351162427e-05, "loss": 0.7137, "step": 592 }, { "epoch": 0.9017297091807641, "grad_norm": 0.39206546198749437, "learning_rate": 1.6152816472729734e-05, "loss": 0.7451, "step": 593 }, { "epoch": 0.9032503326363809, "grad_norm": 0.379717976951851, "learning_rate": 1.6120273012662e-05, "loss": 0.7153, "step": 594 }, { "epoch": 0.9047709560919978, "grad_norm": 0.38621564008551273, "learning_rate": 1.608771416577908e-05, "loss": 0.7559, "step": 595 }, { "epoch": 0.9062915795476145, "grad_norm": 0.4060366023992974, "learning_rate": 1.6055140126992933e-05, "loss": 0.7386, "step": 596 }, { "epoch": 0.9078122030032313, "grad_norm": 0.35024041522069665, "learning_rate": 1.6022551091306466e-05, "loss": 0.7248, "step": 597 }, { "epoch": 0.9093328264588482, "grad_norm": 0.4429630612789365, "learning_rate": 1.5989947253812372e-05, "loss": 0.7344, "step": 598 }, { "epoch": 0.9108534499144649, "grad_norm": 0.35160112786541303, "learning_rate": 1.5957328809691947e-05, "loss": 0.7136, "step": 599 }, { "epoch": 0.9123740733700817, "grad_norm": 0.3607594118044688, "learning_rate": 1.592469595421393e-05, "loss": 0.7152, "step": 600 }, { "epoch": 0.9138946968256986, "grad_norm": 0.3888604747890697, "learning_rate": 1.589204888273333e-05, "loss": 0.7214, "step": 601 }, { "epoch": 0.9154153202813153, "grad_norm": 0.36939791749863715, "learning_rate": 1.585938779069027e-05, "loss": 0.724, "step": 602 }, { "epoch": 0.9169359437369321, "grad_norm": 0.3492211353535467, "learning_rate": 1.5826712873608794e-05, "loss": 0.7059, "step": 603 }, { "epoch": 0.918456567192549, "grad_norm": 0.3485676743224654, "learning_rate": 1.5794024327095722e-05, "loss": 0.6882, "step": 604 }, { "epoch": 0.9199771906481657, "grad_norm": 0.34637043053576727, "learning_rate": 1.5761322346839444e-05, "loss": 0.7329, "step": 605 }, { "epoch": 0.9214978141037825, "grad_norm": 0.39153636395112856, "learning_rate": 1.5728607128608797e-05, "loss": 0.7039, "step": 606 }, { "epoch": 0.9230184375593994, "grad_norm": 0.36173214969406103, "learning_rate": 1.5695878868251847e-05, "loss": 0.7238, "step": 607 }, { "epoch": 0.9245390610150162, "grad_norm": 0.3761817399411723, "learning_rate": 1.5663137761694748e-05, "loss": 0.7287, "step": 608 }, { "epoch": 0.9260596844706329, "grad_norm": 0.35432511798173993, "learning_rate": 1.5630384004940545e-05, "loss": 0.7169, "step": 609 }, { "epoch": 0.9275803079262498, "grad_norm": 0.37885488266265444, "learning_rate": 1.559761779406803e-05, "loss": 0.7456, "step": 610 }, { "epoch": 0.9291009313818666, "grad_norm": 0.3834340488104107, "learning_rate": 1.5564839325230532e-05, "loss": 0.7145, "step": 611 }, { "epoch": 0.9306215548374833, "grad_norm": 0.36202133079807264, "learning_rate": 1.5532048794654782e-05, "loss": 0.7408, "step": 612 }, { "epoch": 0.9321421782931002, "grad_norm": 0.39335695813338395, "learning_rate": 1.549924639863969e-05, "loss": 0.7403, "step": 613 }, { "epoch": 0.933662801748717, "grad_norm": 0.3933661187801479, "learning_rate": 1.546643233355523e-05, "loss": 0.7082, "step": 614 }, { "epoch": 0.9351834252043337, "grad_norm": 0.37654144207241497, "learning_rate": 1.5433606795841207e-05, "loss": 0.7435, "step": 615 }, { "epoch": 0.9367040486599506, "grad_norm": 0.44483184731571473, "learning_rate": 1.5400769982006128e-05, "loss": 0.7156, "step": 616 }, { "epoch": 0.9382246721155674, "grad_norm": 0.42682697285522647, "learning_rate": 1.536792208862598e-05, "loss": 0.696, "step": 617 }, { "epoch": 0.9397452955711841, "grad_norm": 0.40979396640326593, "learning_rate": 1.5335063312343095e-05, "loss": 0.727, "step": 618 }, { "epoch": 0.941265919026801, "grad_norm": 0.3960793962289382, "learning_rate": 1.530219384986495e-05, "loss": 0.7032, "step": 619 }, { "epoch": 0.9427865424824178, "grad_norm": 0.3857008168238225, "learning_rate": 1.5269313897962994e-05, "loss": 0.7502, "step": 620 }, { "epoch": 0.9443071659380345, "grad_norm": 0.3982600098061033, "learning_rate": 1.5236423653471463e-05, "loss": 0.7666, "step": 621 }, { "epoch": 0.9458277893936514, "grad_norm": 0.41685987364222565, "learning_rate": 1.5203523313286227e-05, "loss": 0.7036, "step": 622 }, { "epoch": 0.9473484128492682, "grad_norm": 0.3959161081756824, "learning_rate": 1.5170613074363573e-05, "loss": 0.7311, "step": 623 }, { "epoch": 0.948869036304885, "grad_norm": 0.36920330872863855, "learning_rate": 1.5137693133719065e-05, "loss": 0.7039, "step": 624 }, { "epoch": 0.9503896597605018, "grad_norm": 0.39113501294384384, "learning_rate": 1.5104763688426331e-05, "loss": 0.7506, "step": 625 }, { "epoch": 0.9519102832161186, "grad_norm": 0.39144353205492227, "learning_rate": 1.5071824935615908e-05, "loss": 0.6886, "step": 626 }, { "epoch": 0.9534309066717354, "grad_norm": 0.36594301273514485, "learning_rate": 1.5038877072474041e-05, "loss": 0.7444, "step": 627 }, { "epoch": 0.9549515301273522, "grad_norm": 0.3954444253170849, "learning_rate": 1.500592029624153e-05, "loss": 0.7147, "step": 628 }, { "epoch": 0.956472153582969, "grad_norm": 0.3773327611421358, "learning_rate": 1.4972954804212516e-05, "loss": 0.7453, "step": 629 }, { "epoch": 0.9579927770385859, "grad_norm": 0.41652271596518264, "learning_rate": 1.4939980793733329e-05, "loss": 0.7437, "step": 630 }, { "epoch": 0.9595134004942026, "grad_norm": 0.4094773012771971, "learning_rate": 1.4906998462201288e-05, "loss": 0.7654, "step": 631 }, { "epoch": 0.9610340239498194, "grad_norm": 0.3944195978868427, "learning_rate": 1.4874008007063523e-05, "loss": 0.7452, "step": 632 }, { "epoch": 0.9625546474054363, "grad_norm": 0.3707638628881551, "learning_rate": 1.48410096258158e-05, "loss": 0.7042, "step": 633 }, { "epoch": 0.964075270861053, "grad_norm": 0.3755039123645331, "learning_rate": 1.4808003516001344e-05, "loss": 0.7177, "step": 634 }, { "epoch": 0.9655958943166698, "grad_norm": 0.3723560524603749, "learning_rate": 1.4774989875209624e-05, "loss": 0.7553, "step": 635 }, { "epoch": 0.9671165177722867, "grad_norm": 0.3651623249941728, "learning_rate": 1.474196890107521e-05, "loss": 0.7404, "step": 636 }, { "epoch": 0.9686371412279035, "grad_norm": 0.37966360420328005, "learning_rate": 1.4708940791276568e-05, "loss": 0.7159, "step": 637 }, { "epoch": 0.9701577646835202, "grad_norm": 0.36168885071485135, "learning_rate": 1.4675905743534884e-05, "loss": 0.7541, "step": 638 }, { "epoch": 0.9716783881391371, "grad_norm": 0.36095700915855466, "learning_rate": 1.464286395561287e-05, "loss": 0.7092, "step": 639 }, { "epoch": 0.9731990115947539, "grad_norm": 0.43805492394151224, "learning_rate": 1.46098156253136e-05, "loss": 0.7362, "step": 640 }, { "epoch": 0.9747196350503706, "grad_norm": 0.36203353195893573, "learning_rate": 1.4576760950479295e-05, "loss": 0.7106, "step": 641 }, { "epoch": 0.9762402585059875, "grad_norm": 0.40877067220155616, "learning_rate": 1.454370012899018e-05, "loss": 0.7026, "step": 642 }, { "epoch": 0.9777608819616043, "grad_norm": 0.40374065798245223, "learning_rate": 1.4510633358763258e-05, "loss": 0.7182, "step": 643 }, { "epoch": 0.979281505417221, "grad_norm": 0.3659907217224348, "learning_rate": 1.4477560837751156e-05, "loss": 0.713, "step": 644 }, { "epoch": 0.9808021288728379, "grad_norm": 0.3622601843846982, "learning_rate": 1.4444482763940919e-05, "loss": 0.7327, "step": 645 }, { "epoch": 0.9823227523284547, "grad_norm": 0.3885583144262136, "learning_rate": 1.441139933535284e-05, "loss": 0.6978, "step": 646 }, { "epoch": 0.9838433757840714, "grad_norm": 0.37410385831599374, "learning_rate": 1.437831075003926e-05, "loss": 0.724, "step": 647 }, { "epoch": 0.9853639992396883, "grad_norm": 0.3696274971269831, "learning_rate": 1.4345217206083407e-05, "loss": 0.6965, "step": 648 }, { "epoch": 0.9868846226953051, "grad_norm": 0.41889852029503616, "learning_rate": 1.4312118901598172e-05, "loss": 0.7678, "step": 649 }, { "epoch": 0.9884052461509218, "grad_norm": 0.3429687194640586, "learning_rate": 1.4279016034724961e-05, "loss": 0.7047, "step": 650 }, { "epoch": 0.9899258696065387, "grad_norm": 0.3736752188164251, "learning_rate": 1.4245908803632486e-05, "loss": 0.7362, "step": 651 }, { "epoch": 0.9914464930621555, "grad_norm": 0.38265917497289204, "learning_rate": 1.4212797406515585e-05, "loss": 0.7225, "step": 652 }, { "epoch": 0.9929671165177723, "grad_norm": 0.37332104277853684, "learning_rate": 1.417968204159404e-05, "loss": 0.7312, "step": 653 }, { "epoch": 0.9944877399733891, "grad_norm": 0.42232449088131074, "learning_rate": 1.4146562907111377e-05, "loss": 0.7095, "step": 654 }, { "epoch": 0.9960083634290059, "grad_norm": 0.3868095376144605, "learning_rate": 1.4113440201333701e-05, "loss": 0.6992, "step": 655 }, { "epoch": 0.9975289868846227, "grad_norm": 0.42385907348580026, "learning_rate": 1.4080314122548488e-05, "loss": 0.7347, "step": 656 }, { "epoch": 0.9990496103402395, "grad_norm": 0.41348925205041664, "learning_rate": 1.4047184869063401e-05, "loss": 0.7441, "step": 657 }, { "epoch": 1.0005702337958564, "grad_norm": 0.35648966659894604, "learning_rate": 1.4014052639205127e-05, "loss": 0.6707, "step": 658 }, { "epoch": 1.002090857251473, "grad_norm": 0.34244915908174534, "learning_rate": 1.3980917631318145e-05, "loss": 0.7027, "step": 659 }, { "epoch": 1.00361148070709, "grad_norm": 0.3670346028136332, "learning_rate": 1.394778004376359e-05, "loss": 0.7363, "step": 660 }, { "epoch": 1.0051321041627068, "grad_norm": 0.391603629750881, "learning_rate": 1.3914640074918023e-05, "loss": 0.7351, "step": 661 }, { "epoch": 1.0066527276183235, "grad_norm": 0.3716653098931945, "learning_rate": 1.3881497923172269e-05, "loss": 0.7441, "step": 662 }, { "epoch": 1.0081733510739403, "grad_norm": 0.3755372180582686, "learning_rate": 1.3848353786930213e-05, "loss": 0.6965, "step": 663 }, { "epoch": 1.0096939745295572, "grad_norm": 0.381844726209517, "learning_rate": 1.3815207864607634e-05, "loss": 0.7302, "step": 664 }, { "epoch": 1.0112145979851739, "grad_norm": 0.36415027113970666, "learning_rate": 1.3782060354630988e-05, "loss": 0.6935, "step": 665 }, { "epoch": 1.0127352214407908, "grad_norm": 0.3814827703352279, "learning_rate": 1.3748911455436241e-05, "loss": 0.7288, "step": 666 }, { "epoch": 1.0142558448964076, "grad_norm": 0.3695679755631074, "learning_rate": 1.371576136546768e-05, "loss": 0.7069, "step": 667 }, { "epoch": 1.0157764683520243, "grad_norm": 0.3692625124863184, "learning_rate": 1.3682610283176713e-05, "loss": 0.7308, "step": 668 }, { "epoch": 1.0172970918076412, "grad_norm": 0.3647179500902266, "learning_rate": 1.3649458407020692e-05, "loss": 0.724, "step": 669 }, { "epoch": 1.0003801558639043, "grad_norm": 0.39033436102991464, "learning_rate": 1.3616305935461726e-05, "loss": 0.6831, "step": 670 }, { "epoch": 1.001900779319521, "grad_norm": 0.6228269584548538, "learning_rate": 1.358315306696548e-05, "loss": 0.5697, "step": 671 }, { "epoch": 1.0034214027751378, "grad_norm": 0.4482616727348143, "learning_rate": 1.355e-05, "loss": 0.5192, "step": 672 }, { "epoch": 1.0049420262307547, "grad_norm": 0.5196584857248853, "learning_rate": 1.3516846933034525e-05, "loss": 0.5425, "step": 673 }, { "epoch": 1.0064626496863713, "grad_norm": 0.5838517226989333, "learning_rate": 1.3483694064538279e-05, "loss": 0.5638, "step": 674 }, { "epoch": 1.0079832731419882, "grad_norm": 0.4333646030218863, "learning_rate": 1.3450541592979308e-05, "loss": 0.5361, "step": 675 }, { "epoch": 1.009503896597605, "grad_norm": 0.46529778581028425, "learning_rate": 1.341738971682329e-05, "loss": 0.5338, "step": 676 }, { "epoch": 1.0110245200532217, "grad_norm": 0.50417271156346, "learning_rate": 1.3384238634532324e-05, "loss": 0.5341, "step": 677 }, { "epoch": 1.0125451435088386, "grad_norm": 0.4846246525284431, "learning_rate": 1.3351088544563762e-05, "loss": 0.5454, "step": 678 }, { "epoch": 1.0140657669644555, "grad_norm": 0.43151729793082855, "learning_rate": 1.3317939645369014e-05, "loss": 0.5149, "step": 679 }, { "epoch": 1.0155863904200721, "grad_norm": 0.45346621189906927, "learning_rate": 1.3284792135392369e-05, "loss": 0.5241, "step": 680 }, { "epoch": 1.017107013875689, "grad_norm": 0.4529580634485115, "learning_rate": 1.3251646213069789e-05, "loss": 0.5252, "step": 681 }, { "epoch": 1.018627637331306, "grad_norm": 0.4309687856541669, "learning_rate": 1.3218502076827738e-05, "loss": 0.5041, "step": 682 }, { "epoch": 1.0201482607869226, "grad_norm": 0.40932019500719957, "learning_rate": 1.3185359925081977e-05, "loss": 0.5131, "step": 683 }, { "epoch": 1.0216688842425394, "grad_norm": 0.43569657302929155, "learning_rate": 1.3152219956236414e-05, "loss": 0.5058, "step": 684 }, { "epoch": 1.0231895076981563, "grad_norm": 0.41224009605600054, "learning_rate": 1.3119082368681858e-05, "loss": 0.5102, "step": 685 }, { "epoch": 1.024710131153773, "grad_norm": 0.4633671152656241, "learning_rate": 1.3085947360794881e-05, "loss": 0.5538, "step": 686 }, { "epoch": 1.0262307546093898, "grad_norm": 0.3981390104088507, "learning_rate": 1.3052815130936598e-05, "loss": 0.5053, "step": 687 }, { "epoch": 1.0277513780650067, "grad_norm": 0.42462762773917034, "learning_rate": 1.3019685877451516e-05, "loss": 0.5308, "step": 688 }, { "epoch": 1.0292720015206234, "grad_norm": 0.3928387548640261, "learning_rate": 1.29865597986663e-05, "loss": 0.4768, "step": 689 }, { "epoch": 1.0307926249762402, "grad_norm": 0.4320276733077232, "learning_rate": 1.2953437092888625e-05, "loss": 0.5221, "step": 690 }, { "epoch": 1.0323132484318571, "grad_norm": 0.4334542028887193, "learning_rate": 1.2920317958405962e-05, "loss": 0.5129, "step": 691 }, { "epoch": 1.0338338718874738, "grad_norm": 0.40146788598973115, "learning_rate": 1.2887202593484417e-05, "loss": 0.5052, "step": 692 }, { "epoch": 1.0353544953430907, "grad_norm": 0.41287405977658365, "learning_rate": 1.2854091196367518e-05, "loss": 0.5425, "step": 693 }, { "epoch": 1.0368751187987075, "grad_norm": 0.3708390977190173, "learning_rate": 1.2820983965275043e-05, "loss": 0.5238, "step": 694 }, { "epoch": 1.0383957422543242, "grad_norm": 0.35512557723314947, "learning_rate": 1.278788109840183e-05, "loss": 0.5096, "step": 695 }, { "epoch": 1.039916365709941, "grad_norm": 0.41469287495730844, "learning_rate": 1.2754782793916598e-05, "loss": 0.5159, "step": 696 }, { "epoch": 1.041436989165558, "grad_norm": 0.39113759999764336, "learning_rate": 1.2721689249960743e-05, "loss": 0.5184, "step": 697 }, { "epoch": 1.0429576126211746, "grad_norm": 0.41576137745186253, "learning_rate": 1.2688600664647168e-05, "loss": 0.5251, "step": 698 }, { "epoch": 1.0444782360767915, "grad_norm": 0.37659361938427355, "learning_rate": 1.2655517236059083e-05, "loss": 0.511, "step": 699 }, { "epoch": 1.0459988595324083, "grad_norm": 0.36191609676818887, "learning_rate": 1.2622439162248846e-05, "loss": 0.5142, "step": 700 }, { "epoch": 1.047519482988025, "grad_norm": 0.3501769260644456, "learning_rate": 1.2589366641236745e-05, "loss": 0.5086, "step": 701 }, { "epoch": 1.0490401064436419, "grad_norm": 0.374193201836417, "learning_rate": 1.2556299871009825e-05, "loss": 0.5251, "step": 702 }, { "epoch": 1.0505607298992587, "grad_norm": 0.40667780389565045, "learning_rate": 1.2523239049520706e-05, "loss": 0.5133, "step": 703 }, { "epoch": 1.0520813533548754, "grad_norm": 0.3484682044458917, "learning_rate": 1.2490184374686406e-05, "loss": 0.5078, "step": 704 }, { "epoch": 1.0536019768104923, "grad_norm": 0.3406587720902772, "learning_rate": 1.2457136044387133e-05, "loss": 0.5107, "step": 705 }, { "epoch": 1.0551226002661092, "grad_norm": 0.35983619595762045, "learning_rate": 1.242409425646512e-05, "loss": 0.5349, "step": 706 }, { "epoch": 1.0566432237217258, "grad_norm": 0.3722826497059545, "learning_rate": 1.2391059208723433e-05, "loss": 0.5118, "step": 707 }, { "epoch": 1.0581638471773427, "grad_norm": 0.367794414326388, "learning_rate": 1.2358031098924792e-05, "loss": 0.5012, "step": 708 }, { "epoch": 1.0596844706329596, "grad_norm": 0.379706737726683, "learning_rate": 1.232501012479038e-05, "loss": 0.5234, "step": 709 }, { "epoch": 1.0612050940885762, "grad_norm": 0.3808569060578708, "learning_rate": 1.2291996483998664e-05, "loss": 0.4976, "step": 710 }, { "epoch": 1.062725717544193, "grad_norm": 0.4052638293016256, "learning_rate": 1.2258990374184197e-05, "loss": 0.5115, "step": 711 }, { "epoch": 1.06424634099981, "grad_norm": 0.3884487585023922, "learning_rate": 1.222599199293648e-05, "loss": 0.5469, "step": 712 }, { "epoch": 1.0657669644554266, "grad_norm": 0.3523224705049633, "learning_rate": 1.2193001537798717e-05, "loss": 0.5031, "step": 713 }, { "epoch": 1.0672875879110435, "grad_norm": 0.4035860752243172, "learning_rate": 1.2160019206266676e-05, "loss": 0.5036, "step": 714 }, { "epoch": 1.0688082113666604, "grad_norm": 0.35467473723128, "learning_rate": 1.2127045195787486e-05, "loss": 0.5182, "step": 715 }, { "epoch": 1.070328834822277, "grad_norm": 0.41329760884730493, "learning_rate": 1.2094079703758474e-05, "loss": 0.5212, "step": 716 }, { "epoch": 1.071849458277894, "grad_norm": 0.36806047789304813, "learning_rate": 1.2061122927525962e-05, "loss": 0.5148, "step": 717 }, { "epoch": 1.0733700817335108, "grad_norm": 0.4473641673792849, "learning_rate": 1.20281750643841e-05, "loss": 0.5217, "step": 718 }, { "epoch": 1.0748907051891274, "grad_norm": 0.35449349576414274, "learning_rate": 1.1995236311573668e-05, "loss": 0.5057, "step": 719 }, { "epoch": 1.0764113286447443, "grad_norm": 0.37802608508785335, "learning_rate": 1.1962306866280938e-05, "loss": 0.5127, "step": 720 }, { "epoch": 1.0779319521003612, "grad_norm": 0.38606300562371665, "learning_rate": 1.192938692563643e-05, "loss": 0.5225, "step": 721 }, { "epoch": 1.0794525755559778, "grad_norm": 0.3548827181233496, "learning_rate": 1.1896476686713781e-05, "loss": 0.5097, "step": 722 }, { "epoch": 1.0809731990115947, "grad_norm": 0.38704197291443254, "learning_rate": 1.1863576346528536e-05, "loss": 0.5147, "step": 723 }, { "epoch": 1.0824938224672116, "grad_norm": 0.3788637749591371, "learning_rate": 1.1830686102037011e-05, "loss": 0.526, "step": 724 }, { "epoch": 1.0840144459228283, "grad_norm": 0.36699155176089776, "learning_rate": 1.1797806150135052e-05, "loss": 0.5074, "step": 725 }, { "epoch": 1.0855350693784451, "grad_norm": 0.398207632535489, "learning_rate": 1.1764936687656908e-05, "loss": 0.5263, "step": 726 }, { "epoch": 1.087055692834062, "grad_norm": 0.34386679094163725, "learning_rate": 1.1732077911374022e-05, "loss": 0.5292, "step": 727 }, { "epoch": 1.0885763162896787, "grad_norm": 0.3898697838147865, "learning_rate": 1.1699230017993875e-05, "loss": 0.5157, "step": 728 }, { "epoch": 1.0900969397452955, "grad_norm": 0.35036102017636284, "learning_rate": 1.1666393204158793e-05, "loss": 0.5103, "step": 729 }, { "epoch": 1.0916175632009124, "grad_norm": 0.3766316653297432, "learning_rate": 1.1633567666444775e-05, "loss": 0.5355, "step": 730 }, { "epoch": 1.093138186656529, "grad_norm": 0.3395395584528322, "learning_rate": 1.1600753601360308e-05, "loss": 0.4922, "step": 731 }, { "epoch": 1.094658810112146, "grad_norm": 0.36396997728904185, "learning_rate": 1.1567951205345224e-05, "loss": 0.5219, "step": 732 }, { "epoch": 1.0961794335677628, "grad_norm": 0.3565103395663914, "learning_rate": 1.153516067476947e-05, "loss": 0.5089, "step": 733 }, { "epoch": 1.0977000570233795, "grad_norm": 0.3671854484970251, "learning_rate": 1.1502382205931974e-05, "loss": 0.5259, "step": 734 }, { "epoch": 1.0992206804789963, "grad_norm": 0.35699046104294285, "learning_rate": 1.1469615995059455e-05, "loss": 0.5199, "step": 735 }, { "epoch": 1.1007413039346132, "grad_norm": 0.35814749552055475, "learning_rate": 1.1436862238305256e-05, "loss": 0.4951, "step": 736 }, { "epoch": 1.1022619273902299, "grad_norm": 0.37766411014866846, "learning_rate": 1.1404121131748156e-05, "loss": 0.5133, "step": 737 }, { "epoch": 1.1037825508458468, "grad_norm": 0.38876585742040404, "learning_rate": 1.1371392871391207e-05, "loss": 0.5358, "step": 738 }, { "epoch": 1.1053031743014636, "grad_norm": 0.4211502488863111, "learning_rate": 1.1338677653160557e-05, "loss": 0.5083, "step": 739 }, { "epoch": 1.1068237977570803, "grad_norm": 0.3864747484032165, "learning_rate": 1.1305975672904284e-05, "loss": 0.5258, "step": 740 }, { "epoch": 1.1083444212126972, "grad_norm": 0.39083288027419083, "learning_rate": 1.1273287126391206e-05, "loss": 0.4843, "step": 741 }, { "epoch": 1.109865044668314, "grad_norm": 0.3910337272002525, "learning_rate": 1.1240612209309734e-05, "loss": 0.5479, "step": 742 }, { "epoch": 1.111385668123931, "grad_norm": 0.37980182130941675, "learning_rate": 1.1207951117266668e-05, "loss": 0.4885, "step": 743 }, { "epoch": 1.1129062915795476, "grad_norm": 0.3697135253054006, "learning_rate": 1.1175304045786073e-05, "loss": 0.523, "step": 744 }, { "epoch": 1.1144269150351644, "grad_norm": 0.3772904905798372, "learning_rate": 1.1142671190308056e-05, "loss": 0.5237, "step": 745 }, { "epoch": 1.1159475384907813, "grad_norm": 0.4056017049635124, "learning_rate": 1.1110052746187631e-05, "loss": 0.5137, "step": 746 }, { "epoch": 1.117468161946398, "grad_norm": 0.35112819485311203, "learning_rate": 1.1077448908693534e-05, "loss": 0.5038, "step": 747 }, { "epoch": 1.1189887854020149, "grad_norm": 0.38291495750095184, "learning_rate": 1.104485987300707e-05, "loss": 0.5084, "step": 748 }, { "epoch": 1.1205094088576317, "grad_norm": 0.3922969549725083, "learning_rate": 1.1012285834220924e-05, "loss": 0.5092, "step": 749 }, { "epoch": 1.1220300323132484, "grad_norm": 0.40549020760259696, "learning_rate": 1.0979726987338002e-05, "loss": 0.5204, "step": 750 }, { "epoch": 1.1235506557688653, "grad_norm": 0.40395038468147576, "learning_rate": 1.0947183527270267e-05, "loss": 0.5227, "step": 751 }, { "epoch": 1.1250712792244821, "grad_norm": 0.41236903215126236, "learning_rate": 1.0914655648837576e-05, "loss": 0.5319, "step": 752 }, { "epoch": 1.1265919026800988, "grad_norm": 0.3677643398357805, "learning_rate": 1.0882143546766496e-05, "loss": 0.5038, "step": 753 }, { "epoch": 1.1281125261357157, "grad_norm": 0.3828445757904712, "learning_rate": 1.084964741568916e-05, "loss": 0.5194, "step": 754 }, { "epoch": 1.1296331495913325, "grad_norm": 0.39209056252387725, "learning_rate": 1.0817167450142074e-05, "loss": 0.5013, "step": 755 }, { "epoch": 1.1311537730469492, "grad_norm": 0.3680446404406516, "learning_rate": 1.0784703844564998e-05, "loss": 0.5231, "step": 756 }, { "epoch": 1.132674396502566, "grad_norm": 0.3703842374744353, "learning_rate": 1.0752256793299732e-05, "loss": 0.5153, "step": 757 }, { "epoch": 1.134195019958183, "grad_norm": 0.37037413568566663, "learning_rate": 1.071982649058898e-05, "loss": 0.5124, "step": 758 }, { "epoch": 1.1357156434137996, "grad_norm": 0.34295639916690035, "learning_rate": 1.0687413130575176e-05, "loss": 0.5132, "step": 759 }, { "epoch": 1.1372362668694165, "grad_norm": 0.39357330789283446, "learning_rate": 1.0655016907299346e-05, "loss": 0.5288, "step": 760 }, { "epoch": 1.1387568903250334, "grad_norm": 0.3690059684924214, "learning_rate": 1.0622638014699909e-05, "loss": 0.4947, "step": 761 }, { "epoch": 1.14027751378065, "grad_norm": 0.3797018296519779, "learning_rate": 1.0590276646611539e-05, "loss": 0.5021, "step": 762 }, { "epoch": 1.1417981372362669, "grad_norm": 0.37238784723531165, "learning_rate": 1.0557932996764002e-05, "loss": 0.5111, "step": 763 }, { "epoch": 1.1433187606918838, "grad_norm": 0.34262491030815395, "learning_rate": 1.0525607258781e-05, "loss": 0.5095, "step": 764 }, { "epoch": 1.1448393841475004, "grad_norm": 0.3516626249819126, "learning_rate": 1.0493299626178997e-05, "loss": 0.5256, "step": 765 }, { "epoch": 1.1463600076031173, "grad_norm": 0.36336127776946636, "learning_rate": 1.0461010292366076e-05, "loss": 0.5163, "step": 766 }, { "epoch": 1.1478806310587342, "grad_norm": 0.3660245007970849, "learning_rate": 1.0428739450640768e-05, "loss": 0.5025, "step": 767 }, { "epoch": 1.1494012545143508, "grad_norm": 0.36947456530920786, "learning_rate": 1.039648729419092e-05, "loss": 0.5331, "step": 768 }, { "epoch": 1.1509218779699677, "grad_norm": 0.35068301401351004, "learning_rate": 1.0364254016092499e-05, "loss": 0.4979, "step": 769 }, { "epoch": 1.1524425014255846, "grad_norm": 0.3711143416445139, "learning_rate": 1.033203980930847e-05, "loss": 0.5349, "step": 770 }, { "epoch": 1.1539631248812012, "grad_norm": 0.3739635788047117, "learning_rate": 1.0299844866687616e-05, "loss": 0.4975, "step": 771 }, { "epoch": 1.155483748336818, "grad_norm": 0.33767131159596325, "learning_rate": 1.0267669380963419e-05, "loss": 0.4986, "step": 772 }, { "epoch": 1.157004371792435, "grad_norm": 0.4087303964655199, "learning_rate": 1.0235513544752866e-05, "loss": 0.4995, "step": 773 }, { "epoch": 1.1585249952480516, "grad_norm": 0.3669384792321335, "learning_rate": 1.0203377550555312e-05, "loss": 0.5033, "step": 774 }, { "epoch": 1.1600456187036685, "grad_norm": 0.3582614117799248, "learning_rate": 1.0171261590751334e-05, "loss": 0.5068, "step": 775 }, { "epoch": 1.1615662421592854, "grad_norm": 0.3548960213582146, "learning_rate": 1.0139165857601578e-05, "loss": 0.5219, "step": 776 }, { "epoch": 1.163086865614902, "grad_norm": 0.35769530792522997, "learning_rate": 1.0107090543245597e-05, "loss": 0.4991, "step": 777 }, { "epoch": 1.164607489070519, "grad_norm": 0.3762621019947577, "learning_rate": 1.0075035839700708e-05, "loss": 0.5318, "step": 778 }, { "epoch": 1.1661281125261358, "grad_norm": 0.3428064272464147, "learning_rate": 1.0043001938860842e-05, "loss": 0.5168, "step": 779 }, { "epoch": 1.1676487359817525, "grad_norm": 0.3740132674321784, "learning_rate": 1.001098903249541e-05, "loss": 0.512, "step": 780 }, { "epoch": 1.1691693594373693, "grad_norm": 0.37086418652390696, "learning_rate": 9.978997312248124e-06, "loss": 0.5046, "step": 781 }, { "epoch": 1.1706899828929862, "grad_norm": 0.34389709834427135, "learning_rate": 9.947026969635868e-06, "loss": 0.499, "step": 782 }, { "epoch": 1.1722106063486029, "grad_norm": 0.36925894599740416, "learning_rate": 9.915078196047548e-06, "loss": 0.5269, "step": 783 }, { "epoch": 1.1737312298042197, "grad_norm": 0.34734014615226194, "learning_rate": 9.883151182742967e-06, "loss": 0.5159, "step": 784 }, { "epoch": 1.1752518532598366, "grad_norm": 0.38858923268368156, "learning_rate": 9.851246120851641e-06, "loss": 0.5599, "step": 785 }, { "epoch": 1.1767724767154533, "grad_norm": 0.36352194480464795, "learning_rate": 9.819363201371682e-06, "loss": 0.5077, "step": 786 }, { "epoch": 1.1782931001710701, "grad_norm": 0.38508119384888934, "learning_rate": 9.78750261516864e-06, "loss": 0.5455, "step": 787 }, { "epoch": 1.179813723626687, "grad_norm": 0.35527440995181414, "learning_rate": 9.755664552974386e-06, "loss": 0.5128, "step": 788 }, { "epoch": 1.1813343470823037, "grad_norm": 0.3524825961134292, "learning_rate": 9.723849205385932e-06, "loss": 0.5062, "step": 789 }, { "epoch": 1.1828549705379205, "grad_norm": 0.4044926970167433, "learning_rate": 9.692056762864325e-06, "loss": 0.5412, "step": 790 }, { "epoch": 1.1843755939935374, "grad_norm": 0.35710970899973166, "learning_rate": 9.660287415733477e-06, "loss": 0.52, "step": 791 }, { "epoch": 1.185896217449154, "grad_norm": 0.37803033804258834, "learning_rate": 9.62854135417906e-06, "loss": 0.5052, "step": 792 }, { "epoch": 1.187416840904771, "grad_norm": 0.39768355225535407, "learning_rate": 9.596818768247333e-06, "loss": 0.5092, "step": 793 }, { "epoch": 1.1889374643603878, "grad_norm": 0.348711884248085, "learning_rate": 9.565119847844027e-06, "loss": 0.5267, "step": 794 }, { "epoch": 1.1904580878160045, "grad_norm": 0.3834819895765528, "learning_rate": 9.533444782733183e-06, "loss": 0.5037, "step": 795 }, { "epoch": 1.1919787112716214, "grad_norm": 0.3715822681968465, "learning_rate": 9.501793762536068e-06, "loss": 0.5308, "step": 796 }, { "epoch": 1.1934993347272382, "grad_norm": 0.3652977625843937, "learning_rate": 9.470166976729976e-06, "loss": 0.5372, "step": 797 }, { "epoch": 1.195019958182855, "grad_norm": 0.3364456352273384, "learning_rate": 9.43856461464713e-06, "loss": 0.4963, "step": 798 }, { "epoch": 1.1965405816384718, "grad_norm": 0.3472569441190444, "learning_rate": 9.40698686547354e-06, "loss": 0.5284, "step": 799 }, { "epoch": 1.1980612050940886, "grad_norm": 0.34440523348480445, "learning_rate": 9.375433918247888e-06, "loss": 0.4989, "step": 800 }, { "epoch": 1.1995818285497053, "grad_norm": 0.348206283208153, "learning_rate": 9.343905961860358e-06, "loss": 0.5226, "step": 801 }, { "epoch": 1.2011024520053222, "grad_norm": 0.3505525203415447, "learning_rate": 9.31240318505154e-06, "loss": 0.5047, "step": 802 }, { "epoch": 1.202623075460939, "grad_norm": 0.3515111190341964, "learning_rate": 9.280925776411278e-06, "loss": 0.5127, "step": 803 }, { "epoch": 1.2041436989165557, "grad_norm": 0.3620692993419406, "learning_rate": 9.249473924377573e-06, "loss": 0.5044, "step": 804 }, { "epoch": 1.2056643223721726, "grad_norm": 0.3659564097222614, "learning_rate": 9.218047817235408e-06, "loss": 0.5319, "step": 805 }, { "epoch": 1.2071849458277895, "grad_norm": 0.33548451693694653, "learning_rate": 9.186647643115659e-06, "loss": 0.4965, "step": 806 }, { "epoch": 1.2087055692834061, "grad_norm": 0.37688953100329553, "learning_rate": 9.15527358999394e-06, "loss": 0.5453, "step": 807 }, { "epoch": 1.210226192739023, "grad_norm": 0.3428883181534561, "learning_rate": 9.123925845689515e-06, "loss": 0.5392, "step": 808 }, { "epoch": 1.2117468161946399, "grad_norm": 0.3431087734831433, "learning_rate": 9.092604597864141e-06, "loss": 0.5118, "step": 809 }, { "epoch": 1.2132674396502565, "grad_norm": 0.35251065259476605, "learning_rate": 9.06131003402095e-06, "loss": 0.5054, "step": 810 }, { "epoch": 1.2147880631058734, "grad_norm": 0.3496035139299239, "learning_rate": 9.030042341503332e-06, "loss": 0.5084, "step": 811 }, { "epoch": 1.2163086865614903, "grad_norm": 0.33786656336389453, "learning_rate": 8.998801707493827e-06, "loss": 0.5199, "step": 812 }, { "epoch": 1.217829310017107, "grad_norm": 0.3668696340788812, "learning_rate": 8.967588319012977e-06, "loss": 0.4874, "step": 813 }, { "epoch": 1.2193499334727238, "grad_norm": 0.34867802829711625, "learning_rate": 8.936402362918224e-06, "loss": 0.4939, "step": 814 }, { "epoch": 1.2208705569283407, "grad_norm": 0.33881823564824526, "learning_rate": 8.90524402590278e-06, "loss": 0.5354, "step": 815 }, { "epoch": 1.2223911803839573, "grad_norm": 0.3395916132176579, "learning_rate": 8.874113494494537e-06, "loss": 0.5017, "step": 816 }, { "epoch": 1.2239118038395742, "grad_norm": 0.33632837523359915, "learning_rate": 8.843010955054909e-06, "loss": 0.5006, "step": 817 }, { "epoch": 1.225432427295191, "grad_norm": 0.33436403561258093, "learning_rate": 8.811936593777748e-06, "loss": 0.5052, "step": 818 }, { "epoch": 1.2269530507508077, "grad_norm": 0.3224896536759886, "learning_rate": 8.780890596688211e-06, "loss": 0.4981, "step": 819 }, { "epoch": 1.2284736742064246, "grad_norm": 0.34938066443230276, "learning_rate": 8.749873149641665e-06, "loss": 0.5203, "step": 820 }, { "epoch": 1.2299942976620415, "grad_norm": 0.345563582040282, "learning_rate": 8.718884438322557e-06, "loss": 0.5162, "step": 821 }, { "epoch": 1.2315149211176581, "grad_norm": 0.3428900758147224, "learning_rate": 8.687924648243311e-06, "loss": 0.4962, "step": 822 }, { "epoch": 1.233035544573275, "grad_norm": 0.34750322921749416, "learning_rate": 8.65699396474321e-06, "loss": 0.5186, "step": 823 }, { "epoch": 1.234556168028892, "grad_norm": 0.3589082109646499, "learning_rate": 8.626092572987297e-06, "loss": 0.5078, "step": 824 }, { "epoch": 1.2360767914845086, "grad_norm": 0.3506370697057621, "learning_rate": 8.595220657965268e-06, "loss": 0.5105, "step": 825 }, { "epoch": 1.2375974149401254, "grad_norm": 0.36011717525565096, "learning_rate": 8.564378404490342e-06, "loss": 0.5181, "step": 826 }, { "epoch": 1.2391180383957423, "grad_norm": 0.3402637505055028, "learning_rate": 8.533565997198175e-06, "loss": 0.5186, "step": 827 }, { "epoch": 1.2406386618513592, "grad_norm": 0.3475154384318057, "learning_rate": 8.50278362054577e-06, "loss": 0.4946, "step": 828 }, { "epoch": 1.2421592853069758, "grad_norm": 0.3590598605292272, "learning_rate": 8.472031458810325e-06, "loss": 0.5011, "step": 829 }, { "epoch": 1.2436799087625927, "grad_norm": 0.3483002221345324, "learning_rate": 8.441309696088174e-06, "loss": 0.5231, "step": 830 }, { "epoch": 1.2452005322182096, "grad_norm": 0.3475177777900519, "learning_rate": 8.410618516293662e-06, "loss": 0.5165, "step": 831 }, { "epoch": 1.2467211556738262, "grad_norm": 0.3633656751272662, "learning_rate": 8.37995810315806e-06, "loss": 0.5441, "step": 832 }, { "epoch": 1.2482417791294431, "grad_norm": 0.3447799135446017, "learning_rate": 8.349328640228443e-06, "loss": 0.4933, "step": 833 }, { "epoch": 1.24976240258506, "grad_norm": 0.34981751009747264, "learning_rate": 8.318730310866615e-06, "loss": 0.5008, "step": 834 }, { "epoch": 1.2512830260406766, "grad_norm": 0.361234032357548, "learning_rate": 8.288163298247994e-06, "loss": 0.4979, "step": 835 }, { "epoch": 1.2528036494962935, "grad_norm": 0.34834574638448507, "learning_rate": 8.257627785360528e-06, "loss": 0.5041, "step": 836 }, { "epoch": 1.2543242729519104, "grad_norm": 0.3466954777402415, "learning_rate": 8.227123955003586e-06, "loss": 0.519, "step": 837 }, { "epoch": 1.255844896407527, "grad_norm": 0.3555313535631245, "learning_rate": 8.196651989786872e-06, "loss": 0.513, "step": 838 }, { "epoch": 1.257365519863144, "grad_norm": 0.36059188542064946, "learning_rate": 8.166212072129334e-06, "loss": 0.4975, "step": 839 }, { "epoch": 1.2588861433187608, "grad_norm": 0.33893355712900364, "learning_rate": 8.135804384258074e-06, "loss": 0.4952, "step": 840 }, { "epoch": 1.2604067667743775, "grad_norm": 0.33361223601614015, "learning_rate": 8.10542910820724e-06, "loss": 0.5061, "step": 841 }, { "epoch": 1.2619273902299943, "grad_norm": 0.36477092577701276, "learning_rate": 8.07508642581696e-06, "loss": 0.5305, "step": 842 }, { "epoch": 1.2634480136856112, "grad_norm": 0.3563029319847921, "learning_rate": 8.044776518732229e-06, "loss": 0.4953, "step": 843 }, { "epoch": 1.2649686371412279, "grad_norm": 0.3480130680776454, "learning_rate": 8.014499568401845e-06, "loss": 0.5053, "step": 844 }, { "epoch": 1.2664892605968447, "grad_norm": 0.33347701875752966, "learning_rate": 7.98425575607731e-06, "loss": 0.4956, "step": 845 }, { "epoch": 1.2680098840524616, "grad_norm": 0.33599833733050055, "learning_rate": 7.95404526281174e-06, "loss": 0.5015, "step": 846 }, { "epoch": 1.2695305075080783, "grad_norm": 0.3384957381349461, "learning_rate": 7.923868269458795e-06, "loss": 0.5062, "step": 847 }, { "epoch": 1.2710511309636952, "grad_norm": 0.3252174947797439, "learning_rate": 7.893724956671587e-06, "loss": 0.4884, "step": 848 }, { "epoch": 1.272571754419312, "grad_norm": 0.34540716209835276, "learning_rate": 7.86361550490161e-06, "loss": 0.5045, "step": 849 }, { "epoch": 1.2740923778749287, "grad_norm": 0.3498563998796463, "learning_rate": 7.833540094397626e-06, "loss": 0.5118, "step": 850 }, { "epoch": 1.2756130013305456, "grad_norm": 0.3647153038013536, "learning_rate": 7.803498905204638e-06, "loss": 0.5064, "step": 851 }, { "epoch": 1.2771336247861624, "grad_norm": 0.3508552754714093, "learning_rate": 7.773492117162775e-06, "loss": 0.5035, "step": 852 }, { "epoch": 1.278654248241779, "grad_norm": 0.33239766666827464, "learning_rate": 7.743519909906217e-06, "loss": 0.4986, "step": 853 }, { "epoch": 1.280174871697396, "grad_norm": 0.4315070616641703, "learning_rate": 7.713582462862138e-06, "loss": 0.5331, "step": 854 }, { "epoch": 1.2816954951530128, "grad_norm": 0.36173927943521234, "learning_rate": 7.683679955249609e-06, "loss": 0.5064, "step": 855 }, { "epoch": 1.2832161186086295, "grad_norm": 0.3487948964838659, "learning_rate": 7.653812566078565e-06, "loss": 0.5061, "step": 856 }, { "epoch": 1.2847367420642464, "grad_norm": 0.3484239609890833, "learning_rate": 7.623980474148672e-06, "loss": 0.5023, "step": 857 }, { "epoch": 1.2862573655198632, "grad_norm": 0.36052769608217394, "learning_rate": 7.59418385804832e-06, "loss": 0.5267, "step": 858 }, { "epoch": 1.28777798897548, "grad_norm": 0.37188400515013914, "learning_rate": 7.564422896153491e-06, "loss": 0.5455, "step": 859 }, { "epoch": 1.2892986124310968, "grad_norm": 0.34411324307178126, "learning_rate": 7.534697766626769e-06, "loss": 0.5137, "step": 860 }, { "epoch": 1.2908192358867137, "grad_norm": 0.3364768988197693, "learning_rate": 7.50500864741619e-06, "loss": 0.5209, "step": 861 }, { "epoch": 1.2923398593423303, "grad_norm": 0.3670085290524646, "learning_rate": 7.475355716254242e-06, "loss": 0.5402, "step": 862 }, { "epoch": 1.2938604827979472, "grad_norm": 0.33251162694155534, "learning_rate": 7.445739150656753e-06, "loss": 0.4987, "step": 863 }, { "epoch": 1.295381106253564, "grad_norm": 0.35070274634255805, "learning_rate": 7.416159127921868e-06, "loss": 0.5071, "step": 864 }, { "epoch": 1.2969017297091807, "grad_norm": 0.3795469724864985, "learning_rate": 7.38661582512897e-06, "loss": 0.5399, "step": 865 }, { "epoch": 1.2984223531647976, "grad_norm": 0.3396586485316285, "learning_rate": 7.3571094191375995e-06, "loss": 0.5218, "step": 866 }, { "epoch": 1.2999429766204145, "grad_norm": 0.3360542873886429, "learning_rate": 7.327640086586438e-06, "loss": 0.5034, "step": 867 }, { "epoch": 1.3014636000760311, "grad_norm": 0.3295788458214383, "learning_rate": 7.298208003892223e-06, "loss": 0.5096, "step": 868 }, { "epoch": 1.302984223531648, "grad_norm": 0.35244946035448127, "learning_rate": 7.268813347248696e-06, "loss": 0.5085, "step": 869 }, { "epoch": 1.3045048469872649, "grad_norm": 0.3451919708597768, "learning_rate": 7.239456292625539e-06, "loss": 0.5017, "step": 870 }, { "epoch": 1.3060254704428815, "grad_norm": 0.3208050987232861, "learning_rate": 7.210137015767349e-06, "loss": 0.5024, "step": 871 }, { "epoch": 1.3075460938984984, "grad_norm": 0.35131393844950287, "learning_rate": 7.1808556921925585e-06, "loss": 0.529, "step": 872 }, { "epoch": 1.3090667173541153, "grad_norm": 0.35801843724419535, "learning_rate": 7.1516124971924e-06, "loss": 0.5297, "step": 873 }, { "epoch": 1.310587340809732, "grad_norm": 0.3401955783986041, "learning_rate": 7.122407605829839e-06, "loss": 0.5035, "step": 874 }, { "epoch": 1.3121079642653488, "grad_norm": 0.3371522526619987, "learning_rate": 7.093241192938552e-06, "loss": 0.4876, "step": 875 }, { "epoch": 1.3136285877209657, "grad_norm": 0.35482672035164964, "learning_rate": 7.064113433121862e-06, "loss": 0.5243, "step": 876 }, { "epoch": 1.3151492111765823, "grad_norm": 0.3535329850445862, "learning_rate": 7.03502450075169e-06, "loss": 0.5081, "step": 877 }, { "epoch": 1.3166698346321992, "grad_norm": 0.3423103799188622, "learning_rate": 7.00597456996753e-06, "loss": 0.5121, "step": 878 }, { "epoch": 1.318190458087816, "grad_norm": 0.35140677064268294, "learning_rate": 6.976963814675376e-06, "loss": 0.5171, "step": 879 }, { "epoch": 1.3197110815434328, "grad_norm": 0.352748609509177, "learning_rate": 6.947992408546731e-06, "loss": 0.535, "step": 880 }, { "epoch": 1.3212317049990496, "grad_norm": 0.3317879618506482, "learning_rate": 6.919060525017507e-06, "loss": 0.4942, "step": 881 }, { "epoch": 1.3227523284546665, "grad_norm": 0.3547220499965655, "learning_rate": 6.890168337287031e-06, "loss": 0.5091, "step": 882 }, { "epoch": 1.3242729519102832, "grad_norm": 0.32077453688103713, "learning_rate": 6.861316018316979e-06, "loss": 0.515, "step": 883 }, { "epoch": 1.3257935753659, "grad_norm": 0.3399152548809487, "learning_rate": 6.832503740830382e-06, "loss": 0.507, "step": 884 }, { "epoch": 1.327314198821517, "grad_norm": 0.3318950288383037, "learning_rate": 6.803731677310529e-06, "loss": 0.4966, "step": 885 }, { "epoch": 1.3288348222771336, "grad_norm": 0.3464596127850353, "learning_rate": 6.775000000000004e-06, "loss": 0.5115, "step": 886 }, { "epoch": 1.3303554457327504, "grad_norm": 0.34400466984436623, "learning_rate": 6.746308880899589e-06, "loss": 0.5082, "step": 887 }, { "epoch": 1.3318760691883673, "grad_norm": 0.3700416848449434, "learning_rate": 6.71765849176729e-06, "loss": 0.5302, "step": 888 }, { "epoch": 1.333396692643984, "grad_norm": 0.333378413090238, "learning_rate": 6.689049004117282e-06, "loss": 0.5021, "step": 889 }, { "epoch": 1.3349173160996008, "grad_norm": 0.3384873522425988, "learning_rate": 6.660480589218871e-06, "loss": 0.4805, "step": 890 }, { "epoch": 1.3364379395552177, "grad_norm": 0.34806352457541345, "learning_rate": 6.631953418095496e-06, "loss": 0.4912, "step": 891 }, { "epoch": 1.3379585630108344, "grad_norm": 0.34876873046591195, "learning_rate": 6.603467661523689e-06, "loss": 0.5082, "step": 892 }, { "epoch": 1.3394791864664513, "grad_norm": 0.31774849749377526, "learning_rate": 6.5750234900320585e-06, "loss": 0.4818, "step": 893 }, { "epoch": 1.3409998099220681, "grad_norm": 0.3324374020356446, "learning_rate": 6.546621073900255e-06, "loss": 0.4998, "step": 894 }, { "epoch": 1.3425204333776848, "grad_norm": 0.3353452020425694, "learning_rate": 6.518260583157976e-06, "loss": 0.4921, "step": 895 }, { "epoch": 1.3440410568333017, "grad_norm": 0.3493478007336028, "learning_rate": 6.489942187583926e-06, "loss": 0.5116, "step": 896 }, { "epoch": 1.3455616802889185, "grad_norm": 0.3284929233347785, "learning_rate": 6.461666056704817e-06, "loss": 0.5009, "step": 897 }, { "epoch": 1.3470823037445352, "grad_norm": 0.34202330071520587, "learning_rate": 6.433432359794333e-06, "loss": 0.5034, "step": 898 }, { "epoch": 1.348602927200152, "grad_norm": 0.33111021835858967, "learning_rate": 6.405241265872139e-06, "loss": 0.5044, "step": 899 }, { "epoch": 1.350123550655769, "grad_norm": 0.3354271189663323, "learning_rate": 6.37709294370286e-06, "loss": 0.4976, "step": 900 }, { "epoch": 1.3516441741113856, "grad_norm": 0.33711612171640787, "learning_rate": 6.3489875617950605e-06, "loss": 0.5018, "step": 901 }, { "epoch": 1.3531647975670025, "grad_norm": 0.33268125847154223, "learning_rate": 6.320925288400259e-06, "loss": 0.5061, "step": 902 }, { "epoch": 1.3546854210226194, "grad_norm": 0.3297595687007157, "learning_rate": 6.292906291511883e-06, "loss": 0.5008, "step": 903 }, { "epoch": 1.356206044478236, "grad_norm": 0.33711090972493357, "learning_rate": 6.2649307388643245e-06, "loss": 0.4899, "step": 904 }, { "epoch": 1.3577266679338529, "grad_norm": 0.35714761250864835, "learning_rate": 6.236998797931864e-06, "loss": 0.5336, "step": 905 }, { "epoch": 1.3592472913894698, "grad_norm": 0.33101330668166734, "learning_rate": 6.209110635927724e-06, "loss": 0.5188, "step": 906 }, { "epoch": 1.3607679148450864, "grad_norm": 0.33231897329769716, "learning_rate": 6.181266419803025e-06, "loss": 0.4992, "step": 907 }, { "epoch": 1.3622885383007033, "grad_norm": 0.3391798280349641, "learning_rate": 6.153466316245841e-06, "loss": 0.5145, "step": 908 }, { "epoch": 1.3638091617563202, "grad_norm": 0.3397695996188909, "learning_rate": 6.125710491680132e-06, "loss": 0.5148, "step": 909 }, { "epoch": 1.3653297852119368, "grad_norm": 0.33259300354324434, "learning_rate": 6.097999112264814e-06, "loss": 0.4983, "step": 910 }, { "epoch": 1.3668504086675537, "grad_norm": 0.34841156749227775, "learning_rate": 6.070332343892708e-06, "loss": 0.5197, "step": 911 }, { "epoch": 1.3683710321231706, "grad_norm": 0.3406141056782546, "learning_rate": 6.042710352189592e-06, "loss": 0.49, "step": 912 }, { "epoch": 1.3698916555787872, "grad_norm": 0.3479407700510544, "learning_rate": 6.01513330251319e-06, "loss": 0.5054, "step": 913 }, { "epoch": 1.371412279034404, "grad_norm": 0.34636977510682354, "learning_rate": 5.9876013599521654e-06, "loss": 0.5273, "step": 914 }, { "epoch": 1.372932902490021, "grad_norm": 0.32587606375019723, "learning_rate": 5.9601146893251685e-06, "loss": 0.5032, "step": 915 }, { "epoch": 1.3744535259456376, "grad_norm": 0.34209096131732425, "learning_rate": 5.932673455179826e-06, "loss": 0.5004, "step": 916 }, { "epoch": 1.3759741494012545, "grad_norm": 0.3462622805308902, "learning_rate": 5.9052778217917614e-06, "loss": 0.5121, "step": 917 }, { "epoch": 1.3774947728568714, "grad_norm": 0.3685686058332204, "learning_rate": 5.8779279531636046e-06, "loss": 0.5314, "step": 918 }, { "epoch": 1.379015396312488, "grad_norm": 0.3515509029896568, "learning_rate": 5.850624013024024e-06, "loss": 0.5215, "step": 919 }, { "epoch": 1.380536019768105, "grad_norm": 0.32046588423851496, "learning_rate": 5.82336616482674e-06, "loss": 0.4974, "step": 920 }, { "epoch": 1.3820566432237218, "grad_norm": 0.3808130154691177, "learning_rate": 5.796154571749547e-06, "loss": 0.525, "step": 921 }, { "epoch": 1.3835772666793384, "grad_norm": 0.3454746636565212, "learning_rate": 5.768989396693324e-06, "loss": 0.4923, "step": 922 }, { "epoch": 1.3850978901349553, "grad_norm": 0.3321076118691578, "learning_rate": 5.741870802281081e-06, "loss": 0.4797, "step": 923 }, { "epoch": 1.3866185135905722, "grad_norm": 0.3420626306833587, "learning_rate": 5.714798950856981e-06, "loss": 0.4923, "step": 924 }, { "epoch": 1.3881391370461889, "grad_norm": 0.3353059399098719, "learning_rate": 5.687774004485342e-06, "loss": 0.5073, "step": 925 }, { "epoch": 1.3896597605018057, "grad_norm": 0.3257554102906788, "learning_rate": 5.660796124949711e-06, "loss": 0.4942, "step": 926 }, { "epoch": 1.3911803839574226, "grad_norm": 0.3413851694093131, "learning_rate": 5.633865473751846e-06, "loss": 0.5345, "step": 927 }, { "epoch": 1.3927010074130393, "grad_norm": 0.33413542250569817, "learning_rate": 5.606982212110806e-06, "loss": 0.5205, "step": 928 }, { "epoch": 1.3942216308686561, "grad_norm": 0.3413782795261156, "learning_rate": 5.580146500961927e-06, "loss": 0.4914, "step": 929 }, { "epoch": 1.395742254324273, "grad_norm": 0.34562101413317636, "learning_rate": 5.553358500955903e-06, "loss": 0.5191, "step": 930 }, { "epoch": 1.3972628777798897, "grad_norm": 0.36463942955800077, "learning_rate": 5.526618372457788e-06, "loss": 0.5031, "step": 931 }, { "epoch": 1.3987835012355065, "grad_norm": 0.3490946304863948, "learning_rate": 5.49992627554609e-06, "loss": 0.5117, "step": 932 }, { "epoch": 1.4003041246911234, "grad_norm": 0.35279790273339234, "learning_rate": 5.473282370011742e-06, "loss": 0.5274, "step": 933 }, { "epoch": 1.40182474814674, "grad_norm": 0.349655925155731, "learning_rate": 5.446686815357206e-06, "loss": 0.4965, "step": 934 }, { "epoch": 1.403345371602357, "grad_norm": 0.334218280540011, "learning_rate": 5.420139770795477e-06, "loss": 0.5216, "step": 935 }, { "epoch": 1.4048659950579738, "grad_norm": 0.3630958366505651, "learning_rate": 5.393641395249157e-06, "loss": 0.5125, "step": 936 }, { "epoch": 1.4063866185135905, "grad_norm": 0.3665354616192356, "learning_rate": 5.367191847349496e-06, "loss": 0.5171, "step": 937 }, { "epoch": 1.4079072419692074, "grad_norm": 0.3388173032700996, "learning_rate": 5.340791285435425e-06, "loss": 0.5107, "step": 938 }, { "epoch": 1.4094278654248242, "grad_norm": 0.3408412945651779, "learning_rate": 5.314439867552638e-06, "loss": 0.5092, "step": 939 }, { "epoch": 1.4109484888804409, "grad_norm": 0.32178831689172704, "learning_rate": 5.288137751452625e-06, "loss": 0.5052, "step": 940 }, { "epoch": 1.4124691123360578, "grad_norm": 0.3404604279948066, "learning_rate": 5.261885094591735e-06, "loss": 0.4989, "step": 941 }, { "epoch": 1.4139897357916746, "grad_norm": 0.3528505749791748, "learning_rate": 5.235682054130224e-06, "loss": 0.4883, "step": 942 }, { "epoch": 1.4155103592472913, "grad_norm": 0.33074842020507783, "learning_rate": 5.209528786931329e-06, "loss": 0.5143, "step": 943 }, { "epoch": 1.4170309827029082, "grad_norm": 0.33129196603154915, "learning_rate": 5.183425449560322e-06, "loss": 0.5069, "step": 944 }, { "epoch": 1.418551606158525, "grad_norm": 0.3450555489678585, "learning_rate": 5.1573721982835766e-06, "loss": 0.5194, "step": 945 }, { "epoch": 1.4200722296141417, "grad_norm": 0.32000469036683216, "learning_rate": 5.131369189067611e-06, "loss": 0.4963, "step": 946 }, { "epoch": 1.4215928530697586, "grad_norm": 0.3579257724532904, "learning_rate": 5.105416577578189e-06, "loss": 0.5009, "step": 947 }, { "epoch": 1.4231134765253755, "grad_norm": 0.341889660403509, "learning_rate": 5.079514519179368e-06, "loss": 0.4955, "step": 948 }, { "epoch": 1.424634099980992, "grad_norm": 0.3375109020144137, "learning_rate": 5.053663168932563e-06, "loss": 0.4925, "step": 949 }, { "epoch": 1.426154723436609, "grad_norm": 0.3354140119830031, "learning_rate": 5.02786268159564e-06, "loss": 0.4828, "step": 950 }, { "epoch": 1.4276753468922259, "grad_norm": 0.3411962173320927, "learning_rate": 5.002113211621957e-06, "loss": 0.5059, "step": 951 }, { "epoch": 1.4291959703478425, "grad_norm": 0.3794888650970929, "learning_rate": 4.97641491315949e-06, "loss": 0.5182, "step": 952 }, { "epoch": 1.4307165938034594, "grad_norm": 0.3438813382757282, "learning_rate": 4.950767940049846e-06, "loss": 0.5231, "step": 953 }, { "epoch": 1.4322372172590763, "grad_norm": 0.33702340568088535, "learning_rate": 4.925172445827401e-06, "loss": 0.4898, "step": 954 }, { "epoch": 1.433757840714693, "grad_norm": 0.33891170335451204, "learning_rate": 4.8996285837183344e-06, "loss": 0.4999, "step": 955 }, { "epoch": 1.4352784641703098, "grad_norm": 0.3613620365932021, "learning_rate": 4.874136506639747e-06, "loss": 0.5109, "step": 956 }, { "epoch": 1.4367990876259267, "grad_norm": 0.34896992869532173, "learning_rate": 4.848696367198726e-06, "loss": 0.503, "step": 957 }, { "epoch": 1.4383197110815433, "grad_norm": 0.3319351812805159, "learning_rate": 4.823308317691442e-06, "loss": 0.5189, "step": 958 }, { "epoch": 1.4398403345371602, "grad_norm": 0.3319386837420594, "learning_rate": 4.7979725101022175e-06, "loss": 0.4813, "step": 959 }, { "epoch": 1.441360957992777, "grad_norm": 0.3397974585040773, "learning_rate": 4.772689096102645e-06, "loss": 0.5045, "step": 960 }, { "epoch": 1.4428815814483937, "grad_norm": 0.31839220001298785, "learning_rate": 4.747458227050663e-06, "loss": 0.4815, "step": 961 }, { "epoch": 1.4444022049040106, "grad_norm": 0.32870599013036034, "learning_rate": 4.7222800539896445e-06, "loss": 0.5142, "step": 962 }, { "epoch": 1.4459228283596275, "grad_norm": 0.3342769560881234, "learning_rate": 4.697154727647507e-06, "loss": 0.4943, "step": 963 }, { "epoch": 1.4474434518152441, "grad_norm": 0.34488780755334697, "learning_rate": 4.672082398435805e-06, "loss": 0.5137, "step": 964 }, { "epoch": 1.448964075270861, "grad_norm": 0.3153153534150591, "learning_rate": 4.6470632164488295e-06, "loss": 0.4837, "step": 965 }, { "epoch": 1.450484698726478, "grad_norm": 0.32145993722141425, "learning_rate": 4.622097331462696e-06, "loss": 0.4962, "step": 966 }, { "epoch": 1.4520053221820945, "grad_norm": 0.33298486778180947, "learning_rate": 4.597184892934472e-06, "loss": 0.5103, "step": 967 }, { "epoch": 1.4535259456377114, "grad_norm": 0.3450267637519744, "learning_rate": 4.572326050001267e-06, "loss": 0.4909, "step": 968 }, { "epoch": 1.4550465690933283, "grad_norm": 0.3318326795828598, "learning_rate": 4.547520951479347e-06, "loss": 0.5057, "step": 969 }, { "epoch": 1.456567192548945, "grad_norm": 0.33679166346861333, "learning_rate": 4.522769745863228e-06, "loss": 0.5057, "step": 970 }, { "epoch": 1.4580878160045618, "grad_norm": 0.3395760337602621, "learning_rate": 4.498072581324811e-06, "loss": 0.5358, "step": 971 }, { "epoch": 1.4596084394601787, "grad_norm": 0.3350852965124303, "learning_rate": 4.473429605712482e-06, "loss": 0.516, "step": 972 }, { "epoch": 1.4611290629157954, "grad_norm": 0.3402104641226832, "learning_rate": 4.448840966550214e-06, "loss": 0.5042, "step": 973 }, { "epoch": 1.4626496863714122, "grad_norm": 0.35213392448418906, "learning_rate": 4.424306811036717e-06, "loss": 0.5048, "step": 974 }, { "epoch": 1.4641703098270291, "grad_norm": 0.30760979362082447, "learning_rate": 4.399827286044515e-06, "loss": 0.4976, "step": 975 }, { "epoch": 1.4656909332826458, "grad_norm": 0.3713309741272073, "learning_rate": 4.375402538119116e-06, "loss": 0.4849, "step": 976 }, { "epoch": 1.4672115567382626, "grad_norm": 0.32636579757729905, "learning_rate": 4.351032713478084e-06, "loss": 0.5011, "step": 977 }, { "epoch": 1.4687321801938795, "grad_norm": 0.34427812460067464, "learning_rate": 4.326717958010203e-06, "loss": 0.4985, "step": 978 }, { "epoch": 1.4702528036494962, "grad_norm": 0.3431253440023036, "learning_rate": 4.302458417274573e-06, "loss": 0.504, "step": 979 }, { "epoch": 1.471773427105113, "grad_norm": 0.3219320912973509, "learning_rate": 4.278254236499767e-06, "loss": 0.4872, "step": 980 }, { "epoch": 1.47329405056073, "grad_norm": 0.3259111453940081, "learning_rate": 4.2541055605829445e-06, "loss": 0.4915, "step": 981 }, { "epoch": 1.4748146740163466, "grad_norm": 0.34356473045787483, "learning_rate": 4.2300125340889885e-06, "loss": 0.5041, "step": 982 }, { "epoch": 1.4763352974719635, "grad_norm": 0.3302732477989796, "learning_rate": 4.2059753012496334e-06, "loss": 0.4975, "step": 983 }, { "epoch": 1.4778559209275803, "grad_norm": 0.3139509766944064, "learning_rate": 4.181994005962612e-06, "loss": 0.4987, "step": 984 }, { "epoch": 1.479376544383197, "grad_norm": 0.33007516038018775, "learning_rate": 4.158068791790798e-06, "loss": 0.5094, "step": 985 }, { "epoch": 1.4808971678388139, "grad_norm": 0.3314396852932346, "learning_rate": 4.134199801961319e-06, "loss": 0.5214, "step": 986 }, { "epoch": 1.4824177912944307, "grad_norm": 0.3265821672215369, "learning_rate": 4.110387179364735e-06, "loss": 0.4829, "step": 987 }, { "epoch": 1.4839384147500474, "grad_norm": 0.3332694006250523, "learning_rate": 4.0866310665541576e-06, "loss": 0.5052, "step": 988 }, { "epoch": 1.4854590382056643, "grad_norm": 0.3490720940072708, "learning_rate": 4.062931605744416e-06, "loss": 0.5092, "step": 989 }, { "epoch": 1.4869796616612811, "grad_norm": 0.33398988538744084, "learning_rate": 4.0392889388111746e-06, "loss": 0.5188, "step": 990 }, { "epoch": 1.4885002851168978, "grad_norm": 0.3270648319812971, "learning_rate": 4.015703207290125e-06, "loss": 0.5159, "step": 991 }, { "epoch": 1.4900209085725147, "grad_norm": 0.32741030567450113, "learning_rate": 3.992174552376112e-06, "loss": 0.5025, "step": 992 }, { "epoch": 1.4915415320281316, "grad_norm": 0.32463136930033865, "learning_rate": 3.968703114922282e-06, "loss": 0.4912, "step": 993 }, { "epoch": 1.4930621554837482, "grad_norm": 0.3493187625119753, "learning_rate": 3.945289035439266e-06, "loss": 0.5022, "step": 994 }, { "epoch": 1.494582778939365, "grad_norm": 0.3393915308815615, "learning_rate": 3.921932454094324e-06, "loss": 0.5121, "step": 995 }, { "epoch": 1.496103402394982, "grad_norm": 0.3361281345069889, "learning_rate": 3.8986335107104904e-06, "loss": 0.5113, "step": 996 }, { "epoch": 1.4976240258505986, "grad_norm": 0.33203560338943144, "learning_rate": 3.875392344765772e-06, "loss": 0.4754, "step": 997 }, { "epoch": 1.4991446493062155, "grad_norm": 0.324328611770242, "learning_rate": 3.852209095392288e-06, "loss": 0.4993, "step": 998 }, { "epoch": 1.5006652727618324, "grad_norm": 0.3558742459808321, "learning_rate": 3.829083901375433e-06, "loss": 0.5047, "step": 999 }, { "epoch": 1.502185896217449, "grad_norm": 0.3338340547562902, "learning_rate": 3.8060169011530668e-06, "loss": 0.4759, "step": 1000 }, { "epoch": 1.5037065196730661, "grad_norm": 0.35120584802546506, "learning_rate": 3.783008232814671e-06, "loss": 0.5294, "step": 1001 }, { "epoch": 1.5052271431286828, "grad_norm": 0.3353936427085398, "learning_rate": 3.760058034100533e-06, "loss": 0.5284, "step": 1002 }, { "epoch": 1.5067477665842994, "grad_norm": 0.33688668654855014, "learning_rate": 3.7371664424008973e-06, "loss": 0.4959, "step": 1003 }, { "epoch": 1.5082683900399165, "grad_norm": 0.32266166581151934, "learning_rate": 3.714333594755177e-06, "loss": 0.4752, "step": 1004 }, { "epoch": 1.5097890134955332, "grad_norm": 0.33412231708735246, "learning_rate": 3.691559627851118e-06, "loss": 0.5179, "step": 1005 }, { "epoch": 1.5113096369511498, "grad_norm": 0.3243441675650943, "learning_rate": 3.6688446780239596e-06, "loss": 0.4943, "step": 1006 }, { "epoch": 1.512830260406767, "grad_norm": 0.327154574269167, "learning_rate": 3.646188881255658e-06, "loss": 0.492, "step": 1007 }, { "epoch": 1.5143508838623836, "grad_norm": 0.3192294634078481, "learning_rate": 3.623592373174045e-06, "loss": 0.4771, "step": 1008 }, { "epoch": 1.5158715073180002, "grad_norm": 0.3372732645788767, "learning_rate": 3.601055289052031e-06, "loss": 0.4949, "step": 1009 }, { "epoch": 1.5173921307736173, "grad_norm": 0.31829723772061774, "learning_rate": 3.578577763806769e-06, "loss": 0.49, "step": 1010 }, { "epoch": 1.518912754229234, "grad_norm": 0.3374508150505602, "learning_rate": 3.556159931998888e-06, "loss": 0.5138, "step": 1011 }, { "epoch": 1.5204333776848507, "grad_norm": 0.3368177667255345, "learning_rate": 3.5338019278316426e-06, "loss": 0.5332, "step": 1012 }, { "epoch": 1.5219540011404677, "grad_norm": 0.35595499914728906, "learning_rate": 3.5115038851501605e-06, "loss": 0.5093, "step": 1013 }, { "epoch": 1.5234746245960844, "grad_norm": 0.3549641645181648, "learning_rate": 3.4892659374405844e-06, "loss": 0.5251, "step": 1014 }, { "epoch": 1.524995248051701, "grad_norm": 0.32092817344554087, "learning_rate": 3.467088217829321e-06, "loss": 0.5086, "step": 1015 }, { "epoch": 1.5265158715073182, "grad_norm": 0.3163937191442778, "learning_rate": 3.4449708590822104e-06, "loss": 0.4997, "step": 1016 }, { "epoch": 1.5280364949629348, "grad_norm": 0.3204985454805181, "learning_rate": 3.422913993603756e-06, "loss": 0.523, "step": 1017 }, { "epoch": 1.5295571184185515, "grad_norm": 0.3282138238031523, "learning_rate": 3.4009177534363184e-06, "loss": 0.5096, "step": 1018 }, { "epoch": 1.5310777418741686, "grad_norm": 0.3243188648838239, "learning_rate": 3.378982270259323e-06, "loss": 0.4974, "step": 1019 }, { "epoch": 1.5325983653297852, "grad_norm": 0.32762061133060594, "learning_rate": 3.3571076753884802e-06, "loss": 0.5174, "step": 1020 }, { "epoch": 1.5341189887854019, "grad_norm": 0.3315702913947939, "learning_rate": 3.335294099774996e-06, "loss": 0.5317, "step": 1021 }, { "epoch": 1.535639612241019, "grad_norm": 0.3146514780388357, "learning_rate": 3.313541674004791e-06, "loss": 0.4823, "step": 1022 }, { "epoch": 1.5371602356966356, "grad_norm": 0.31324712226441537, "learning_rate": 3.291850528297705e-06, "loss": 0.4936, "step": 1023 }, { "epoch": 1.5386808591522523, "grad_norm": 0.3220338054628576, "learning_rate": 3.270220792506735e-06, "loss": 0.4897, "step": 1024 }, { "epoch": 1.5402014826078694, "grad_norm": 0.33327886740248275, "learning_rate": 3.2486525961172487e-06, "loss": 0.5118, "step": 1025 }, { "epoch": 1.541722106063486, "grad_norm": 0.31839568655177714, "learning_rate": 3.2271460682462175e-06, "loss": 0.4913, "step": 1026 }, { "epoch": 1.5432427295191027, "grad_norm": 0.3275702952073424, "learning_rate": 3.2057013376414218e-06, "loss": 0.5022, "step": 1027 }, { "epoch": 1.5447633529747198, "grad_norm": 0.3265904791104775, "learning_rate": 3.1843185326807096e-06, "loss": 0.5013, "step": 1028 }, { "epoch": 1.5462839764303364, "grad_norm": 0.34562281052781296, "learning_rate": 3.162997781371215e-06, "loss": 0.5245, "step": 1029 }, { "epoch": 1.547804599885953, "grad_norm": 0.3304005470169874, "learning_rate": 3.141739211348575e-06, "loss": 0.5034, "step": 1030 }, { "epoch": 1.5493252233415702, "grad_norm": 0.32880228905554193, "learning_rate": 3.1205429498761974e-06, "loss": 0.5015, "step": 1031 }, { "epoch": 1.5508458467971868, "grad_norm": 0.32336045302160143, "learning_rate": 3.0994091238444775e-06, "loss": 0.511, "step": 1032 }, { "epoch": 1.5523664702528035, "grad_norm": 0.3311162660651907, "learning_rate": 3.0783378597700453e-06, "loss": 0.5114, "step": 1033 }, { "epoch": 1.5538870937084206, "grad_norm": 0.3375600053400384, "learning_rate": 3.0573292837949997e-06, "loss": 0.4995, "step": 1034 }, { "epoch": 1.5554077171640373, "grad_norm": 0.3184801308022672, "learning_rate": 3.036383521686171e-06, "loss": 0.491, "step": 1035 }, { "epoch": 1.556928340619654, "grad_norm": 0.3453112317336798, "learning_rate": 3.015500698834343e-06, "loss": 0.5348, "step": 1036 }, { "epoch": 1.558448964075271, "grad_norm": 0.3178400015966866, "learning_rate": 2.9946809402535396e-06, "loss": 0.5184, "step": 1037 }, { "epoch": 1.5599695875308877, "grad_norm": 0.32719134398419775, "learning_rate": 2.9739243705802274e-06, "loss": 0.515, "step": 1038 }, { "epoch": 1.5614902109865043, "grad_norm": 0.3205800396648508, "learning_rate": 2.953231114072617e-06, "loss": 0.4964, "step": 1039 }, { "epoch": 1.5630108344421214, "grad_norm": 0.316255767582057, "learning_rate": 2.9326012946098793e-06, "loss": 0.4772, "step": 1040 }, { "epoch": 1.564531457897738, "grad_norm": 0.3175652603893164, "learning_rate": 2.9120350356914355e-06, "loss": 0.5086, "step": 1041 }, { "epoch": 1.5660520813533547, "grad_norm": 0.32582238160345084, "learning_rate": 2.891532460436206e-06, "loss": 0.513, "step": 1042 }, { "epoch": 1.5675727048089718, "grad_norm": 0.3296575006626612, "learning_rate": 2.8710936915818587e-06, "loss": 0.5023, "step": 1043 }, { "epoch": 1.5690933282645885, "grad_norm": 0.31611337448881216, "learning_rate": 2.8507188514840983e-06, "loss": 0.5156, "step": 1044 }, { "epoch": 1.5706139517202051, "grad_norm": 0.31450333160108834, "learning_rate": 2.8304080621159222e-06, "loss": 0.5039, "step": 1045 }, { "epoch": 1.5721345751758222, "grad_norm": 0.3213166122638692, "learning_rate": 2.8101614450668923e-06, "loss": 0.4891, "step": 1046 }, { "epoch": 1.5736551986314389, "grad_norm": 0.3410745502394496, "learning_rate": 2.7899791215423954e-06, "loss": 0.5091, "step": 1047 }, { "epoch": 1.5751758220870555, "grad_norm": 0.3888744291624106, "learning_rate": 2.76986121236294e-06, "loss": 0.5075, "step": 1048 }, { "epoch": 1.5766964455426726, "grad_norm": 0.32027232788552623, "learning_rate": 2.7498078379634194e-06, "loss": 0.4863, "step": 1049 }, { "epoch": 1.5782170689982893, "grad_norm": 0.3219460654212709, "learning_rate": 2.7298191183923903e-06, "loss": 0.4919, "step": 1050 }, { "epoch": 1.579737692453906, "grad_norm": 0.3189327947333556, "learning_rate": 2.709895173311352e-06, "loss": 0.507, "step": 1051 }, { "epoch": 1.581258315909523, "grad_norm": 0.34438583434356085, "learning_rate": 2.690036121994041e-06, "loss": 0.4899, "step": 1052 }, { "epoch": 1.5827789393651397, "grad_norm": 0.3398976699129184, "learning_rate": 2.6702420833257135e-06, "loss": 0.5225, "step": 1053 }, { "epoch": 1.5842995628207563, "grad_norm": 0.3102911393257611, "learning_rate": 2.6505131758024167e-06, "loss": 0.5059, "step": 1054 }, { "epoch": 1.5858201862763734, "grad_norm": 0.324270478165599, "learning_rate": 2.630849517530311e-06, "loss": 0.4981, "step": 1055 }, { "epoch": 1.58734080973199, "grad_norm": 0.3546913242376741, "learning_rate": 2.6112512262249243e-06, "loss": 0.5361, "step": 1056 }, { "epoch": 1.5888614331876068, "grad_norm": 0.34869377738298407, "learning_rate": 2.591718419210495e-06, "loss": 0.5103, "step": 1057 }, { "epoch": 1.5903820566432239, "grad_norm": 0.3123532728864231, "learning_rate": 2.572251213419218e-06, "loss": 0.4932, "step": 1058 }, { "epoch": 1.5919026800988405, "grad_norm": 0.32330942665052126, "learning_rate": 2.5528497253905883e-06, "loss": 0.5037, "step": 1059 }, { "epoch": 1.5934233035544574, "grad_norm": 0.31181836853952816, "learning_rate": 2.5335140712706637e-06, "loss": 0.4866, "step": 1060 }, { "epoch": 1.5949439270100743, "grad_norm": 0.3030338685118879, "learning_rate": 2.5142443668114184e-06, "loss": 0.4618, "step": 1061 }, { "epoch": 1.596464550465691, "grad_norm": 0.3270760691359473, "learning_rate": 2.495040727369997e-06, "loss": 0.5345, "step": 1062 }, { "epoch": 1.5979851739213078, "grad_norm": 0.31027653370341124, "learning_rate": 2.4759032679080646e-06, "loss": 0.4881, "step": 1063 }, { "epoch": 1.5995057973769247, "grad_norm": 0.3062497695212509, "learning_rate": 2.4568321029910926e-06, "loss": 0.4787, "step": 1064 }, { "epoch": 1.6010264208325413, "grad_norm": 0.32621381275453964, "learning_rate": 2.4378273467876895e-06, "loss": 0.525, "step": 1065 }, { "epoch": 1.6025470442881582, "grad_norm": 0.3318069849664119, "learning_rate": 2.4188891130689163e-06, "loss": 0.4909, "step": 1066 }, { "epoch": 1.604067667743775, "grad_norm": 0.32860192649298986, "learning_rate": 2.4000175152075832e-06, "loss": 0.5047, "step": 1067 }, { "epoch": 1.6055882911993917, "grad_norm": 0.32129364609495753, "learning_rate": 2.3812126661776048e-06, "loss": 0.49, "step": 1068 }, { "epoch": 1.6071089146550086, "grad_norm": 0.34479880933344564, "learning_rate": 2.3624746785533015e-06, "loss": 0.5202, "step": 1069 }, { "epoch": 1.6086295381106255, "grad_norm": 0.32050670560042915, "learning_rate": 2.3438036645087323e-06, "loss": 0.5045, "step": 1070 }, { "epoch": 1.6101501615662421, "grad_norm": 0.3236770692201634, "learning_rate": 2.325199735817016e-06, "loss": 0.5007, "step": 1071 }, { "epoch": 1.611670785021859, "grad_norm": 0.3182094882577364, "learning_rate": 2.306663003849674e-06, "loss": 0.4972, "step": 1072 }, { "epoch": 1.6131914084774759, "grad_norm": 0.31720365603712886, "learning_rate": 2.2881935795759588e-06, "loss": 0.4716, "step": 1073 }, { "epoch": 1.6147120319330925, "grad_norm": 0.32974314670568317, "learning_rate": 2.2697915735621873e-06, "loss": 0.484, "step": 1074 }, { "epoch": 1.6162326553887094, "grad_norm": 0.30577586884242175, "learning_rate": 2.2514570959710736e-06, "loss": 0.4735, "step": 1075 }, { "epoch": 1.6177532788443263, "grad_norm": 0.3197001472064691, "learning_rate": 2.2331902565610855e-06, "loss": 0.5218, "step": 1076 }, { "epoch": 1.619273902299943, "grad_norm": 0.3166847446707726, "learning_rate": 2.214991164685776e-06, "loss": 0.4903, "step": 1077 }, { "epoch": 1.6207945257555598, "grad_norm": 0.3195718257381146, "learning_rate": 2.1968599292931262e-06, "loss": 0.4971, "step": 1078 }, { "epoch": 1.6223151492111767, "grad_norm": 0.31426671771687575, "learning_rate": 2.178796658924904e-06, "loss": 0.5064, "step": 1079 }, { "epoch": 1.6238357726667934, "grad_norm": 0.3212716102723848, "learning_rate": 2.160801461715996e-06, "loss": 0.5079, "step": 1080 }, { "epoch": 1.6253563961224102, "grad_norm": 0.3191289490920015, "learning_rate": 2.1428744453937935e-06, "loss": 0.4761, "step": 1081 }, { "epoch": 1.626877019578027, "grad_norm": 0.3159849608628463, "learning_rate": 2.125015717277504e-06, "loss": 0.5033, "step": 1082 }, { "epoch": 1.6283976430336438, "grad_norm": 0.31696066878201823, "learning_rate": 2.1072253842775457e-06, "loss": 0.4788, "step": 1083 }, { "epoch": 1.6299182664892606, "grad_norm": 0.3258225759542187, "learning_rate": 2.089503552894881e-06, "loss": 0.5099, "step": 1084 }, { "epoch": 1.6314388899448775, "grad_norm": 0.3097721381921143, "learning_rate": 2.0718503292204057e-06, "loss": 0.4837, "step": 1085 }, { "epoch": 1.6329595134004942, "grad_norm": 0.32471560408203043, "learning_rate": 2.054265818934283e-06, "loss": 0.523, "step": 1086 }, { "epoch": 1.634480136856111, "grad_norm": 0.30086700392629573, "learning_rate": 2.036750127305341e-06, "loss": 0.4751, "step": 1087 }, { "epoch": 1.636000760311728, "grad_norm": 0.306302856546338, "learning_rate": 2.0193033591904125e-06, "loss": 0.4734, "step": 1088 }, { "epoch": 1.6375213837673446, "grad_norm": 0.30581159890991627, "learning_rate": 2.001925619033737e-06, "loss": 0.4881, "step": 1089 }, { "epoch": 1.6390420072229615, "grad_norm": 0.3148442376719998, "learning_rate": 1.9846170108663187e-06, "loss": 0.4862, "step": 1090 }, { "epoch": 1.6405626306785783, "grad_norm": 0.3187412303526785, "learning_rate": 1.9673776383052978e-06, "loss": 0.5088, "step": 1091 }, { "epoch": 1.642083254134195, "grad_norm": 0.3237394151376111, "learning_rate": 1.9502076045533484e-06, "loss": 0.5127, "step": 1092 }, { "epoch": 1.6436038775898119, "grad_norm": 0.30397789760314514, "learning_rate": 1.9331070123980477e-06, "loss": 0.4906, "step": 1093 }, { "epoch": 1.6451245010454287, "grad_norm": 0.31316588193164063, "learning_rate": 1.916075964211268e-06, "loss": 0.496, "step": 1094 }, { "epoch": 1.6466451245010454, "grad_norm": 0.33024840735397576, "learning_rate": 1.8991145619485479e-06, "loss": 0.5203, "step": 1095 }, { "epoch": 1.6481657479566623, "grad_norm": 0.3409025779210624, "learning_rate": 1.8822229071485085e-06, "loss": 0.5002, "step": 1096 }, { "epoch": 1.6496863714122791, "grad_norm": 0.3287972930467973, "learning_rate": 1.8654011009322228e-06, "loss": 0.4995, "step": 1097 }, { "epoch": 1.6512069948678958, "grad_norm": 0.32623077457196087, "learning_rate": 1.8486492440026244e-06, "loss": 0.487, "step": 1098 }, { "epoch": 1.6527276183235127, "grad_norm": 0.3106687896956913, "learning_rate": 1.83196743664389e-06, "loss": 0.4946, "step": 1099 }, { "epoch": 1.6542482417791295, "grad_norm": 0.32812705879714626, "learning_rate": 1.8153557787208556e-06, "loss": 0.4903, "step": 1100 }, { "epoch": 1.6557688652347462, "grad_norm": 0.31890170792842726, "learning_rate": 1.7988143696784124e-06, "loss": 0.4821, "step": 1101 }, { "epoch": 1.657289488690363, "grad_norm": 0.31284347487731023, "learning_rate": 1.7823433085409e-06, "loss": 0.4931, "step": 1102 }, { "epoch": 1.65881011214598, "grad_norm": 0.30982765014249525, "learning_rate": 1.7659426939115348e-06, "loss": 0.4683, "step": 1103 }, { "epoch": 1.6603307356015966, "grad_norm": 0.3164620121768728, "learning_rate": 1.7496126239717968e-06, "loss": 0.5007, "step": 1104 }, { "epoch": 1.6618513590572135, "grad_norm": 0.3479489881630459, "learning_rate": 1.7333531964808702e-06, "loss": 0.5055, "step": 1105 }, { "epoch": 1.6633719825128304, "grad_norm": 0.3208880246840334, "learning_rate": 1.7171645087750255e-06, "loss": 0.4876, "step": 1106 }, { "epoch": 1.664892605968447, "grad_norm": 0.32500628409148635, "learning_rate": 1.7010466577670633e-06, "loss": 0.5185, "step": 1107 }, { "epoch": 1.666413229424064, "grad_norm": 0.3151870780999753, "learning_rate": 1.684999739945713e-06, "loss": 0.5224, "step": 1108 }, { "epoch": 1.6679338528796808, "grad_norm": 0.30364068565385566, "learning_rate": 1.669023851375082e-06, "loss": 0.4843, "step": 1109 }, { "epoch": 1.6694544763352974, "grad_norm": 0.32280756678101397, "learning_rate": 1.6531190876940478e-06, "loss": 0.5129, "step": 1110 }, { "epoch": 1.6709750997909143, "grad_norm": 0.3058304028983153, "learning_rate": 1.6372855441157166e-06, "loss": 0.4708, "step": 1111 }, { "epoch": 1.6724957232465312, "grad_norm": 0.34052506206539823, "learning_rate": 1.6215233154268233e-06, "loss": 0.5272, "step": 1112 }, { "epoch": 1.6740163467021478, "grad_norm": 0.32203401759522543, "learning_rate": 1.6058324959871942e-06, "loss": 0.4975, "step": 1113 }, { "epoch": 1.6755369701577647, "grad_norm": 0.3192131968898841, "learning_rate": 1.5902131797291616e-06, "loss": 0.4892, "step": 1114 }, { "epoch": 1.6770575936133816, "grad_norm": 0.30564977758041467, "learning_rate": 1.5746654601570045e-06, "loss": 0.4891, "step": 1115 }, { "epoch": 1.6785782170689982, "grad_norm": 0.3174124356131291, "learning_rate": 1.559189430346397e-06, "loss": 0.4948, "step": 1116 }, { "epoch": 1.6800988405246151, "grad_norm": 0.3094722335828649, "learning_rate": 1.5437851829438422e-06, "loss": 0.488, "step": 1117 }, { "epoch": 1.681619463980232, "grad_norm": 0.3131260119981047, "learning_rate": 1.5284528101661273e-06, "loss": 0.4664, "step": 1118 }, { "epoch": 1.6831400874358486, "grad_norm": 0.31529735958006433, "learning_rate": 1.5131924037997535e-06, "loss": 0.4951, "step": 1119 }, { "epoch": 1.6846607108914655, "grad_norm": 0.30173098838134094, "learning_rate": 1.4980040552004121e-06, "loss": 0.4903, "step": 1120 }, { "epoch": 1.6861813343470824, "grad_norm": 0.30988057448857725, "learning_rate": 1.4828878552924145e-06, "loss": 0.49, "step": 1121 }, { "epoch": 1.687701957802699, "grad_norm": 0.3077026541913675, "learning_rate": 1.4678438945681649e-06, "loss": 0.4849, "step": 1122 }, { "epoch": 1.689222581258316, "grad_norm": 0.31664882915707754, "learning_rate": 1.4528722630876041e-06, "loss": 0.4972, "step": 1123 }, { "epoch": 1.6907432047139328, "grad_norm": 0.31667781766589775, "learning_rate": 1.4379730504776807e-06, "loss": 0.4907, "step": 1124 }, { "epoch": 1.6922638281695495, "grad_norm": 0.30777745107859716, "learning_rate": 1.4231463459318138e-06, "loss": 0.4791, "step": 1125 }, { "epoch": 1.6937844516251663, "grad_norm": 0.3802082342644203, "learning_rate": 1.4083922382093484e-06, "loss": 0.4925, "step": 1126 }, { "epoch": 1.6953050750807832, "grad_norm": 0.319724112059326, "learning_rate": 1.3937108156350437e-06, "loss": 0.5069, "step": 1127 }, { "epoch": 1.6968256985363999, "grad_norm": 0.32744646943463035, "learning_rate": 1.379102166098519e-06, "loss": 0.4892, "step": 1128 }, { "epoch": 1.6983463219920167, "grad_norm": 0.33236823986934483, "learning_rate": 1.3645663770537586e-06, "loss": 0.5106, "step": 1129 }, { "epoch": 1.6998669454476336, "grad_norm": 0.2988679855910158, "learning_rate": 1.350103535518552e-06, "loss": 0.4792, "step": 1130 }, { "epoch": 1.7013875689032503, "grad_norm": 0.31594423002452454, "learning_rate": 1.3357137280740085e-06, "loss": 0.5113, "step": 1131 }, { "epoch": 1.7029081923588671, "grad_norm": 0.30589751459809705, "learning_rate": 1.3213970408640052e-06, "loss": 0.4611, "step": 1132 }, { "epoch": 1.704428815814484, "grad_norm": 0.32052631965810274, "learning_rate": 1.307153559594711e-06, "loss": 0.5103, "step": 1133 }, { "epoch": 1.7059494392701007, "grad_norm": 0.3293578066194276, "learning_rate": 1.2929833695340287e-06, "loss": 0.4955, "step": 1134 }, { "epoch": 1.7074700627257176, "grad_norm": 0.30301511320734814, "learning_rate": 1.2788865555111225e-06, "loss": 0.457, "step": 1135 }, { "epoch": 1.7089906861813344, "grad_norm": 0.31166180214752437, "learning_rate": 1.264863201915883e-06, "loss": 0.4976, "step": 1136 }, { "epoch": 1.710511309636951, "grad_norm": 0.3163518521392497, "learning_rate": 1.2509133926984408e-06, "loss": 0.4966, "step": 1137 }, { "epoch": 1.712031933092568, "grad_norm": 0.3540193363362305, "learning_rate": 1.2370372113686581e-06, "loss": 0.5157, "step": 1138 }, { "epoch": 1.7135525565481848, "grad_norm": 0.3054649208614663, "learning_rate": 1.2232347409956184e-06, "loss": 0.4674, "step": 1139 }, { "epoch": 1.7150731800038015, "grad_norm": 0.31124066442535125, "learning_rate": 1.2095060642071457e-06, "loss": 0.4922, "step": 1140 }, { "epoch": 1.7165938034594184, "grad_norm": 0.31790128208883095, "learning_rate": 1.1958512631893e-06, "loss": 0.4973, "step": 1141 }, { "epoch": 1.7181144269150352, "grad_norm": 0.3165756985087125, "learning_rate": 1.1822704196858915e-06, "loss": 0.503, "step": 1142 }, { "epoch": 1.719635050370652, "grad_norm": 0.32017884174072114, "learning_rate": 1.1687636149979783e-06, "loss": 0.4865, "step": 1143 }, { "epoch": 1.7211556738262688, "grad_norm": 0.31138001106614027, "learning_rate": 1.155330929983396e-06, "loss": 0.5083, "step": 1144 }, { "epoch": 1.7226762972818856, "grad_norm": 0.3178786337486812, "learning_rate": 1.1419724450562654e-06, "loss": 0.4909, "step": 1145 }, { "epoch": 1.7241969207375023, "grad_norm": 0.30532938139355004, "learning_rate": 1.1286882401865111e-06, "loss": 0.472, "step": 1146 }, { "epoch": 1.7257175441931192, "grad_norm": 0.3296827200280989, "learning_rate": 1.1154783948993806e-06, "loss": 0.5244, "step": 1147 }, { "epoch": 1.727238167648736, "grad_norm": 0.31508066429540393, "learning_rate": 1.1023429882749769e-06, "loss": 0.5102, "step": 1148 }, { "epoch": 1.7287587911043527, "grad_norm": 0.3196487275495865, "learning_rate": 1.0892820989477799e-06, "loss": 0.4768, "step": 1149 }, { "epoch": 1.7302794145599696, "grad_norm": 0.30207628232666384, "learning_rate": 1.0762958051061657e-06, "loss": 0.4691, "step": 1150 }, { "epoch": 1.7318000380155865, "grad_norm": 0.3054462008267432, "learning_rate": 1.0633841844919598e-06, "loss": 0.4862, "step": 1151 }, { "epoch": 1.7333206614712031, "grad_norm": 0.3078006467585517, "learning_rate": 1.0505473143999502e-06, "loss": 0.4875, "step": 1152 }, { "epoch": 1.73484128492682, "grad_norm": 0.3053091170945442, "learning_rate": 1.0377852716774458e-06, "loss": 0.5073, "step": 1153 }, { "epoch": 1.7363619083824369, "grad_norm": 0.3076690595978604, "learning_rate": 1.0250981327237949e-06, "loss": 0.4949, "step": 1154 }, { "epoch": 1.7378825318380535, "grad_norm": 0.32895341988079463, "learning_rate": 1.012485973489944e-06, "loss": 0.4841, "step": 1155 }, { "epoch": 1.7394031552936704, "grad_norm": 0.3253480760618867, "learning_rate": 9.99948869477969e-07, "loss": 0.5004, "step": 1156 }, { "epoch": 1.7409237787492873, "grad_norm": 0.3259517227944411, "learning_rate": 9.874868957406476e-07, "loss": 0.5077, "step": 1157 }, { "epoch": 1.742444402204904, "grad_norm": 0.30986669451274373, "learning_rate": 9.751001268809732e-07, "loss": 0.4904, "step": 1158 }, { "epoch": 1.7439650256605208, "grad_norm": 0.3174807363161579, "learning_rate": 9.627886370517466e-07, "loss": 0.5058, "step": 1159 }, { "epoch": 1.7454856491161377, "grad_norm": 0.3060614656643119, "learning_rate": 9.505524999551009e-07, "loss": 0.4807, "step": 1160 }, { "epoch": 1.7470062725717543, "grad_norm": 0.3104685490140639, "learning_rate": 9.383917888420816e-07, "loss": 0.4892, "step": 1161 }, { "epoch": 1.7485268960273712, "grad_norm": 0.31773900649056236, "learning_rate": 9.263065765122013e-07, "loss": 0.4974, "step": 1162 }, { "epoch": 1.750047519482988, "grad_norm": 0.3218554742602446, "learning_rate": 9.142969353129946e-07, "loss": 0.4846, "step": 1163 }, { "epoch": 1.7515681429386047, "grad_norm": 0.3060712871920658, "learning_rate": 9.023629371396e-07, "loss": 0.4787, "step": 1164 }, { "epoch": 1.7530887663942216, "grad_norm": 0.3057039287255903, "learning_rate": 8.905046534343228e-07, "loss": 0.4888, "step": 1165 }, { "epoch": 1.7546093898498385, "grad_norm": 0.33483369302675053, "learning_rate": 8.787221551862048e-07, "loss": 0.5001, "step": 1166 }, { "epoch": 1.7561300133054552, "grad_norm": 0.3104906521034538, "learning_rate": 8.670155129306e-07, "loss": 0.496, "step": 1167 }, { "epoch": 1.757650636761072, "grad_norm": 0.3073575609788421, "learning_rate": 8.55384796748759e-07, "loss": 0.4905, "step": 1168 }, { "epoch": 1.759171260216689, "grad_norm": 0.3150067268140591, "learning_rate": 8.438300762674001e-07, "loss": 0.4872, "step": 1169 }, { "epoch": 1.7606918836723056, "grad_norm": 0.30605216319902684, "learning_rate": 8.323514206583039e-07, "loss": 0.4935, "step": 1170 }, { "epoch": 1.7622125071279224, "grad_norm": 0.3177954531343794, "learning_rate": 8.20948898637881e-07, "loss": 0.4916, "step": 1171 }, { "epoch": 1.7637331305835393, "grad_norm": 0.31828297184631565, "learning_rate": 8.09622578466782e-07, "loss": 0.4982, "step": 1172 }, { "epoch": 1.765253754039156, "grad_norm": 0.3112173769721995, "learning_rate": 7.983725279494758e-07, "loss": 0.4857, "step": 1173 }, { "epoch": 1.7667743774947728, "grad_norm": 0.31585454715709904, "learning_rate": 7.871988144338443e-07, "loss": 0.4784, "step": 1174 }, { "epoch": 1.7682950009503897, "grad_norm": 0.32127393064790427, "learning_rate": 7.761015048107861e-07, "loss": 0.5075, "step": 1175 }, { "epoch": 1.7698156244060064, "grad_norm": 0.30794451953754653, "learning_rate": 7.650806655138031e-07, "loss": 0.5038, "step": 1176 }, { "epoch": 1.7713362478616232, "grad_norm": 0.31181423286855536, "learning_rate": 7.541363625186259e-07, "loss": 0.5005, "step": 1177 }, { "epoch": 1.7728568713172401, "grad_norm": 0.32245652816729514, "learning_rate": 7.432686613427891e-07, "loss": 0.5095, "step": 1178 }, { "epoch": 1.7743774947728568, "grad_norm": 0.3242738181594943, "learning_rate": 7.324776270452638e-07, "loss": 0.52, "step": 1179 }, { "epoch": 1.7758981182284737, "grad_norm": 0.3110243759601029, "learning_rate": 7.217633242260516e-07, "loss": 0.464, "step": 1180 }, { "epoch": 1.7774187416840905, "grad_norm": 0.3319624462691413, "learning_rate": 7.111258170258142e-07, "loss": 0.5098, "step": 1181 }, { "epoch": 1.7789393651397072, "grad_norm": 0.3068597543181992, "learning_rate": 7.005651691254696e-07, "loss": 0.4815, "step": 1182 }, { "epoch": 1.780459988595324, "grad_norm": 0.3157827076249044, "learning_rate": 6.900814437458295e-07, "loss": 0.5024, "step": 1183 }, { "epoch": 1.781980612050941, "grad_norm": 0.30638021830468265, "learning_rate": 6.79674703647207e-07, "loss": 0.4798, "step": 1184 }, { "epoch": 1.7835012355065576, "grad_norm": 0.33125316725002335, "learning_rate": 6.693450111290479e-07, "loss": 0.5069, "step": 1185 }, { "epoch": 1.7850218589621745, "grad_norm": 0.325386774890156, "learning_rate": 6.59092428029562e-07, "loss": 0.5246, "step": 1186 }, { "epoch": 1.7865424824177913, "grad_norm": 0.2985022483714369, "learning_rate": 6.489170157253382e-07, "loss": 0.4852, "step": 1187 }, { "epoch": 1.788063105873408, "grad_norm": 0.3102718525215574, "learning_rate": 6.388188351309907e-07, "loss": 0.5006, "step": 1188 }, { "epoch": 1.7895837293290249, "grad_norm": 0.31889832653696915, "learning_rate": 6.287979466987909e-07, "loss": 0.49, "step": 1189 }, { "epoch": 1.7911043527846418, "grad_norm": 0.31148436424823517, "learning_rate": 6.188544104183059e-07, "loss": 0.4733, "step": 1190 }, { "epoch": 1.7926249762402584, "grad_norm": 0.31317825721031184, "learning_rate": 6.089882858160299e-07, "loss": 0.4879, "step": 1191 }, { "epoch": 1.7941455996958753, "grad_norm": 0.31818518365850107, "learning_rate": 5.991996319550448e-07, "loss": 0.5198, "step": 1192 }, { "epoch": 1.7956662231514922, "grad_norm": 0.3120540099643981, "learning_rate": 5.894885074346511e-07, "loss": 0.4969, "step": 1193 }, { "epoch": 1.7971868466071088, "grad_norm": 0.31290703317847507, "learning_rate": 5.79854970390031e-07, "loss": 0.4902, "step": 1194 }, { "epoch": 1.7987074700627257, "grad_norm": 0.3128686096799172, "learning_rate": 5.702990784918843e-07, "loss": 0.4837, "step": 1195 }, { "epoch": 1.8002280935183426, "grad_norm": 0.3152764097292228, "learning_rate": 5.608208889460964e-07, "loss": 0.5082, "step": 1196 }, { "epoch": 1.8017487169739592, "grad_norm": 0.3238711254005141, "learning_rate": 5.514204584933931e-07, "loss": 0.4932, "step": 1197 }, { "epoch": 1.803269340429576, "grad_norm": 0.32742976059180334, "learning_rate": 5.420978434089952e-07, "loss": 0.5006, "step": 1198 }, { "epoch": 1.804789963885193, "grad_norm": 0.3120702271210168, "learning_rate": 5.328530995022905e-07, "loss": 0.5077, "step": 1199 }, { "epoch": 1.8063105873408096, "grad_norm": 0.3176329082136289, "learning_rate": 5.236862821164854e-07, "loss": 0.5182, "step": 1200 }, { "epoch": 1.8078312107964265, "grad_norm": 0.3014290469656112, "learning_rate": 5.145974461282964e-07, "loss": 0.484, "step": 1201 }, { "epoch": 1.8093518342520434, "grad_norm": 0.3214449630653542, "learning_rate": 5.055866459475958e-07, "loss": 0.4808, "step": 1202 }, { "epoch": 1.81087245770766, "grad_norm": 0.32637831086774916, "learning_rate": 4.966539355171074e-07, "loss": 0.5317, "step": 1203 }, { "epoch": 1.812393081163277, "grad_norm": 0.3111952143588352, "learning_rate": 4.877993683120647e-07, "loss": 0.48, "step": 1204 }, { "epoch": 1.8139137046188938, "grad_norm": 0.3159439961230018, "learning_rate": 4.790229973399132e-07, "loss": 0.5355, "step": 1205 }, { "epoch": 1.8154343280745104, "grad_norm": 0.32606917992981577, "learning_rate": 4.7032487513997023e-07, "loss": 0.4854, "step": 1206 }, { "epoch": 1.8169549515301273, "grad_norm": 0.3099232220873522, "learning_rate": 4.617050537831259e-07, "loss": 0.4744, "step": 1207 }, { "epoch": 1.8184755749857442, "grad_norm": 0.3189690329444183, "learning_rate": 4.5316358487152045e-07, "loss": 0.5154, "step": 1208 }, { "epoch": 1.8199961984413608, "grad_norm": 0.3245800528868221, "learning_rate": 4.447005195382475e-07, "loss": 0.486, "step": 1209 }, { "epoch": 1.8215168218969777, "grad_norm": 0.3265685397280179, "learning_rate": 4.3631590844703997e-07, "loss": 0.5012, "step": 1210 }, { "epoch": 1.8230374453525946, "grad_norm": 0.3158110920991978, "learning_rate": 4.2800980179196426e-07, "loss": 0.4975, "step": 1211 }, { "epoch": 1.8245580688082113, "grad_norm": 0.30590044892564083, "learning_rate": 4.197822492971273e-07, "loss": 0.4761, "step": 1212 }, { "epoch": 1.8260786922638281, "grad_norm": 0.3296107652675287, "learning_rate": 4.1163330021637403e-07, "loss": 0.5024, "step": 1213 }, { "epoch": 1.827599315719445, "grad_norm": 0.31081520867244933, "learning_rate": 4.0356300333299696e-07, "loss": 0.4861, "step": 1214 }, { "epoch": 1.8291199391750617, "grad_norm": 0.31577863324052, "learning_rate": 3.9557140695943544e-07, "loss": 0.5023, "step": 1215 }, { "epoch": 1.8306405626306785, "grad_norm": 0.30718130485703893, "learning_rate": 3.8765855893699576e-07, "loss": 0.4847, "step": 1216 }, { "epoch": 1.8321611860862954, "grad_norm": 0.3044770096969053, "learning_rate": 3.798245066355609e-07, "loss": 0.4774, "step": 1217 }, { "epoch": 1.833681809541912, "grad_norm": 0.3233110025620584, "learning_rate": 3.7206929695330764e-07, "loss": 0.516, "step": 1218 }, { "epoch": 1.835202432997529, "grad_norm": 0.3184528078482147, "learning_rate": 3.643929763164223e-07, "loss": 0.5002, "step": 1219 }, { "epoch": 1.8367230564531458, "grad_norm": 0.3162950503495864, "learning_rate": 3.567955906788252e-07, "loss": 0.5208, "step": 1220 }, { "epoch": 1.8382436799087625, "grad_norm": 0.3032249377512437, "learning_rate": 3.492771855219034e-07, "loss": 0.4749, "step": 1221 }, { "epoch": 1.8397643033643793, "grad_norm": 0.32645828201108573, "learning_rate": 3.418378058542228e-07, "loss": 0.5212, "step": 1222 }, { "epoch": 1.8412849268199962, "grad_norm": 0.31381954625492664, "learning_rate": 3.344774962112713e-07, "loss": 0.4697, "step": 1223 }, { "epoch": 1.8428055502756129, "grad_norm": 0.3163864792118917, "learning_rate": 3.2719630065518326e-07, "loss": 0.5161, "step": 1224 }, { "epoch": 1.8443261737312298, "grad_norm": 0.29974584672007193, "learning_rate": 3.199942627744915e-07, "loss": 0.4769, "step": 1225 }, { "epoch": 1.8458467971868466, "grad_norm": 0.32581659125357904, "learning_rate": 3.1287142568384115e-07, "loss": 0.485, "step": 1226 }, { "epoch": 1.8473674206424633, "grad_norm": 0.32074295034449724, "learning_rate": 3.058278320237554e-07, "loss": 0.4896, "step": 1227 }, { "epoch": 1.8488880440980802, "grad_norm": 0.30909473328904485, "learning_rate": 2.9886352396036584e-07, "loss": 0.4768, "step": 1228 }, { "epoch": 1.850408667553697, "grad_norm": 0.32336521053629386, "learning_rate": 2.919785431851674e-07, "loss": 0.4999, "step": 1229 }, { "epoch": 1.8519292910093137, "grad_norm": 0.305906498433603, "learning_rate": 2.851729309147642e-07, "loss": 0.4998, "step": 1230 }, { "epoch": 1.8534499144649306, "grad_norm": 0.30991453199000357, "learning_rate": 2.7844672789062724e-07, "loss": 0.5011, "step": 1231 }, { "epoch": 1.8549705379205474, "grad_norm": 0.314943371794473, "learning_rate": 2.717999743788416e-07, "loss": 0.5096, "step": 1232 }, { "epoch": 1.856491161376164, "grad_norm": 0.3150834801717786, "learning_rate": 2.652327101698796e-07, "loss": 0.504, "step": 1233 }, { "epoch": 1.858011784831781, "grad_norm": 0.32620085243083274, "learning_rate": 2.587449745783506e-07, "loss": 0.5239, "step": 1234 }, { "epoch": 1.8595324082873979, "grad_norm": 0.31548675603667015, "learning_rate": 2.5233680644276673e-07, "loss": 0.4889, "step": 1235 }, { "epoch": 1.8610530317430145, "grad_norm": 0.32956663567551076, "learning_rate": 2.4600824412532154e-07, "loss": 0.5108, "step": 1236 }, { "epoch": 1.8625736551986314, "grad_norm": 0.29811803143834026, "learning_rate": 2.3975932551164486e-07, "loss": 0.4616, "step": 1237 }, { "epoch": 1.8640942786542483, "grad_norm": 0.3118381439138631, "learning_rate": 2.335900880105892e-07, "loss": 0.5051, "step": 1238 }, { "epoch": 1.865614902109865, "grad_norm": 0.30830068205472616, "learning_rate": 2.27500568553992e-07, "loss": 0.5012, "step": 1239 }, { "epoch": 1.867135525565482, "grad_norm": 0.304947912980323, "learning_rate": 2.2149080359646954e-07, "loss": 0.4583, "step": 1240 }, { "epoch": 1.8686561490210987, "grad_norm": 0.3114996653388056, "learning_rate": 2.1556082911518989e-07, "loss": 0.4898, "step": 1241 }, { "epoch": 1.8701767724767153, "grad_norm": 0.30177416022258724, "learning_rate": 2.0971068060965908e-07, "loss": 0.4965, "step": 1242 }, { "epoch": 1.8716973959323324, "grad_norm": 0.3309823897680051, "learning_rate": 2.0394039310150775e-07, "loss": 0.5185, "step": 1243 }, { "epoch": 1.873218019387949, "grad_norm": 0.32115239410578134, "learning_rate": 1.9825000113428182e-07, "loss": 0.5228, "step": 1244 }, { "epoch": 1.8747386428435657, "grad_norm": 0.3075640176251186, "learning_rate": 1.9263953877323952e-07, "loss": 0.4882, "step": 1245 }, { "epoch": 1.8762592662991828, "grad_norm": 0.31724674747730425, "learning_rate": 1.8710903960514225e-07, "loss": 0.5207, "step": 1246 }, { "epoch": 1.8777798897547995, "grad_norm": 0.3164911565117331, "learning_rate": 1.8165853673805754e-07, "loss": 0.5032, "step": 1247 }, { "epoch": 1.8793005132104161, "grad_norm": 0.30881540155384896, "learning_rate": 1.7628806280115286e-07, "loss": 0.5033, "step": 1248 }, { "epoch": 1.8808211366660332, "grad_norm": 0.31442427065508066, "learning_rate": 1.7099764994451522e-07, "loss": 0.5076, "step": 1249 }, { "epoch": 1.8823417601216499, "grad_norm": 0.31278434986007647, "learning_rate": 1.65787329838945e-07, "loss": 0.4989, "step": 1250 }, { "epoch": 1.8838623835772665, "grad_norm": 0.3025781565220422, "learning_rate": 1.60657133675771e-07, "loss": 0.5014, "step": 1251 }, { "epoch": 1.8853830070328836, "grad_norm": 0.30236403268840834, "learning_rate": 1.556070921666637e-07, "loss": 0.4653, "step": 1252 }, { "epoch": 1.8869036304885003, "grad_norm": 0.32144829401638425, "learning_rate": 1.5063723554345349e-07, "loss": 0.5022, "step": 1253 }, { "epoch": 1.888424253944117, "grad_norm": 0.3081811486598015, "learning_rate": 1.4574759355794846e-07, "loss": 0.468, "step": 1254 }, { "epoch": 1.889944877399734, "grad_norm": 0.3110080086508686, "learning_rate": 1.4093819548175095e-07, "loss": 0.4821, "step": 1255 }, { "epoch": 1.8914655008553507, "grad_norm": 0.30999522834801807, "learning_rate": 1.3620907010609206e-07, "loss": 0.475, "step": 1256 }, { "epoch": 1.8929861243109674, "grad_norm": 0.3244026681645176, "learning_rate": 1.3156024574165258e-07, "loss": 0.4967, "step": 1257 }, { "epoch": 1.8945067477665845, "grad_norm": 0.30596722710536933, "learning_rate": 1.2699175021839463e-07, "loss": 0.4794, "step": 1258 }, { "epoch": 1.896027371222201, "grad_norm": 0.33281008993735, "learning_rate": 1.2250361088539598e-07, "loss": 0.5202, "step": 1259 }, { "epoch": 1.8975479946778178, "grad_norm": 0.3092714395241654, "learning_rate": 1.1809585461068329e-07, "loss": 0.4897, "step": 1260 }, { "epoch": 1.8990686181334349, "grad_norm": 0.30328721741220116, "learning_rate": 1.1376850778108004e-07, "loss": 0.4734, "step": 1261 }, { "epoch": 1.9005892415890515, "grad_norm": 0.31671920046718893, "learning_rate": 1.0952159630203802e-07, "loss": 0.495, "step": 1262 }, { "epoch": 1.9021098650446682, "grad_norm": 0.301311429340991, "learning_rate": 1.0535514559748699e-07, "loss": 0.4879, "step": 1263 }, { "epoch": 1.9036304885002853, "grad_norm": 0.31829608262735665, "learning_rate": 1.0126918060968266e-07, "loss": 0.5054, "step": 1264 }, { "epoch": 1.905151111955902, "grad_norm": 0.3087155473121018, "learning_rate": 9.726372579905783e-08, "loss": 0.4972, "step": 1265 }, { "epoch": 1.9066717354115186, "grad_norm": 0.3080173664665244, "learning_rate": 9.33388051440734e-08, "loss": 0.5339, "step": 1266 }, { "epoch": 1.9081923588671357, "grad_norm": 0.3282773688288346, "learning_rate": 8.949444214107697e-08, "loss": 0.4775, "step": 1267 }, { "epoch": 1.9097129823227523, "grad_norm": 0.33035543404288714, "learning_rate": 8.573065980416298e-08, "loss": 0.5071, "step": 1268 }, { "epoch": 1.911233605778369, "grad_norm": 0.3215706061649872, "learning_rate": 8.204748066503122e-08, "loss": 0.5094, "step": 1269 }, { "epoch": 1.912754229233986, "grad_norm": 0.2988811222168615, "learning_rate": 7.844492677285754e-08, "loss": 0.4984, "step": 1270 }, { "epoch": 1.9142748526896027, "grad_norm": 0.31431504366013957, "learning_rate": 7.492301969415391e-08, "loss": 0.5072, "step": 1271 }, { "epoch": 1.9157954761452194, "grad_norm": 0.3079234038112652, "learning_rate": 7.148178051264804e-08, "loss": 0.4883, "step": 1272 }, { "epoch": 1.9173160996008365, "grad_norm": 0.30748269412345614, "learning_rate": 6.812122982915408e-08, "loss": 0.5152, "step": 1273 }, { "epoch": 1.9188367230564531, "grad_norm": 0.3111607368191631, "learning_rate": 6.484138776144467e-08, "loss": 0.4992, "step": 1274 }, { "epoch": 1.9203573465120698, "grad_norm": 0.3130854739102124, "learning_rate": 6.164227394413516e-08, "loss": 0.4947, "step": 1275 }, { "epoch": 1.921877969967687, "grad_norm": 0.29735483922959655, "learning_rate": 5.852390752856624e-08, "loss": 0.4717, "step": 1276 }, { "epoch": 1.9233985934233035, "grad_norm": 0.30465543049913124, "learning_rate": 5.548630718268813e-08, "loss": 0.4977, "step": 1277 }, { "epoch": 1.9249192168789202, "grad_norm": 0.3133152057012344, "learning_rate": 5.252949109094924e-08, "loss": 0.4839, "step": 1278 }, { "epoch": 1.9264398403345373, "grad_norm": 0.32301869827083196, "learning_rate": 4.965347695418483e-08, "loss": 0.4755, "step": 1279 }, { "epoch": 1.927960463790154, "grad_norm": 0.30143456487467507, "learning_rate": 4.685828198951479e-08, "loss": 0.4901, "step": 1280 }, { "epoch": 1.9294810872457706, "grad_norm": 0.3198896489356579, "learning_rate": 4.4143922930239736e-08, "loss": 0.5014, "step": 1281 }, { "epoch": 1.9310017107013877, "grad_norm": 0.29774733042477136, "learning_rate": 4.1510416025737274e-08, "loss": 0.4598, "step": 1282 }, { "epoch": 1.9325223341570044, "grad_norm": 0.3117201437762231, "learning_rate": 3.895777704137174e-08, "loss": 0.4718, "step": 1283 }, { "epoch": 1.934042957612621, "grad_norm": 0.30970105466061604, "learning_rate": 3.648602125839037e-08, "loss": 0.4925, "step": 1284 }, { "epoch": 1.9355635810682381, "grad_norm": 0.3032281003832038, "learning_rate": 3.409516347384058e-08, "loss": 0.4876, "step": 1285 }, { "epoch": 1.9370842045238548, "grad_norm": 0.29947248473886334, "learning_rate": 3.178521800047818e-08, "loss": 0.4652, "step": 1286 }, { "epoch": 1.9386048279794714, "grad_norm": 0.3101511675568205, "learning_rate": 2.9556198666678676e-08, "loss": 0.487, "step": 1287 }, { "epoch": 1.9401254514350885, "grad_norm": 0.3031822788379352, "learning_rate": 2.7408118816360456e-08, "loss": 0.4906, "step": 1288 }, { "epoch": 1.9416460748907052, "grad_norm": 0.3022071103062291, "learning_rate": 2.5340991308897615e-08, "loss": 0.4749, "step": 1289 }, { "epoch": 1.9431666983463218, "grad_norm": 0.3376492826020699, "learning_rate": 2.3354828519049215e-08, "loss": 0.5193, "step": 1290 }, { "epoch": 1.944687321801939, "grad_norm": 0.3049763317749277, "learning_rate": 2.144964233688558e-08, "loss": 0.492, "step": 1291 }, { "epoch": 1.9462079452575556, "grad_norm": 0.31915116281549216, "learning_rate": 1.9625444167708558e-08, "loss": 0.5077, "step": 1292 }, { "epoch": 1.9477285687131722, "grad_norm": 0.29789807615910413, "learning_rate": 1.7882244931992866e-08, "loss": 0.4802, "step": 1293 }, { "epoch": 1.9492491921687893, "grad_norm": 0.29936729470728324, "learning_rate": 1.622005506531838e-08, "loss": 0.4683, "step": 1294 }, { "epoch": 1.950769815624406, "grad_norm": 0.30557270295233097, "learning_rate": 1.4638884518302449e-08, "loss": 0.493, "step": 1295 }, { "epoch": 1.9522904390800226, "grad_norm": 0.3152644678698385, "learning_rate": 1.313874275654573e-08, "loss": 0.4872, "step": 1296 }, { "epoch": 1.9538110625356397, "grad_norm": 0.32155409817146613, "learning_rate": 1.1719638760576535e-08, "loss": 0.4898, "step": 1297 }, { "epoch": 1.9553316859912564, "grad_norm": 0.322441859921911, "learning_rate": 1.0381581025792154e-08, "loss": 0.5123, "step": 1298 }, { "epoch": 1.9568523094468733, "grad_norm": 0.3075162182309576, "learning_rate": 9.124577562409214e-09, "loss": 0.5075, "step": 1299 }, { "epoch": 1.9583729329024901, "grad_norm": 0.31214568646689916, "learning_rate": 7.948635895418555e-09, "loss": 0.5012, "step": 1300 }, { "epoch": 1.9598935563581068, "grad_norm": 0.3160685661400899, "learning_rate": 6.853763064538582e-09, "loss": 0.5009, "step": 1301 }, { "epoch": 1.9614141798137237, "grad_norm": 0.31359686192724745, "learning_rate": 5.8399656241716505e-09, "loss": 0.4962, "step": 1302 }, { "epoch": 1.9629348032693406, "grad_norm": 0.31993025547008763, "learning_rate": 4.907249643367956e-09, "loss": 0.4837, "step": 1303 }, { "epoch": 1.9644554267249572, "grad_norm": 0.30154655369767214, "learning_rate": 4.055620705787927e-09, "loss": 0.4907, "step": 1304 }, { "epoch": 1.965976050180574, "grad_norm": 0.3082339737443702, "learning_rate": 3.2850839096646157e-09, "loss": 0.49, "step": 1305 }, { "epoch": 1.967496673636191, "grad_norm": 0.31262848669610493, "learning_rate": 2.5956438677796293e-09, "loss": 0.5033, "step": 1306 }, { "epoch": 1.9690172970918076, "grad_norm": 0.3078427537941908, "learning_rate": 1.9873047074345453e-09, "loss": 0.4793, "step": 1307 }, { "epoch": 1.9705379205474245, "grad_norm": 0.3074339790589038, "learning_rate": 1.4600700704193227e-09, "loss": 0.4843, "step": 1308 }, { "epoch": 1.9720585440030414, "grad_norm": 0.31784494867260554, "learning_rate": 1.0139431130017695e-09, "loss": 0.4984, "step": 1309 }, { "epoch": 1.973579167458658, "grad_norm": 0.32037784960225285, "learning_rate": 6.489265058974569e-10, "loss": 0.5143, "step": 1310 }, { "epoch": 1.975099790914275, "grad_norm": 0.3115474055785006, "learning_rate": 3.6502243426069227e-10, "loss": 0.4958, "step": 1311 }, { "epoch": 1.9766204143698918, "grad_norm": 0.30740455562554947, "learning_rate": 1.622325976709804e-10, "loss": 0.4825, "step": 1312 }, { "epoch": 1.9781410378255084, "grad_norm": 0.3192120605150713, "learning_rate": 4.055821011798022e-11, "loss": 0.5028, "step": 1313 }, { "epoch": 1.9796616612811253, "grad_norm": 0.3128442843272166, "learning_rate": 0.0, "loss": 0.4886, "step": 1314 } ], "logging_steps": 1, "max_steps": 1314, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 329, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4190576087400448.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }