Llama-2-7b-sft-human-rm / trainer_state.json
AmberYifan's picture
Model save
73a6adf verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 2502,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.001199040767386091,
"grad_norm": 20.168596571210266,
"learning_rate": 1.99203187250996e-09,
"logits/chosen": -1.484375,
"logits/rejected": -1.4609375,
"logps/chosen": -226.0,
"logps/rejected": -218.0,
"loss": 0.6914,
"rewards/accuracies": 0.0,
"rewards/chosen": 0.0,
"rewards/margins": 0.0,
"rewards/rejected": 0.0,
"step": 1
},
{
"epoch": 0.011990407673860911,
"grad_norm": 18.74753353986141,
"learning_rate": 1.99203187250996e-08,
"logits/chosen": -1.5625,
"logits/rejected": -1.5390625,
"logps/chosen": -185.0,
"logps/rejected": -176.0,
"loss": 0.6943,
"rewards/accuracies": 0.125,
"rewards/chosen": 0.006256103515625,
"rewards/margins": -0.0093994140625,
"rewards/rejected": 0.015625,
"step": 10
},
{
"epoch": 0.023980815347721823,
"grad_norm": 20.00098472032445,
"learning_rate": 3.98406374501992e-08,
"logits/chosen": -1.5,
"logits/rejected": -1.4921875,
"logps/chosen": -182.0,
"logps/rejected": -175.0,
"loss": 0.6993,
"rewards/accuracies": 0.1875,
"rewards/chosen": 0.00750732421875,
"rewards/margins": -0.01190185546875,
"rewards/rejected": 0.0194091796875,
"step": 20
},
{
"epoch": 0.03597122302158273,
"grad_norm": 19.082683285927587,
"learning_rate": 5.97609561752988e-08,
"logits/chosen": -1.4765625,
"logits/rejected": -1.5,
"logps/chosen": -157.0,
"logps/rejected": -167.0,
"loss": 0.6934,
"rewards/accuracies": 0.20000000298023224,
"rewards/chosen": 0.01531982421875,
"rewards/margins": 0.000308990478515625,
"rewards/rejected": 0.0150146484375,
"step": 30
},
{
"epoch": 0.047961630695443645,
"grad_norm": 18.804321980067716,
"learning_rate": 7.96812749003984e-08,
"logits/chosen": -1.4453125,
"logits/rejected": -1.4375,
"logps/chosen": -172.0,
"logps/rejected": -160.0,
"loss": 0.6978,
"rewards/accuracies": 0.17499999701976776,
"rewards/chosen": 0.004058837890625,
"rewards/margins": -0.01251220703125,
"rewards/rejected": 0.0166015625,
"step": 40
},
{
"epoch": 0.05995203836930456,
"grad_norm": 17.83886367720322,
"learning_rate": 9.9601593625498e-08,
"logits/chosen": -1.4921875,
"logits/rejected": -1.4921875,
"logps/chosen": -185.0,
"logps/rejected": -179.0,
"loss": 0.6931,
"rewards/accuracies": 0.23749999701976776,
"rewards/chosen": 0.0106201171875,
"rewards/margins": -0.000926971435546875,
"rewards/rejected": 0.0115966796875,
"step": 50
},
{
"epoch": 0.07194244604316546,
"grad_norm": 22.450580380910267,
"learning_rate": 1.195219123505976e-07,
"logits/chosen": -1.5078125,
"logits/rejected": -1.5078125,
"logps/chosen": -170.0,
"logps/rejected": -180.0,
"loss": 0.6885,
"rewards/accuracies": 0.2874999940395355,
"rewards/chosen": 0.01190185546875,
"rewards/margins": 0.01031494140625,
"rewards/rejected": 0.0015716552734375,
"step": 60
},
{
"epoch": 0.08393285371702638,
"grad_norm": 19.042392981886824,
"learning_rate": 1.394422310756972e-07,
"logits/chosen": -1.546875,
"logits/rejected": -1.4765625,
"logps/chosen": -180.0,
"logps/rejected": -172.0,
"loss": 0.6887,
"rewards/accuracies": 0.20000000298023224,
"rewards/chosen": 0.01531982421875,
"rewards/margins": -0.00061798095703125,
"rewards/rejected": 0.0159912109375,
"step": 70
},
{
"epoch": 0.09592326139088729,
"grad_norm": 19.035173662979318,
"learning_rate": 1.593625498007968e-07,
"logits/chosen": -1.5390625,
"logits/rejected": -1.5234375,
"logps/chosen": -178.0,
"logps/rejected": -191.0,
"loss": 0.6917,
"rewards/accuracies": 0.22499999403953552,
"rewards/chosen": 0.01190185546875,
"rewards/margins": -0.00250244140625,
"rewards/rejected": 0.014404296875,
"step": 80
},
{
"epoch": 0.1079136690647482,
"grad_norm": 19.033868457427005,
"learning_rate": 1.7928286852589642e-07,
"logits/chosen": -1.484375,
"logits/rejected": -1.46875,
"logps/chosen": -197.0,
"logps/rejected": -190.0,
"loss": 0.6935,
"rewards/accuracies": 0.20000000298023224,
"rewards/chosen": 0.0174560546875,
"rewards/margins": 0.0031280517578125,
"rewards/rejected": 0.014404296875,
"step": 90
},
{
"epoch": 0.11990407673860912,
"grad_norm": 19.178986080183737,
"learning_rate": 1.99203187250996e-07,
"logits/chosen": -1.4296875,
"logits/rejected": -1.46875,
"logps/chosen": -162.0,
"logps/rejected": -175.0,
"loss": 0.6941,
"rewards/accuracies": 0.21250000596046448,
"rewards/chosen": 0.00909423828125,
"rewards/margins": -0.00750732421875,
"rewards/rejected": 0.0166015625,
"step": 100
},
{
"epoch": 0.13189448441247004,
"grad_norm": 18.51014085479843,
"learning_rate": 2.191235059760956e-07,
"logits/chosen": -1.53125,
"logits/rejected": -1.5,
"logps/chosen": -170.0,
"logps/rejected": -186.0,
"loss": 0.69,
"rewards/accuracies": 0.3375000059604645,
"rewards/chosen": 0.0322265625,
"rewards/margins": 0.015625,
"rewards/rejected": 0.0166015625,
"step": 110
},
{
"epoch": 0.14388489208633093,
"grad_norm": 18.632268534190302,
"learning_rate": 2.390438247011952e-07,
"logits/chosen": -1.5,
"logits/rejected": -1.5078125,
"logps/chosen": -176.0,
"logps/rejected": -158.0,
"loss": 0.6893,
"rewards/accuracies": 0.3125,
"rewards/chosen": 0.031982421875,
"rewards/margins": 0.0234375,
"rewards/rejected": 0.0084228515625,
"step": 120
},
{
"epoch": 0.15587529976019185,
"grad_norm": 20.464332592180558,
"learning_rate": 2.589641434262948e-07,
"logits/chosen": -1.4765625,
"logits/rejected": -1.4375,
"logps/chosen": -181.0,
"logps/rejected": -173.0,
"loss": 0.6818,
"rewards/accuracies": 0.36250001192092896,
"rewards/chosen": 0.024658203125,
"rewards/margins": 0.017822265625,
"rewards/rejected": 0.00689697265625,
"step": 130
},
{
"epoch": 0.16786570743405277,
"grad_norm": 18.785864772733888,
"learning_rate": 2.788844621513944e-07,
"logits/chosen": -1.4453125,
"logits/rejected": -1.484375,
"logps/chosen": -184.0,
"logps/rejected": -178.0,
"loss": 0.6853,
"rewards/accuracies": 0.3125,
"rewards/chosen": 0.032470703125,
"rewards/margins": 0.01470947265625,
"rewards/rejected": 0.017822265625,
"step": 140
},
{
"epoch": 0.17985611510791366,
"grad_norm": 18.715152675718606,
"learning_rate": 2.98804780876494e-07,
"logits/chosen": -1.515625,
"logits/rejected": -1.453125,
"logps/chosen": -181.0,
"logps/rejected": -183.0,
"loss": 0.6824,
"rewards/accuracies": 0.4000000059604645,
"rewards/chosen": 0.0294189453125,
"rewards/margins": 0.027587890625,
"rewards/rejected": 0.0018768310546875,
"step": 150
},
{
"epoch": 0.19184652278177458,
"grad_norm": 19.40478993623526,
"learning_rate": 3.187250996015936e-07,
"logits/chosen": -1.4453125,
"logits/rejected": -1.4453125,
"logps/chosen": -174.0,
"logps/rejected": -182.0,
"loss": 0.6854,
"rewards/accuracies": 0.38749998807907104,
"rewards/chosen": 0.0419921875,
"rewards/margins": 0.026611328125,
"rewards/rejected": 0.01531982421875,
"step": 160
},
{
"epoch": 0.2038369304556355,
"grad_norm": 18.06542256703035,
"learning_rate": 3.386454183266932e-07,
"logits/chosen": -1.5078125,
"logits/rejected": -1.484375,
"logps/chosen": -191.0,
"logps/rejected": -188.0,
"loss": 0.6775,
"rewards/accuracies": 0.512499988079071,
"rewards/chosen": 0.043212890625,
"rewards/margins": 0.045654296875,
"rewards/rejected": -0.00250244140625,
"step": 170
},
{
"epoch": 0.2158273381294964,
"grad_norm": 18.751232483418,
"learning_rate": 3.5856573705179284e-07,
"logits/chosen": -1.4609375,
"logits/rejected": -1.4375,
"logps/chosen": -180.0,
"logps/rejected": -183.0,
"loss": 0.673,
"rewards/accuracies": 0.574999988079071,
"rewards/chosen": 0.0218505859375,
"rewards/margins": 0.04638671875,
"rewards/rejected": -0.0244140625,
"step": 180
},
{
"epoch": 0.2278177458033573,
"grad_norm": 17.469498378813427,
"learning_rate": 3.784860557768924e-07,
"logits/chosen": -1.5078125,
"logits/rejected": -1.5234375,
"logps/chosen": -169.0,
"logps/rejected": -158.0,
"loss": 0.6819,
"rewards/accuracies": 0.48750001192092896,
"rewards/chosen": 0.01312255859375,
"rewards/margins": 0.031005859375,
"rewards/rejected": -0.017822265625,
"step": 190
},
{
"epoch": 0.23980815347721823,
"grad_norm": 19.476666022603656,
"learning_rate": 3.98406374501992e-07,
"logits/chosen": -1.4609375,
"logits/rejected": -1.4453125,
"logps/chosen": -181.0,
"logps/rejected": -177.0,
"loss": 0.6788,
"rewards/accuracies": 0.44999998807907104,
"rewards/chosen": 0.01190185546875,
"rewards/margins": 0.0281982421875,
"rewards/rejected": -0.0162353515625,
"step": 200
},
{
"epoch": 0.2517985611510791,
"grad_norm": 15.783370845580528,
"learning_rate": 4.1832669322709163e-07,
"logits/chosen": -1.484375,
"logits/rejected": -1.5234375,
"logps/chosen": -174.0,
"logps/rejected": -182.0,
"loss": 0.6673,
"rewards/accuracies": 0.48750001192092896,
"rewards/chosen": 0.029052734375,
"rewards/margins": 0.05859375,
"rewards/rejected": -0.0294189453125,
"step": 210
},
{
"epoch": 0.2637889688249401,
"grad_norm": 17.69748999194774,
"learning_rate": 4.382470119521912e-07,
"logits/chosen": -1.4453125,
"logits/rejected": -1.4453125,
"logps/chosen": -185.0,
"logps/rejected": -179.0,
"loss": 0.6637,
"rewards/accuracies": 0.6000000238418579,
"rewards/chosen": -0.0244140625,
"rewards/margins": 0.058837890625,
"rewards/rejected": -0.08349609375,
"step": 220
},
{
"epoch": 0.27577937649880097,
"grad_norm": 18.608453625360337,
"learning_rate": 4.581673306772908e-07,
"logits/chosen": -1.484375,
"logits/rejected": -1.46875,
"logps/chosen": -174.0,
"logps/rejected": -178.0,
"loss": 0.6603,
"rewards/accuracies": 0.574999988079071,
"rewards/chosen": -0.045654296875,
"rewards/margins": 0.07470703125,
"rewards/rejected": -0.1201171875,
"step": 230
},
{
"epoch": 0.28776978417266186,
"grad_norm": 20.06258278752274,
"learning_rate": 4.780876494023904e-07,
"logits/chosen": -1.4765625,
"logits/rejected": -1.4765625,
"logps/chosen": -186.0,
"logps/rejected": -172.0,
"loss": 0.652,
"rewards/accuracies": 0.6499999761581421,
"rewards/chosen": -0.022216796875,
"rewards/margins": 0.11474609375,
"rewards/rejected": -0.13671875,
"step": 240
},
{
"epoch": 0.2997601918465228,
"grad_norm": 18.760059177733364,
"learning_rate": 4.9800796812749e-07,
"logits/chosen": -1.3984375,
"logits/rejected": -1.3984375,
"logps/chosen": -161.0,
"logps/rejected": -173.0,
"loss": 0.649,
"rewards/accuracies": 0.550000011920929,
"rewards/chosen": -0.130859375,
"rewards/margins": 0.07568359375,
"rewards/rejected": -0.2060546875,
"step": 250
},
{
"epoch": 0.3117505995203837,
"grad_norm": 18.23235067079184,
"learning_rate": 4.980008884940026e-07,
"logits/chosen": -1.4453125,
"logits/rejected": -1.4375,
"logps/chosen": -174.0,
"logps/rejected": -174.0,
"loss": 0.649,
"rewards/accuracies": 0.6625000238418579,
"rewards/chosen": -0.1357421875,
"rewards/margins": 0.1435546875,
"rewards/rejected": -0.279296875,
"step": 260
},
{
"epoch": 0.3237410071942446,
"grad_norm": 16.049599214987126,
"learning_rate": 4.957796534873389e-07,
"logits/chosen": -1.453125,
"logits/rejected": -1.3984375,
"logps/chosen": -175.0,
"logps/rejected": -171.0,
"loss": 0.6365,
"rewards/accuracies": 0.699999988079071,
"rewards/chosen": -0.2109375,
"rewards/margins": 0.150390625,
"rewards/rejected": -0.36328125,
"step": 270
},
{
"epoch": 0.33573141486810554,
"grad_norm": 16.211195894914958,
"learning_rate": 4.935584184806753e-07,
"logits/chosen": -1.4375,
"logits/rejected": -1.3984375,
"logps/chosen": -176.0,
"logps/rejected": -177.0,
"loss": 0.624,
"rewards/accuracies": 0.5874999761581421,
"rewards/chosen": -0.330078125,
"rewards/margins": 0.1513671875,
"rewards/rejected": -0.48046875,
"step": 280
},
{
"epoch": 0.34772182254196643,
"grad_norm": 20.12569666248882,
"learning_rate": 4.913371834740116e-07,
"logits/chosen": -1.46875,
"logits/rejected": -1.4296875,
"logps/chosen": -176.0,
"logps/rejected": -180.0,
"loss": 0.6387,
"rewards/accuracies": 0.5,
"rewards/chosen": -0.498046875,
"rewards/margins": 0.11279296875,
"rewards/rejected": -0.609375,
"step": 290
},
{
"epoch": 0.3597122302158273,
"grad_norm": 17.476880445567698,
"learning_rate": 4.891159484673478e-07,
"logits/chosen": -1.4296875,
"logits/rejected": -1.4375,
"logps/chosen": -187.0,
"logps/rejected": -189.0,
"loss": 0.624,
"rewards/accuracies": 0.625,
"rewards/chosen": -0.58984375,
"rewards/margins": 0.1591796875,
"rewards/rejected": -0.74609375,
"step": 300
},
{
"epoch": 0.37170263788968827,
"grad_norm": 15.507480637830666,
"learning_rate": 4.868947134606841e-07,
"logits/chosen": -1.4140625,
"logits/rejected": -1.4375,
"logps/chosen": -186.0,
"logps/rejected": -196.0,
"loss": 0.5885,
"rewards/accuracies": 0.7124999761581421,
"rewards/chosen": -0.734375,
"rewards/margins": 0.3203125,
"rewards/rejected": -1.0546875,
"step": 310
},
{
"epoch": 0.38369304556354916,
"grad_norm": 18.563307603684667,
"learning_rate": 4.846734784540204e-07,
"logits/chosen": -1.4765625,
"logits/rejected": -1.453125,
"logps/chosen": -199.0,
"logps/rejected": -203.0,
"loss": 0.587,
"rewards/accuracies": 0.6499999761581421,
"rewards/chosen": -0.92578125,
"rewards/margins": 0.2734375,
"rewards/rejected": -1.1953125,
"step": 320
},
{
"epoch": 0.39568345323741005,
"grad_norm": 23.31061683348206,
"learning_rate": 4.824522434473567e-07,
"logits/chosen": -1.46875,
"logits/rejected": -1.4140625,
"logps/chosen": -203.0,
"logps/rejected": -202.0,
"loss": 0.5723,
"rewards/accuracies": 0.637499988079071,
"rewards/chosen": -1.3125,
"rewards/margins": 0.29296875,
"rewards/rejected": -1.609375,
"step": 330
},
{
"epoch": 0.407673860911271,
"grad_norm": 16.604718922811248,
"learning_rate": 4.80231008440693e-07,
"logits/chosen": -1.421875,
"logits/rejected": -1.3984375,
"logps/chosen": -194.0,
"logps/rejected": -195.0,
"loss": 0.55,
"rewards/accuracies": 0.75,
"rewards/chosen": -1.09375,
"rewards/margins": 0.466796875,
"rewards/rejected": -1.5625,
"step": 340
},
{
"epoch": 0.4196642685851319,
"grad_norm": 21.812258860080746,
"learning_rate": 4.780097734340293e-07,
"logits/chosen": -1.4375,
"logits/rejected": -1.4140625,
"logps/chosen": -191.0,
"logps/rejected": -199.0,
"loss": 0.5831,
"rewards/accuracies": 0.6499999761581421,
"rewards/chosen": -1.2734375,
"rewards/margins": 0.349609375,
"rewards/rejected": -1.6171875,
"step": 350
},
{
"epoch": 0.4316546762589928,
"grad_norm": 21.975218415102187,
"learning_rate": 4.757885384273656e-07,
"logits/chosen": -1.453125,
"logits/rejected": -1.453125,
"logps/chosen": -192.0,
"logps/rejected": -204.0,
"loss": 0.5577,
"rewards/accuracies": 0.675000011920929,
"rewards/chosen": -1.65625,
"rewards/margins": 0.4375,
"rewards/rejected": -2.09375,
"step": 360
},
{
"epoch": 0.44364508393285373,
"grad_norm": 18.21517367189722,
"learning_rate": 4.7356730342070187e-07,
"logits/chosen": -1.40625,
"logits/rejected": -1.421875,
"logps/chosen": -198.0,
"logps/rejected": -202.0,
"loss": 0.5576,
"rewards/accuracies": 0.762499988079071,
"rewards/chosen": -1.6171875,
"rewards/margins": 0.5390625,
"rewards/rejected": -2.15625,
"step": 370
},
{
"epoch": 0.4556354916067146,
"grad_norm": 17.44710723350182,
"learning_rate": 4.713460684140382e-07,
"logits/chosen": -1.40625,
"logits/rejected": -1.421875,
"logps/chosen": -195.0,
"logps/rejected": -195.0,
"loss": 0.5889,
"rewards/accuracies": 0.625,
"rewards/chosen": -1.65625,
"rewards/margins": 0.310546875,
"rewards/rejected": -1.96875,
"step": 380
},
{
"epoch": 0.4676258992805755,
"grad_norm": 17.175546716735866,
"learning_rate": 4.691248334073745e-07,
"logits/chosen": -1.359375,
"logits/rejected": -1.4296875,
"logps/chosen": -187.0,
"logps/rejected": -199.0,
"loss": 0.5365,
"rewards/accuracies": 0.6625000238418579,
"rewards/chosen": -1.390625,
"rewards/margins": 0.4375,
"rewards/rejected": -1.828125,
"step": 390
},
{
"epoch": 0.47961630695443647,
"grad_norm": 21.590406692659766,
"learning_rate": 4.6690359840071075e-07,
"logits/chosen": -1.40625,
"logits/rejected": -1.3515625,
"logps/chosen": -171.0,
"logps/rejected": -185.0,
"loss": 0.5199,
"rewards/accuracies": 0.737500011920929,
"rewards/chosen": -0.6640625,
"rewards/margins": 0.55078125,
"rewards/rejected": -1.21875,
"step": 400
},
{
"epoch": 0.49160671462829736,
"grad_norm": 17.161105944130238,
"learning_rate": 4.646823633940471e-07,
"logits/chosen": -1.4609375,
"logits/rejected": -1.5,
"logps/chosen": -185.0,
"logps/rejected": -194.0,
"loss": 0.5379,
"rewards/accuracies": 0.699999988079071,
"rewards/chosen": -0.5625,
"rewards/margins": 0.5078125,
"rewards/rejected": -1.0703125,
"step": 410
},
{
"epoch": 0.5035971223021583,
"grad_norm": 20.897701687493683,
"learning_rate": 4.6246112838738336e-07,
"logits/chosen": -1.4140625,
"logits/rejected": -1.421875,
"logps/chosen": -187.0,
"logps/rejected": -187.0,
"loss": 0.5262,
"rewards/accuracies": 0.737500011920929,
"rewards/chosen": -1.8125,
"rewards/margins": 0.62890625,
"rewards/rejected": -2.4375,
"step": 420
},
{
"epoch": 0.5155875299760192,
"grad_norm": 21.505932353833877,
"learning_rate": 4.602398933807197e-07,
"logits/chosen": -1.4453125,
"logits/rejected": -1.5078125,
"logps/chosen": -192.0,
"logps/rejected": -225.0,
"loss": 0.5387,
"rewards/accuracies": 0.699999988079071,
"rewards/chosen": -1.5078125,
"rewards/margins": 0.435546875,
"rewards/rejected": -1.9375,
"step": 430
},
{
"epoch": 0.5275779376498801,
"grad_norm": 20.010972596147028,
"learning_rate": 4.5801865837405597e-07,
"logits/chosen": -1.4296875,
"logits/rejected": -1.4375,
"logps/chosen": -200.0,
"logps/rejected": -203.0,
"loss": 0.5316,
"rewards/accuracies": 0.699999988079071,
"rewards/chosen": -1.1796875,
"rewards/margins": 0.53125,
"rewards/rejected": -1.703125,
"step": 440
},
{
"epoch": 0.539568345323741,
"grad_norm": 19.522837250813602,
"learning_rate": 4.5579742336739224e-07,
"logits/chosen": -1.421875,
"logits/rejected": -1.375,
"logps/chosen": -188.0,
"logps/rejected": -178.0,
"loss": 0.5376,
"rewards/accuracies": 0.8125,
"rewards/chosen": -0.8046875,
"rewards/margins": 0.7578125,
"rewards/rejected": -1.5625,
"step": 450
},
{
"epoch": 0.5515587529976019,
"grad_norm": 19.08163221164428,
"learning_rate": 4.535761883607285e-07,
"logits/chosen": -1.453125,
"logits/rejected": -1.4453125,
"logps/chosen": -209.0,
"logps/rejected": -197.0,
"loss": 0.5262,
"rewards/accuracies": 0.699999988079071,
"rewards/chosen": -1.0546875,
"rewards/margins": 0.42578125,
"rewards/rejected": -1.4765625,
"step": 460
},
{
"epoch": 0.5635491606714629,
"grad_norm": 19.252670782360564,
"learning_rate": 4.513549533540648e-07,
"logits/chosen": -1.4453125,
"logits/rejected": -1.4453125,
"logps/chosen": -188.0,
"logps/rejected": -195.0,
"loss": 0.4992,
"rewards/accuracies": 0.762499988079071,
"rewards/chosen": -0.91796875,
"rewards/margins": 0.5859375,
"rewards/rejected": -1.5,
"step": 470
},
{
"epoch": 0.5755395683453237,
"grad_norm": 25.774716921623085,
"learning_rate": 4.491337183474012e-07,
"logits/chosen": -1.4375,
"logits/rejected": -1.421875,
"logps/chosen": -199.0,
"logps/rejected": -194.0,
"loss": 0.5192,
"rewards/accuracies": 0.737500011920929,
"rewards/chosen": -1.71875,
"rewards/margins": 0.55078125,
"rewards/rejected": -2.28125,
"step": 480
},
{
"epoch": 0.5875299760191847,
"grad_norm": 14.548009215483642,
"learning_rate": 4.4691248334073746e-07,
"logits/chosen": -1.4765625,
"logits/rejected": -1.4375,
"logps/chosen": -197.0,
"logps/rejected": -211.0,
"loss": 0.4748,
"rewards/accuracies": 0.8125,
"rewards/chosen": -1.6484375,
"rewards/margins": 0.77734375,
"rewards/rejected": -2.421875,
"step": 490
},
{
"epoch": 0.5995203836930456,
"grad_norm": 25.516181574963138,
"learning_rate": 4.4469124833407373e-07,
"logits/chosen": -1.3828125,
"logits/rejected": -1.375,
"logps/chosen": -188.0,
"logps/rejected": -190.0,
"loss": 0.5534,
"rewards/accuracies": 0.6875,
"rewards/chosen": -2.03125,
"rewards/margins": 0.3984375,
"rewards/rejected": -2.421875,
"step": 500
},
{
"epoch": 0.6115107913669064,
"grad_norm": 22.58747620927269,
"learning_rate": 4.4247001332741e-07,
"logits/chosen": -1.40625,
"logits/rejected": -1.4140625,
"logps/chosen": -189.0,
"logps/rejected": -210.0,
"loss": 0.4916,
"rewards/accuracies": 0.7875000238418579,
"rewards/chosen": -1.5078125,
"rewards/margins": 0.8046875,
"rewards/rejected": -2.3125,
"step": 510
},
{
"epoch": 0.6235011990407674,
"grad_norm": 17.897088219787737,
"learning_rate": 4.402487783207463e-07,
"logits/chosen": -1.4453125,
"logits/rejected": -1.421875,
"logps/chosen": -188.0,
"logps/rejected": -209.0,
"loss": 0.4577,
"rewards/accuracies": 0.7749999761581421,
"rewards/chosen": -1.4765625,
"rewards/margins": 0.9296875,
"rewards/rejected": -2.40625,
"step": 520
},
{
"epoch": 0.6354916067146283,
"grad_norm": 16.86391324984424,
"learning_rate": 4.380275433140826e-07,
"logits/chosen": -1.421875,
"logits/rejected": -1.421875,
"logps/chosen": -189.0,
"logps/rejected": -201.0,
"loss": 0.4514,
"rewards/accuracies": 0.8374999761581421,
"rewards/chosen": -1.4609375,
"rewards/margins": 0.93359375,
"rewards/rejected": -2.390625,
"step": 530
},
{
"epoch": 0.6474820143884892,
"grad_norm": 21.115226866999087,
"learning_rate": 4.358063083074189e-07,
"logits/chosen": -1.4375,
"logits/rejected": -1.3984375,
"logps/chosen": -203.0,
"logps/rejected": -216.0,
"loss": 0.4654,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": -2.03125,
"rewards/margins": 1.0234375,
"rewards/rejected": -3.0625,
"step": 540
},
{
"epoch": 0.6594724220623501,
"grad_norm": 20.504857524123096,
"learning_rate": 4.335850733007552e-07,
"logits/chosen": -1.3984375,
"logits/rejected": -1.421875,
"logps/chosen": -196.0,
"logps/rejected": -207.0,
"loss": 0.4951,
"rewards/accuracies": 0.699999988079071,
"rewards/chosen": -2.4375,
"rewards/margins": 0.67578125,
"rewards/rejected": -3.125,
"step": 550
},
{
"epoch": 0.6714628297362111,
"grad_norm": 12.606241771319064,
"learning_rate": 4.313638382940915e-07,
"logits/chosen": -1.4296875,
"logits/rejected": -1.4296875,
"logps/chosen": -193.0,
"logps/rejected": -211.0,
"loss": 0.5071,
"rewards/accuracies": 0.7749999761581421,
"rewards/chosen": -1.8125,
"rewards/margins": 0.8671875,
"rewards/rejected": -2.671875,
"step": 560
},
{
"epoch": 0.6834532374100719,
"grad_norm": 21.06544638125072,
"learning_rate": 4.291426032874278e-07,
"logits/chosen": -1.421875,
"logits/rejected": -1.4296875,
"logps/chosen": -191.0,
"logps/rejected": -228.0,
"loss": 0.466,
"rewards/accuracies": 0.75,
"rewards/chosen": -1.703125,
"rewards/margins": 0.8984375,
"rewards/rejected": -2.59375,
"step": 570
},
{
"epoch": 0.6954436450839329,
"grad_norm": 20.105608002851536,
"learning_rate": 4.269213682807641e-07,
"logits/chosen": -1.4453125,
"logits/rejected": -1.4296875,
"logps/chosen": -197.0,
"logps/rejected": -193.0,
"loss": 0.461,
"rewards/accuracies": 0.7124999761581421,
"rewards/chosen": -1.2578125,
"rewards/margins": 0.87109375,
"rewards/rejected": -2.125,
"step": 580
},
{
"epoch": 0.7074340527577938,
"grad_norm": 19.41006530799305,
"learning_rate": 4.247001332741004e-07,
"logits/chosen": -1.453125,
"logits/rejected": -1.484375,
"logps/chosen": -190.0,
"logps/rejected": -208.0,
"loss": 0.4635,
"rewards/accuracies": 0.7875000238418579,
"rewards/chosen": -1.1015625,
"rewards/margins": 0.8828125,
"rewards/rejected": -1.984375,
"step": 590
},
{
"epoch": 0.7194244604316546,
"grad_norm": 19.775747633899154,
"learning_rate": 4.2247889826743666e-07,
"logits/chosen": -1.453125,
"logits/rejected": -1.4609375,
"logps/chosen": -196.0,
"logps/rejected": -202.0,
"loss": 0.532,
"rewards/accuracies": 0.824999988079071,
"rewards/chosen": -1.125,
"rewards/margins": 0.74609375,
"rewards/rejected": -1.8671875,
"step": 600
},
{
"epoch": 0.7314148681055156,
"grad_norm": 20.975128500826628,
"learning_rate": 4.2025766326077294e-07,
"logits/chosen": -1.4375,
"logits/rejected": -1.4453125,
"logps/chosen": -191.0,
"logps/rejected": -200.0,
"loss": 0.4693,
"rewards/accuracies": 0.737500011920929,
"rewards/chosen": -1.6171875,
"rewards/margins": 0.82421875,
"rewards/rejected": -2.4375,
"step": 610
},
{
"epoch": 0.7434052757793765,
"grad_norm": 19.86967099333255,
"learning_rate": 4.1803642825410926e-07,
"logits/chosen": -1.46875,
"logits/rejected": -1.4296875,
"logps/chosen": -187.0,
"logps/rejected": -208.0,
"loss": 0.4997,
"rewards/accuracies": 0.7875000238418579,
"rewards/chosen": -1.34375,
"rewards/margins": 0.77734375,
"rewards/rejected": -2.125,
"step": 620
},
{
"epoch": 0.7553956834532374,
"grad_norm": 18.483183442280176,
"learning_rate": 4.158151932474456e-07,
"logits/chosen": -1.4609375,
"logits/rejected": -1.4609375,
"logps/chosen": -200.0,
"logps/rejected": -199.0,
"loss": 0.4335,
"rewards/accuracies": 0.8500000238418579,
"rewards/chosen": -1.453125,
"rewards/margins": 1.0546875,
"rewards/rejected": -2.5,
"step": 630
},
{
"epoch": 0.7673860911270983,
"grad_norm": 27.16526184426748,
"learning_rate": 4.1359395824078187e-07,
"logits/chosen": -1.4375,
"logits/rejected": -1.4453125,
"logps/chosen": -199.0,
"logps/rejected": -222.0,
"loss": 0.4636,
"rewards/accuracies": 0.7749999761581421,
"rewards/chosen": -2.15625,
"rewards/margins": 0.92578125,
"rewards/rejected": -3.078125,
"step": 640
},
{
"epoch": 0.7793764988009593,
"grad_norm": 15.242340480747277,
"learning_rate": 4.1137272323411815e-07,
"logits/chosen": -1.484375,
"logits/rejected": -1.4765625,
"logps/chosen": -197.0,
"logps/rejected": -216.0,
"loss": 0.4529,
"rewards/accuracies": 0.7875000238418579,
"rewards/chosen": -2.578125,
"rewards/margins": 0.9296875,
"rewards/rejected": -3.515625,
"step": 650
},
{
"epoch": 0.7913669064748201,
"grad_norm": 19.772842905766947,
"learning_rate": 4.091514882274544e-07,
"logits/chosen": -1.453125,
"logits/rejected": -1.4296875,
"logps/chosen": -195.0,
"logps/rejected": -194.0,
"loss": 0.5111,
"rewards/accuracies": 0.6625000238418579,
"rewards/chosen": -1.8828125,
"rewards/margins": 0.59375,
"rewards/rejected": -2.46875,
"step": 660
},
{
"epoch": 0.8033573141486811,
"grad_norm": 20.04801706897612,
"learning_rate": 4.069302532207907e-07,
"logits/chosen": -1.46875,
"logits/rejected": -1.453125,
"logps/chosen": -205.0,
"logps/rejected": -205.0,
"loss": 0.4599,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": -1.4765625,
"rewards/margins": 1.1484375,
"rewards/rejected": -2.625,
"step": 670
},
{
"epoch": 0.815347721822542,
"grad_norm": 16.63035578458501,
"learning_rate": 4.047090182141271e-07,
"logits/chosen": -1.4375,
"logits/rejected": -1.3984375,
"logps/chosen": -194.0,
"logps/rejected": -208.0,
"loss": 0.4646,
"rewards/accuracies": 0.7250000238418579,
"rewards/chosen": -2.34375,
"rewards/margins": 0.65234375,
"rewards/rejected": -3.0,
"step": 680
},
{
"epoch": 0.8273381294964028,
"grad_norm": 16.867616740093794,
"learning_rate": 4.0248778320746336e-07,
"logits/chosen": -1.3984375,
"logits/rejected": -1.453125,
"logps/chosen": -194.0,
"logps/rejected": -213.0,
"loss": 0.4477,
"rewards/accuracies": 0.7749999761581421,
"rewards/chosen": -1.8515625,
"rewards/margins": 1.0078125,
"rewards/rejected": -2.859375,
"step": 690
},
{
"epoch": 0.8393285371702638,
"grad_norm": 15.441471505065069,
"learning_rate": 4.0026654820079964e-07,
"logits/chosen": -1.46875,
"logits/rejected": -1.4375,
"logps/chosen": -207.0,
"logps/rejected": -219.0,
"loss": 0.4527,
"rewards/accuracies": 0.8500000238418579,
"rewards/chosen": -1.6875,
"rewards/margins": 1.0234375,
"rewards/rejected": -2.71875,
"step": 700
},
{
"epoch": 0.8513189448441247,
"grad_norm": 17.143760908308213,
"learning_rate": 3.980453131941359e-07,
"logits/chosen": -1.4609375,
"logits/rejected": -1.4453125,
"logps/chosen": -197.0,
"logps/rejected": -209.0,
"loss": 0.4297,
"rewards/accuracies": 0.824999988079071,
"rewards/chosen": -1.546875,
"rewards/margins": 1.03125,
"rewards/rejected": -2.578125,
"step": 710
},
{
"epoch": 0.8633093525179856,
"grad_norm": 14.198886301598709,
"learning_rate": 3.958240781874722e-07,
"logits/chosen": -1.4765625,
"logits/rejected": -1.453125,
"logps/chosen": -198.0,
"logps/rejected": -198.0,
"loss": 0.4112,
"rewards/accuracies": 0.875,
"rewards/chosen": -1.640625,
"rewards/margins": 1.0546875,
"rewards/rejected": -2.6875,
"step": 720
},
{
"epoch": 0.8752997601918465,
"grad_norm": 19.10768891251934,
"learning_rate": 3.936028431808085e-07,
"logits/chosen": -1.4921875,
"logits/rejected": -1.421875,
"logps/chosen": -200.0,
"logps/rejected": -198.0,
"loss": 0.3902,
"rewards/accuracies": 0.875,
"rewards/chosen": -1.6640625,
"rewards/margins": 1.109375,
"rewards/rejected": -2.765625,
"step": 730
},
{
"epoch": 0.8872901678657075,
"grad_norm": 15.879596623648883,
"learning_rate": 3.913816081741448e-07,
"logits/chosen": -1.4140625,
"logits/rejected": -1.421875,
"logps/chosen": -186.0,
"logps/rejected": -204.0,
"loss": 0.3893,
"rewards/accuracies": 0.887499988079071,
"rewards/chosen": -1.796875,
"rewards/margins": 1.375,
"rewards/rejected": -3.171875,
"step": 740
},
{
"epoch": 0.8992805755395683,
"grad_norm": 17.436078043677007,
"learning_rate": 3.8916037316748113e-07,
"logits/chosen": -1.421875,
"logits/rejected": -1.421875,
"logps/chosen": -193.0,
"logps/rejected": -200.0,
"loss": 0.4263,
"rewards/accuracies": 0.75,
"rewards/chosen": -2.3125,
"rewards/margins": 0.91796875,
"rewards/rejected": -3.234375,
"step": 750
},
{
"epoch": 0.9112709832134293,
"grad_norm": 15.704331891226987,
"learning_rate": 3.869391381608174e-07,
"logits/chosen": -1.4375,
"logits/rejected": -1.453125,
"logps/chosen": -201.0,
"logps/rejected": -211.0,
"loss": 0.3875,
"rewards/accuracies": 0.8125,
"rewards/chosen": -1.9609375,
"rewards/margins": 1.015625,
"rewards/rejected": -2.984375,
"step": 760
},
{
"epoch": 0.9232613908872902,
"grad_norm": 19.28806504511243,
"learning_rate": 3.847179031541537e-07,
"logits/chosen": -1.4921875,
"logits/rejected": -1.4765625,
"logps/chosen": -191.0,
"logps/rejected": -190.0,
"loss": 0.421,
"rewards/accuracies": 0.862500011920929,
"rewards/chosen": -1.453125,
"rewards/margins": 1.046875,
"rewards/rejected": -2.5,
"step": 770
},
{
"epoch": 0.935251798561151,
"grad_norm": 17.706917404985195,
"learning_rate": 3.8249666814749e-07,
"logits/chosen": -1.4375,
"logits/rejected": -1.40625,
"logps/chosen": -205.0,
"logps/rejected": -201.0,
"loss": 0.3674,
"rewards/accuracies": 0.875,
"rewards/chosen": -1.78125,
"rewards/margins": 1.25,
"rewards/rejected": -3.03125,
"step": 780
},
{
"epoch": 0.947242206235012,
"grad_norm": 14.788710817068432,
"learning_rate": 3.802754331408263e-07,
"logits/chosen": -1.4453125,
"logits/rejected": -1.4453125,
"logps/chosen": -198.0,
"logps/rejected": -222.0,
"loss": 0.4113,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": -2.171875,
"rewards/margins": 1.078125,
"rewards/rejected": -3.25,
"step": 790
},
{
"epoch": 0.9592326139088729,
"grad_norm": 20.22248928411642,
"learning_rate": 3.7805419813416256e-07,
"logits/chosen": -1.4921875,
"logits/rejected": -1.4375,
"logps/chosen": -192.0,
"logps/rejected": -204.0,
"loss": 0.3646,
"rewards/accuracies": 0.824999988079071,
"rewards/chosen": -2.3125,
"rewards/margins": 1.15625,
"rewards/rejected": -3.484375,
"step": 800
},
{
"epoch": 0.9712230215827338,
"grad_norm": 13.171309874145647,
"learning_rate": 3.7583296312749884e-07,
"logits/chosen": -1.3984375,
"logits/rejected": -1.4296875,
"logps/chosen": -198.0,
"logps/rejected": -215.0,
"loss": 0.4278,
"rewards/accuracies": 0.862500011920929,
"rewards/chosen": -2.234375,
"rewards/margins": 1.34375,
"rewards/rejected": -3.578125,
"step": 810
},
{
"epoch": 0.9832134292565947,
"grad_norm": 19.50680457515295,
"learning_rate": 3.7361172812083517e-07,
"logits/chosen": -1.46875,
"logits/rejected": -1.4140625,
"logps/chosen": -195.0,
"logps/rejected": -196.0,
"loss": 0.3729,
"rewards/accuracies": 0.8500000238418579,
"rewards/chosen": -1.640625,
"rewards/margins": 1.1953125,
"rewards/rejected": -2.828125,
"step": 820
},
{
"epoch": 0.9952038369304557,
"grad_norm": 19.43880275481757,
"learning_rate": 3.713904931141715e-07,
"logits/chosen": -1.5,
"logits/rejected": -1.4765625,
"logps/chosen": -202.0,
"logps/rejected": -202.0,
"loss": 0.3305,
"rewards/accuracies": 0.9125000238418579,
"rewards/chosen": -1.3203125,
"rewards/margins": 1.2890625,
"rewards/rejected": -2.609375,
"step": 830
},
{
"epoch": 1.0,
"eval_logits/chosen": -1.4765625,
"eval_logits/rejected": -1.484375,
"eval_logps/chosen": -190.0,
"eval_logps/rejected": -195.0,
"eval_loss": 0.6123461127281189,
"eval_rewards/accuracies": 0.6029411554336548,
"eval_rewards/chosen": -1.2265625,
"eval_rewards/margins": 0.63671875,
"eval_rewards/rejected": -1.8671875,
"eval_runtime": 23.6977,
"eval_samples_per_second": 16.921,
"eval_steps_per_second": 0.717,
"step": 834
},
{
"epoch": 1.0071942446043165,
"grad_norm": 17.925966726812263,
"learning_rate": 3.691692581075078e-07,
"logits/chosen": -1.484375,
"logits/rejected": -1.5078125,
"logps/chosen": -198.0,
"logps/rejected": -228.0,
"loss": 0.3292,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -1.28125,
"rewards/margins": 1.6328125,
"rewards/rejected": -2.921875,
"step": 840
},
{
"epoch": 1.0191846522781776,
"grad_norm": 15.13245137359342,
"learning_rate": 3.6694802310084405e-07,
"logits/chosen": -1.4921875,
"logits/rejected": -1.5234375,
"logps/chosen": -191.0,
"logps/rejected": -209.0,
"loss": 0.2941,
"rewards/accuracies": 0.887499988079071,
"rewards/chosen": -1.390625,
"rewards/margins": 1.390625,
"rewards/rejected": -2.78125,
"step": 850
},
{
"epoch": 1.0311750599520384,
"grad_norm": 15.85613126792236,
"learning_rate": 3.6472678809418033e-07,
"logits/chosen": -1.453125,
"logits/rejected": -1.453125,
"logps/chosen": -197.0,
"logps/rejected": -200.0,
"loss": 0.3029,
"rewards/accuracies": 0.887499988079071,
"rewards/chosen": -1.28125,
"rewards/margins": 1.65625,
"rewards/rejected": -2.9375,
"step": 860
},
{
"epoch": 1.0431654676258992,
"grad_norm": 17.934043297524138,
"learning_rate": 3.625055530875166e-07,
"logits/chosen": -1.484375,
"logits/rejected": -1.4765625,
"logps/chosen": -201.0,
"logps/rejected": -217.0,
"loss": 0.2741,
"rewards/accuracies": 0.9125000238418579,
"rewards/chosen": -1.59375,
"rewards/margins": 1.828125,
"rewards/rejected": -3.421875,
"step": 870
},
{
"epoch": 1.0551558752997603,
"grad_norm": 12.6431049483951,
"learning_rate": 3.6028431808085294e-07,
"logits/chosen": -1.5078125,
"logits/rejected": -1.46875,
"logps/chosen": -203.0,
"logps/rejected": -221.0,
"loss": 0.2871,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -1.7890625,
"rewards/margins": 1.5859375,
"rewards/rejected": -3.375,
"step": 880
},
{
"epoch": 1.0671462829736211,
"grad_norm": 14.781119124887237,
"learning_rate": 3.5806308307418926e-07,
"logits/chosen": -1.5078125,
"logits/rejected": -1.5078125,
"logps/chosen": -200.0,
"logps/rejected": -222.0,
"loss": 0.2536,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -1.90625,
"rewards/margins": 1.828125,
"rewards/rejected": -3.71875,
"step": 890
},
{
"epoch": 1.079136690647482,
"grad_norm": 13.675786724224842,
"learning_rate": 3.5584184806752554e-07,
"logits/chosen": -1.46875,
"logits/rejected": -1.453125,
"logps/chosen": -191.0,
"logps/rejected": -216.0,
"loss": 0.2305,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -1.84375,
"rewards/margins": 2.03125,
"rewards/rejected": -3.875,
"step": 900
},
{
"epoch": 1.091127098321343,
"grad_norm": 15.133495835048825,
"learning_rate": 3.536206130608618e-07,
"logits/chosen": -1.4609375,
"logits/rejected": -1.484375,
"logps/chosen": -191.0,
"logps/rejected": -208.0,
"loss": 0.2446,
"rewards/accuracies": 0.9125000238418579,
"rewards/chosen": -1.7734375,
"rewards/margins": 2.015625,
"rewards/rejected": -3.796875,
"step": 910
},
{
"epoch": 1.1031175059952039,
"grad_norm": 17.223627578304455,
"learning_rate": 3.513993780541981e-07,
"logits/chosen": -1.453125,
"logits/rejected": -1.4921875,
"logps/chosen": -172.0,
"logps/rejected": -210.0,
"loss": 0.2433,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -1.796875,
"rewards/margins": 1.9765625,
"rewards/rejected": -3.765625,
"step": 920
},
{
"epoch": 1.1151079136690647,
"grad_norm": 12.606788377121838,
"learning_rate": 3.491781430475344e-07,
"logits/chosen": -1.5390625,
"logits/rejected": -1.5546875,
"logps/chosen": -192.0,
"logps/rejected": -231.0,
"loss": 0.2498,
"rewards/accuracies": 0.887499988079071,
"rewards/chosen": -1.78125,
"rewards/margins": 1.796875,
"rewards/rejected": -3.578125,
"step": 930
},
{
"epoch": 1.1270983213429258,
"grad_norm": 20.113412627019386,
"learning_rate": 3.469569080408707e-07,
"logits/chosen": -1.4453125,
"logits/rejected": -1.4375,
"logps/chosen": -185.0,
"logps/rejected": -204.0,
"loss": 0.258,
"rewards/accuracies": 0.8999999761581421,
"rewards/chosen": -1.5234375,
"rewards/margins": 2.125,
"rewards/rejected": -3.640625,
"step": 940
},
{
"epoch": 1.1390887290167866,
"grad_norm": 22.896453855913,
"learning_rate": 3.4473567303420703e-07,
"logits/chosen": -1.53125,
"logits/rejected": -1.515625,
"logps/chosen": -200.0,
"logps/rejected": -217.0,
"loss": 0.2356,
"rewards/accuracies": 0.9624999761581421,
"rewards/chosen": -2.0,
"rewards/margins": 2.09375,
"rewards/rejected": -4.09375,
"step": 950
},
{
"epoch": 1.1510791366906474,
"grad_norm": 17.597413512159072,
"learning_rate": 3.425144380275433e-07,
"logits/chosen": -1.484375,
"logits/rejected": -1.53125,
"logps/chosen": -213.0,
"logps/rejected": -235.0,
"loss": 0.2713,
"rewards/accuracies": 0.925000011920929,
"rewards/chosen": -2.359375,
"rewards/margins": 1.8984375,
"rewards/rejected": -4.25,
"step": 960
},
{
"epoch": 1.1630695443645085,
"grad_norm": 14.323268951448473,
"learning_rate": 3.402932030208796e-07,
"logits/chosen": -1.515625,
"logits/rejected": -1.5,
"logps/chosen": -210.0,
"logps/rejected": -235.0,
"loss": 0.2388,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -2.40625,
"rewards/margins": 2.171875,
"rewards/rejected": -4.5625,
"step": 970
},
{
"epoch": 1.1750599520383693,
"grad_norm": 13.502480458522491,
"learning_rate": 3.380719680142159e-07,
"logits/chosen": -1.515625,
"logits/rejected": -1.546875,
"logps/chosen": -192.0,
"logps/rejected": -209.0,
"loss": 0.2241,
"rewards/accuracies": 0.9125000238418579,
"rewards/chosen": -1.6953125,
"rewards/margins": 2.078125,
"rewards/rejected": -3.765625,
"step": 980
},
{
"epoch": 1.1870503597122302,
"grad_norm": 12.261596643016196,
"learning_rate": 3.358507330075522e-07,
"logits/chosen": -1.5546875,
"logits/rejected": -1.5390625,
"logps/chosen": -199.0,
"logps/rejected": -219.0,
"loss": 0.2554,
"rewards/accuracies": 0.925000011920929,
"rewards/chosen": -2.140625,
"rewards/margins": 1.953125,
"rewards/rejected": -4.09375,
"step": 990
},
{
"epoch": 1.1990407673860912,
"grad_norm": 11.618624189476918,
"learning_rate": 3.3362949800088847e-07,
"logits/chosen": -1.5078125,
"logits/rejected": -1.4921875,
"logps/chosen": -204.0,
"logps/rejected": -222.0,
"loss": 0.2051,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -1.71875,
"rewards/margins": 2.171875,
"rewards/rejected": -3.890625,
"step": 1000
},
{
"epoch": 1.211031175059952,
"grad_norm": 22.814769403029445,
"learning_rate": 3.3140826299422474e-07,
"logits/chosen": -1.546875,
"logits/rejected": -1.53125,
"logps/chosen": -201.0,
"logps/rejected": -229.0,
"loss": 0.2417,
"rewards/accuracies": 0.9375,
"rewards/chosen": -2.015625,
"rewards/margins": 2.25,
"rewards/rejected": -4.28125,
"step": 1010
},
{
"epoch": 1.223021582733813,
"grad_norm": 23.105005233123116,
"learning_rate": 3.291870279875611e-07,
"logits/chosen": -1.546875,
"logits/rejected": -1.578125,
"logps/chosen": -209.0,
"logps/rejected": -235.0,
"loss": 0.2123,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -2.4375,
"rewards/margins": 2.21875,
"rewards/rejected": -4.65625,
"step": 1020
},
{
"epoch": 1.235011990407674,
"grad_norm": 16.59163456560542,
"learning_rate": 3.269657929808974e-07,
"logits/chosen": -1.546875,
"logits/rejected": -1.515625,
"logps/chosen": -206.0,
"logps/rejected": -229.0,
"loss": 0.2316,
"rewards/accuracies": 0.925000011920929,
"rewards/chosen": -2.3125,
"rewards/margins": 1.953125,
"rewards/rejected": -4.28125,
"step": 1030
},
{
"epoch": 1.2470023980815348,
"grad_norm": 16.97136930640841,
"learning_rate": 3.247445579742337e-07,
"logits/chosen": -1.4921875,
"logits/rejected": -1.4609375,
"logps/chosen": -196.0,
"logps/rejected": -228.0,
"loss": 0.233,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -2.640625,
"rewards/margins": 2.078125,
"rewards/rejected": -4.71875,
"step": 1040
},
{
"epoch": 1.2589928057553956,
"grad_norm": 11.658846274024977,
"learning_rate": 3.2252332296756996e-07,
"logits/chosen": -1.484375,
"logits/rejected": -1.5546875,
"logps/chosen": -191.0,
"logps/rejected": -217.0,
"loss": 0.2061,
"rewards/accuracies": 0.925000011920929,
"rewards/chosen": -2.078125,
"rewards/margins": 2.265625,
"rewards/rejected": -4.34375,
"step": 1050
},
{
"epoch": 1.2709832134292567,
"grad_norm": 17.907683758231656,
"learning_rate": 3.2030208796090623e-07,
"logits/chosen": -1.5546875,
"logits/rejected": -1.515625,
"logps/chosen": -199.0,
"logps/rejected": -218.0,
"loss": 0.2564,
"rewards/accuracies": 0.9375,
"rewards/chosen": -2.125,
"rewards/margins": 2.46875,
"rewards/rejected": -4.59375,
"step": 1060
},
{
"epoch": 1.2829736211031175,
"grad_norm": 13.891330543214426,
"learning_rate": 3.180808529542425e-07,
"logits/chosen": -1.578125,
"logits/rejected": -1.53125,
"logps/chosen": -220.0,
"logps/rejected": -236.0,
"loss": 0.2253,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -2.5625,
"rewards/margins": 2.078125,
"rewards/rejected": -4.625,
"step": 1070
},
{
"epoch": 1.2949640287769784,
"grad_norm": 12.255143244628048,
"learning_rate": 3.1585961794757884e-07,
"logits/chosen": -1.5703125,
"logits/rejected": -1.5546875,
"logps/chosen": -213.0,
"logps/rejected": -220.0,
"loss": 0.1986,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -2.15625,
"rewards/margins": 2.25,
"rewards/rejected": -4.40625,
"step": 1080
},
{
"epoch": 1.3069544364508392,
"grad_norm": 16.986661579673758,
"learning_rate": 3.1363838294091517e-07,
"logits/chosen": -1.5546875,
"logits/rejected": -1.53125,
"logps/chosen": -201.0,
"logps/rejected": -215.0,
"loss": 0.1726,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -1.5078125,
"rewards/margins": 2.515625,
"rewards/rejected": -4.0,
"step": 1090
},
{
"epoch": 1.3189448441247003,
"grad_norm": 11.362437294228451,
"learning_rate": 3.1141714793425145e-07,
"logits/chosen": -1.5546875,
"logits/rejected": -1.5546875,
"logps/chosen": -217.0,
"logps/rejected": -240.0,
"loss": 0.1936,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -2.265625,
"rewards/margins": 2.34375,
"rewards/rejected": -4.625,
"step": 1100
},
{
"epoch": 1.330935251798561,
"grad_norm": 18.06174034251652,
"learning_rate": 3.091959129275877e-07,
"logits/chosen": -1.5,
"logits/rejected": -1.4921875,
"logps/chosen": -184.0,
"logps/rejected": -205.0,
"loss": 0.2132,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -1.78125,
"rewards/margins": 2.40625,
"rewards/rejected": -4.1875,
"step": 1110
},
{
"epoch": 1.3429256594724222,
"grad_norm": 21.343977198752015,
"learning_rate": 3.06974677920924e-07,
"logits/chosen": -1.5625,
"logits/rejected": -1.5078125,
"logps/chosen": -221.0,
"logps/rejected": -240.0,
"loss": 0.2055,
"rewards/accuracies": 0.9125000238418579,
"rewards/chosen": -2.90625,
"rewards/margins": 2.171875,
"rewards/rejected": -5.09375,
"step": 1120
},
{
"epoch": 1.354916067146283,
"grad_norm": 19.364261897999658,
"learning_rate": 3.0475344291426033e-07,
"logits/chosen": -1.546875,
"logits/rejected": -1.515625,
"logps/chosen": -196.0,
"logps/rejected": -230.0,
"loss": 0.1925,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -2.140625,
"rewards/margins": 2.90625,
"rewards/rejected": -5.03125,
"step": 1130
},
{
"epoch": 1.3669064748201438,
"grad_norm": 12.49182324344103,
"learning_rate": 3.025322079075966e-07,
"logits/chosen": -1.484375,
"logits/rejected": -1.484375,
"logps/chosen": -197.0,
"logps/rejected": -239.0,
"loss": 0.2206,
"rewards/accuracies": 0.9375,
"rewards/chosen": -2.453125,
"rewards/margins": 2.65625,
"rewards/rejected": -5.09375,
"step": 1140
},
{
"epoch": 1.3788968824940047,
"grad_norm": 12.604128787864076,
"learning_rate": 3.003109729009329e-07,
"logits/chosen": -1.5078125,
"logits/rejected": -1.5234375,
"logps/chosen": -197.0,
"logps/rejected": -235.0,
"loss": 0.2102,
"rewards/accuracies": 0.9375,
"rewards/chosen": -2.671875,
"rewards/margins": 2.1875,
"rewards/rejected": -4.875,
"step": 1150
},
{
"epoch": 1.3908872901678657,
"grad_norm": 16.0572765821895,
"learning_rate": 2.980897378942692e-07,
"logits/chosen": -1.5546875,
"logits/rejected": -1.546875,
"logps/chosen": -207.0,
"logps/rejected": -226.0,
"loss": 0.1899,
"rewards/accuracies": 0.925000011920929,
"rewards/chosen": -2.6875,
"rewards/margins": 2.171875,
"rewards/rejected": -4.84375,
"step": 1160
},
{
"epoch": 1.4028776978417266,
"grad_norm": 14.62570261847443,
"learning_rate": 2.958685028876055e-07,
"logits/chosen": -1.5703125,
"logits/rejected": -1.546875,
"logps/chosen": -204.0,
"logps/rejected": -223.0,
"loss": 0.2256,
"rewards/accuracies": 0.9125000238418579,
"rewards/chosen": -2.1875,
"rewards/margins": 2.28125,
"rewards/rejected": -4.46875,
"step": 1170
},
{
"epoch": 1.4148681055155876,
"grad_norm": 16.69089195431221,
"learning_rate": 2.936472678809418e-07,
"logits/chosen": -1.5703125,
"logits/rejected": -1.5703125,
"logps/chosen": -215.0,
"logps/rejected": -234.0,
"loss": 0.2164,
"rewards/accuracies": 0.925000011920929,
"rewards/chosen": -2.625,
"rewards/margins": 2.265625,
"rewards/rejected": -4.90625,
"step": 1180
},
{
"epoch": 1.4268585131894485,
"grad_norm": 14.208012618540776,
"learning_rate": 2.914260328742781e-07,
"logits/chosen": -1.5625,
"logits/rejected": -1.546875,
"logps/chosen": -206.0,
"logps/rejected": -226.0,
"loss": 0.1728,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -2.90625,
"rewards/margins": 2.21875,
"rewards/rejected": -5.125,
"step": 1190
},
{
"epoch": 1.4388489208633093,
"grad_norm": 19.814051144030756,
"learning_rate": 2.8920479786761437e-07,
"logits/chosen": -1.59375,
"logits/rejected": -1.59375,
"logps/chosen": -206.0,
"logps/rejected": -232.0,
"loss": 0.2096,
"rewards/accuracies": 0.9624999761581421,
"rewards/chosen": -2.484375,
"rewards/margins": 2.40625,
"rewards/rejected": -4.875,
"step": 1200
},
{
"epoch": 1.4508393285371701,
"grad_norm": 16.967462896165113,
"learning_rate": 2.8698356286095065e-07,
"logits/chosen": -1.59375,
"logits/rejected": -1.5703125,
"logps/chosen": -220.0,
"logps/rejected": -239.0,
"loss": 0.1481,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -2.5625,
"rewards/margins": 2.796875,
"rewards/rejected": -5.34375,
"step": 1210
},
{
"epoch": 1.4628297362110312,
"grad_norm": 15.206385611452236,
"learning_rate": 2.847623278542869e-07,
"logits/chosen": -1.609375,
"logits/rejected": -1.609375,
"logps/chosen": -208.0,
"logps/rejected": -229.0,
"loss": 0.1846,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -2.5625,
"rewards/margins": 2.375,
"rewards/rejected": -4.9375,
"step": 1220
},
{
"epoch": 1.474820143884892,
"grad_norm": 17.869462519073895,
"learning_rate": 2.825410928476233e-07,
"logits/chosen": -1.515625,
"logits/rejected": -1.53125,
"logps/chosen": -187.0,
"logps/rejected": -229.0,
"loss": 0.1821,
"rewards/accuracies": 0.9375,
"rewards/chosen": -2.703125,
"rewards/margins": 2.578125,
"rewards/rejected": -5.28125,
"step": 1230
},
{
"epoch": 1.486810551558753,
"grad_norm": 11.533712996176597,
"learning_rate": 2.803198578409596e-07,
"logits/chosen": -1.609375,
"logits/rejected": -1.5625,
"logps/chosen": -204.0,
"logps/rejected": -228.0,
"loss": 0.2133,
"rewards/accuracies": 0.925000011920929,
"rewards/chosen": -2.703125,
"rewards/margins": 2.515625,
"rewards/rejected": -5.21875,
"step": 1240
},
{
"epoch": 1.498800959232614,
"grad_norm": 11.144943270392623,
"learning_rate": 2.7809862283429586e-07,
"logits/chosen": -1.5859375,
"logits/rejected": -1.5703125,
"logps/chosen": -193.0,
"logps/rejected": -239.0,
"loss": 0.1494,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -2.3125,
"rewards/margins": 3.078125,
"rewards/rejected": -5.375,
"step": 1250
},
{
"epoch": 1.5107913669064748,
"grad_norm": 15.93725977415015,
"learning_rate": 2.7587738782763214e-07,
"logits/chosen": -1.5625,
"logits/rejected": -1.59375,
"logps/chosen": -196.0,
"logps/rejected": -236.0,
"loss": 0.204,
"rewards/accuracies": 0.8999999761581421,
"rewards/chosen": -3.53125,
"rewards/margins": 2.546875,
"rewards/rejected": -6.0625,
"step": 1260
},
{
"epoch": 1.5227817745803356,
"grad_norm": 23.59190890231536,
"learning_rate": 2.736561528209684e-07,
"logits/chosen": -1.640625,
"logits/rejected": -1.640625,
"logps/chosen": -210.0,
"logps/rejected": -244.0,
"loss": 0.1853,
"rewards/accuracies": 0.9624999761581421,
"rewards/chosen": -3.09375,
"rewards/margins": 2.625,
"rewards/rejected": -5.71875,
"step": 1270
},
{
"epoch": 1.5347721822541966,
"grad_norm": 22.364117581141223,
"learning_rate": 2.7143491781430474e-07,
"logits/chosen": -1.5078125,
"logits/rejected": -1.5390625,
"logps/chosen": -189.0,
"logps/rejected": -229.0,
"loss": 0.1952,
"rewards/accuracies": 0.9375,
"rewards/chosen": -2.78125,
"rewards/margins": 2.296875,
"rewards/rejected": -5.09375,
"step": 1280
},
{
"epoch": 1.5467625899280577,
"grad_norm": 11.715030229741336,
"learning_rate": 2.692136828076411e-07,
"logits/chosen": -1.578125,
"logits/rejected": -1.5546875,
"logps/chosen": -210.0,
"logps/rejected": -234.0,
"loss": 0.1717,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -2.96875,
"rewards/margins": 2.609375,
"rewards/rejected": -5.5625,
"step": 1290
},
{
"epoch": 1.5587529976019185,
"grad_norm": 16.74676425409118,
"learning_rate": 2.6699244780097735e-07,
"logits/chosen": -1.609375,
"logits/rejected": -1.59375,
"logps/chosen": -206.0,
"logps/rejected": -231.0,
"loss": 0.1883,
"rewards/accuracies": 0.9624999761581421,
"rewards/chosen": -2.875,
"rewards/margins": 2.515625,
"rewards/rejected": -5.375,
"step": 1300
},
{
"epoch": 1.5707434052757794,
"grad_norm": 16.31256275621072,
"learning_rate": 2.6477121279431363e-07,
"logits/chosen": -1.59375,
"logits/rejected": -1.5859375,
"logps/chosen": -194.0,
"logps/rejected": -226.0,
"loss": 0.2258,
"rewards/accuracies": 0.9375,
"rewards/chosen": -2.5,
"rewards/margins": 2.4375,
"rewards/rejected": -4.9375,
"step": 1310
},
{
"epoch": 1.5827338129496402,
"grad_norm": 15.174270462866287,
"learning_rate": 2.625499777876499e-07,
"logits/chosen": -1.59375,
"logits/rejected": -1.5703125,
"logps/chosen": -204.0,
"logps/rejected": -229.0,
"loss": 0.1577,
"rewards/accuracies": 0.925000011920929,
"rewards/chosen": -3.375,
"rewards/margins": 2.421875,
"rewards/rejected": -5.78125,
"step": 1320
},
{
"epoch": 1.594724220623501,
"grad_norm": 20.949271053556263,
"learning_rate": 2.6032874278098623e-07,
"logits/chosen": -1.625,
"logits/rejected": -1.53125,
"logps/chosen": -218.0,
"logps/rejected": -234.0,
"loss": 0.1542,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -2.46875,
"rewards/margins": 3.078125,
"rewards/rejected": -5.5625,
"step": 1330
},
{
"epoch": 1.6067146282973621,
"grad_norm": 21.582166006491406,
"learning_rate": 2.581075077743225e-07,
"logits/chosen": -1.5625,
"logits/rejected": -1.484375,
"logps/chosen": -215.0,
"logps/rejected": -231.0,
"loss": 0.1642,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -3.0625,
"rewards/margins": 2.703125,
"rewards/rejected": -5.75,
"step": 1340
},
{
"epoch": 1.6187050359712232,
"grad_norm": 12.490052192808845,
"learning_rate": 2.558862727676588e-07,
"logits/chosen": -1.625,
"logits/rejected": -1.546875,
"logps/chosen": -211.0,
"logps/rejected": -227.0,
"loss": 0.1549,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.71875,
"rewards/margins": 2.921875,
"rewards/rejected": -5.625,
"step": 1350
},
{
"epoch": 1.630695443645084,
"grad_norm": 14.27659020234575,
"learning_rate": 2.536650377609951e-07,
"logits/chosen": -1.6484375,
"logits/rejected": -1.625,
"logps/chosen": -208.0,
"logps/rejected": -230.0,
"loss": 0.1123,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -2.515625,
"rewards/margins": 3.0625,
"rewards/rejected": -5.5625,
"step": 1360
},
{
"epoch": 1.6426858513189448,
"grad_norm": 14.433767508436462,
"learning_rate": 2.514438027543314e-07,
"logits/chosen": -1.6015625,
"logits/rejected": -1.5625,
"logps/chosen": -206.0,
"logps/rejected": -235.0,
"loss": 0.1312,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -2.359375,
"rewards/margins": 3.359375,
"rewards/rejected": -5.71875,
"step": 1370
},
{
"epoch": 1.6546762589928057,
"grad_norm": 13.424589730956683,
"learning_rate": 2.492225677476677e-07,
"logits/chosen": -1.609375,
"logits/rejected": -1.6171875,
"logps/chosen": -198.0,
"logps/rejected": -242.0,
"loss": 0.1744,
"rewards/accuracies": 0.925000011920929,
"rewards/chosen": -2.609375,
"rewards/margins": 2.71875,
"rewards/rejected": -5.34375,
"step": 1380
},
{
"epoch": 1.6666666666666665,
"grad_norm": 23.17900601362576,
"learning_rate": 2.47001332741004e-07,
"logits/chosen": -1.625,
"logits/rejected": -1.5703125,
"logps/chosen": -203.0,
"logps/rejected": -217.0,
"loss": 0.1801,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -2.109375,
"rewards/margins": 2.8125,
"rewards/rejected": -4.90625,
"step": 1390
},
{
"epoch": 1.6786570743405276,
"grad_norm": 15.035036612326222,
"learning_rate": 2.447800977343403e-07,
"logits/chosen": -1.5859375,
"logits/rejected": -1.5625,
"logps/chosen": -195.0,
"logps/rejected": -226.0,
"loss": 0.1659,
"rewards/accuracies": 0.9125000238418579,
"rewards/chosen": -2.78125,
"rewards/margins": 2.796875,
"rewards/rejected": -5.5625,
"step": 1400
},
{
"epoch": 1.6906474820143886,
"grad_norm": 23.256101786444912,
"learning_rate": 2.425588627276766e-07,
"logits/chosen": -1.59375,
"logits/rejected": -1.5859375,
"logps/chosen": -199.0,
"logps/rejected": -238.0,
"loss": 0.15,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -2.734375,
"rewards/margins": 3.140625,
"rewards/rejected": -5.875,
"step": 1410
},
{
"epoch": 1.7026378896882495,
"grad_norm": 16.44114210272047,
"learning_rate": 2.403376277210129e-07,
"logits/chosen": -1.5859375,
"logits/rejected": -1.578125,
"logps/chosen": -207.0,
"logps/rejected": -242.0,
"loss": 0.1413,
"rewards/accuracies": 0.925000011920929,
"rewards/chosen": -3.265625,
"rewards/margins": 2.890625,
"rewards/rejected": -6.125,
"step": 1420
},
{
"epoch": 1.7146282973621103,
"grad_norm": 19.104299358854444,
"learning_rate": 2.3811639271434916e-07,
"logits/chosen": -1.65625,
"logits/rejected": -1.5390625,
"logps/chosen": -216.0,
"logps/rejected": -240.0,
"loss": 0.1638,
"rewards/accuracies": 0.9375,
"rewards/chosen": -3.265625,
"rewards/margins": 2.875,
"rewards/rejected": -6.15625,
"step": 1430
},
{
"epoch": 1.7266187050359711,
"grad_norm": 8.57838400028329,
"learning_rate": 2.3589515770768546e-07,
"logits/chosen": -1.59375,
"logits/rejected": -1.5546875,
"logps/chosen": -221.0,
"logps/rejected": -236.0,
"loss": 0.1491,
"rewards/accuracies": 0.9375,
"rewards/chosen": -3.09375,
"rewards/margins": 2.875,
"rewards/rejected": -5.96875,
"step": 1440
},
{
"epoch": 1.738609112709832,
"grad_norm": 12.545882511992975,
"learning_rate": 2.3367392270102177e-07,
"logits/chosen": -1.6328125,
"logits/rejected": -1.6171875,
"logps/chosen": -200.0,
"logps/rejected": -252.0,
"loss": 0.1236,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -2.328125,
"rewards/margins": 3.359375,
"rewards/rejected": -5.6875,
"step": 1450
},
{
"epoch": 1.750599520383693,
"grad_norm": 16.11932503670354,
"learning_rate": 2.3145268769435804e-07,
"logits/chosen": -1.6328125,
"logits/rejected": -1.578125,
"logps/chosen": -214.0,
"logps/rejected": -242.0,
"loss": 0.1705,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -3.078125,
"rewards/margins": 3.125,
"rewards/rejected": -6.21875,
"step": 1460
},
{
"epoch": 1.762589928057554,
"grad_norm": 19.767212861451107,
"learning_rate": 2.2923145268769435e-07,
"logits/chosen": -1.65625,
"logits/rejected": -1.6171875,
"logps/chosen": -197.0,
"logps/rejected": -232.0,
"loss": 0.1243,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -2.96875,
"rewards/margins": 3.59375,
"rewards/rejected": -6.5625,
"step": 1470
},
{
"epoch": 1.774580335731415,
"grad_norm": 17.770676013426908,
"learning_rate": 2.2701021768103065e-07,
"logits/chosen": -1.609375,
"logits/rejected": -1.5859375,
"logps/chosen": -212.0,
"logps/rejected": -244.0,
"loss": 0.162,
"rewards/accuracies": 0.9125000238418579,
"rewards/chosen": -3.5,
"rewards/margins": 3.0625,
"rewards/rejected": -6.5625,
"step": 1480
},
{
"epoch": 1.7865707434052758,
"grad_norm": 9.354271003661482,
"learning_rate": 2.2478898267436695e-07,
"logits/chosen": -1.6328125,
"logits/rejected": -1.6328125,
"logps/chosen": -207.0,
"logps/rejected": -245.0,
"loss": 0.1463,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -3.421875,
"rewards/margins": 3.109375,
"rewards/rejected": -6.53125,
"step": 1490
},
{
"epoch": 1.7985611510791366,
"grad_norm": 16.23301041511246,
"learning_rate": 2.2256774766770323e-07,
"logits/chosen": -1.5859375,
"logits/rejected": -1.53125,
"logps/chosen": -205.0,
"logps/rejected": -218.0,
"loss": 0.147,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -3.046875,
"rewards/margins": 3.0625,
"rewards/rejected": -6.125,
"step": 1500
},
{
"epoch": 1.8105515587529974,
"grad_norm": 10.88674570833714,
"learning_rate": 2.2034651266103953e-07,
"logits/chosen": -1.59375,
"logits/rejected": -1.5546875,
"logps/chosen": -186.0,
"logps/rejected": -227.0,
"loss": 0.1292,
"rewards/accuracies": 0.9624999761581421,
"rewards/chosen": -2.734375,
"rewards/margins": 3.109375,
"rewards/rejected": -5.84375,
"step": 1510
},
{
"epoch": 1.8225419664268585,
"grad_norm": 11.649861361880111,
"learning_rate": 2.1812527765437583e-07,
"logits/chosen": -1.6328125,
"logits/rejected": -1.578125,
"logps/chosen": -211.0,
"logps/rejected": -235.0,
"loss": 0.1467,
"rewards/accuracies": 0.9624999761581421,
"rewards/chosen": -2.5,
"rewards/margins": 3.046875,
"rewards/rejected": -5.5625,
"step": 1520
},
{
"epoch": 1.8345323741007196,
"grad_norm": 17.611539910897193,
"learning_rate": 2.159040426477121e-07,
"logits/chosen": -1.640625,
"logits/rejected": -1.625,
"logps/chosen": -216.0,
"logps/rejected": -258.0,
"loss": 0.1337,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -3.546875,
"rewards/margins": 3.0625,
"rewards/rejected": -6.59375,
"step": 1530
},
{
"epoch": 1.8465227817745804,
"grad_norm": 20.32024320505495,
"learning_rate": 2.1368280764104841e-07,
"logits/chosen": -1.640625,
"logits/rejected": -1.6015625,
"logps/chosen": -210.0,
"logps/rejected": -262.0,
"loss": 0.1444,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -3.59375,
"rewards/margins": 3.421875,
"rewards/rejected": -7.0,
"step": 1540
},
{
"epoch": 1.8585131894484412,
"grad_norm": 10.61654419093803,
"learning_rate": 2.1146157263438472e-07,
"logits/chosen": -1.59375,
"logits/rejected": -1.609375,
"logps/chosen": -208.0,
"logps/rejected": -241.0,
"loss": 0.1117,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -3.140625,
"rewards/margins": 3.390625,
"rewards/rejected": -6.53125,
"step": 1550
},
{
"epoch": 1.870503597122302,
"grad_norm": 15.023137073497749,
"learning_rate": 2.09240337627721e-07,
"logits/chosen": -1.6640625,
"logits/rejected": -1.59375,
"logps/chosen": -216.0,
"logps/rejected": -225.0,
"loss": 0.1284,
"rewards/accuracies": 0.9375,
"rewards/chosen": -3.40625,
"rewards/margins": 2.96875,
"rewards/rejected": -6.375,
"step": 1560
},
{
"epoch": 1.882494004796163,
"grad_norm": 23.48973810623808,
"learning_rate": 2.070191026210573e-07,
"logits/chosen": -1.5703125,
"logits/rejected": -1.6015625,
"logps/chosen": -214.0,
"logps/rejected": -239.0,
"loss": 0.1468,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -2.953125,
"rewards/margins": 2.96875,
"rewards/rejected": -5.9375,
"step": 1570
},
{
"epoch": 1.894484412470024,
"grad_norm": 9.16497112579326,
"learning_rate": 2.047978676143936e-07,
"logits/chosen": -1.625,
"logits/rejected": -1.578125,
"logps/chosen": -220.0,
"logps/rejected": -242.0,
"loss": 0.1041,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -2.84375,
"rewards/margins": 3.453125,
"rewards/rejected": -6.28125,
"step": 1580
},
{
"epoch": 1.906474820143885,
"grad_norm": 11.983080775973534,
"learning_rate": 2.025766326077299e-07,
"logits/chosen": -1.640625,
"logits/rejected": -1.609375,
"logps/chosen": -207.0,
"logps/rejected": -248.0,
"loss": 0.1184,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -2.984375,
"rewards/margins": 3.28125,
"rewards/rejected": -6.25,
"step": 1590
},
{
"epoch": 1.9184652278177459,
"grad_norm": 20.978396501783582,
"learning_rate": 2.0035539760106618e-07,
"logits/chosen": -1.625,
"logits/rejected": -1.6328125,
"logps/chosen": -212.0,
"logps/rejected": -256.0,
"loss": 0.1042,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -2.859375,
"rewards/margins": 3.53125,
"rewards/rejected": -6.40625,
"step": 1600
},
{
"epoch": 1.9304556354916067,
"grad_norm": 9.307796901849636,
"learning_rate": 1.9813416259440246e-07,
"logits/chosen": -1.625,
"logits/rejected": -1.6015625,
"logps/chosen": -209.0,
"logps/rejected": -233.0,
"loss": 0.1347,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -3.171875,
"rewards/margins": 2.984375,
"rewards/rejected": -6.15625,
"step": 1610
},
{
"epoch": 1.9424460431654675,
"grad_norm": 22.24759097187433,
"learning_rate": 1.9591292758773879e-07,
"logits/chosen": -1.625,
"logits/rejected": -1.6015625,
"logps/chosen": -196.0,
"logps/rejected": -251.0,
"loss": 0.1543,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -2.765625,
"rewards/margins": 3.640625,
"rewards/rejected": -6.40625,
"step": 1620
},
{
"epoch": 1.9544364508393284,
"grad_norm": 22.879947315900267,
"learning_rate": 1.9369169258107506e-07,
"logits/chosen": -1.6328125,
"logits/rejected": -1.5703125,
"logps/chosen": -214.0,
"logps/rejected": -239.0,
"loss": 0.1174,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -2.84375,
"rewards/margins": 3.359375,
"rewards/rejected": -6.21875,
"step": 1630
},
{
"epoch": 1.9664268585131894,
"grad_norm": 10.630808616318225,
"learning_rate": 1.9147045757441137e-07,
"logits/chosen": -1.6328125,
"logits/rejected": -1.609375,
"logps/chosen": -205.0,
"logps/rejected": -249.0,
"loss": 0.113,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -3.6875,
"rewards/margins": 3.140625,
"rewards/rejected": -6.8125,
"step": 1640
},
{
"epoch": 1.9784172661870505,
"grad_norm": 20.455023304923063,
"learning_rate": 1.8924922256774767e-07,
"logits/chosen": -1.5625,
"logits/rejected": -1.484375,
"logps/chosen": -214.0,
"logps/rejected": -230.0,
"loss": 0.133,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -3.46875,
"rewards/margins": 2.84375,
"rewards/rejected": -6.3125,
"step": 1650
},
{
"epoch": 1.9904076738609113,
"grad_norm": 12.942728073820737,
"learning_rate": 1.8702798756108395e-07,
"logits/chosen": -1.65625,
"logits/rejected": -1.578125,
"logps/chosen": -216.0,
"logps/rejected": -234.0,
"loss": 0.1466,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -2.71875,
"rewards/margins": 3.078125,
"rewards/rejected": -5.78125,
"step": 1660
},
{
"epoch": 2.0,
"eval_logits/chosen": -1.6484375,
"eval_logits/rejected": -1.6640625,
"eval_logps/chosen": -219.0,
"eval_logps/rejected": -228.0,
"eval_loss": 0.7706417441368103,
"eval_rewards/accuracies": 0.6397058963775635,
"eval_rewards/chosen": -4.15625,
"eval_rewards/margins": 1.03125,
"eval_rewards/rejected": -5.1875,
"eval_runtime": 23.2393,
"eval_samples_per_second": 17.255,
"eval_steps_per_second": 0.732,
"step": 1668
},
{
"epoch": 2.002398081534772,
"grad_norm": 9.573444683512719,
"learning_rate": 1.8480675255442025e-07,
"logits/chosen": -1.6796875,
"logits/rejected": -1.640625,
"logps/chosen": -216.0,
"logps/rejected": -256.0,
"loss": 0.1096,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -3.6875,
"rewards/margins": 3.40625,
"rewards/rejected": -7.09375,
"step": 1670
},
{
"epoch": 2.014388489208633,
"grad_norm": 13.805204899223623,
"learning_rate": 1.8258551754775655e-07,
"logits/chosen": -1.6796875,
"logits/rejected": -1.625,
"logps/chosen": -219.0,
"logps/rejected": -236.0,
"loss": 0.0794,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -3.015625,
"rewards/margins": 3.375,
"rewards/rejected": -6.40625,
"step": 1680
},
{
"epoch": 2.026378896882494,
"grad_norm": 5.467275936133308,
"learning_rate": 1.8036428254109286e-07,
"logits/chosen": -1.6171875,
"logits/rejected": -1.5625,
"logps/chosen": -216.0,
"logps/rejected": -276.0,
"loss": 0.0638,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.71875,
"rewards/margins": 3.8125,
"rewards/rejected": -7.53125,
"step": 1690
},
{
"epoch": 2.038369304556355,
"grad_norm": 10.769684058873244,
"learning_rate": 1.7814304753442913e-07,
"logits/chosen": -1.6796875,
"logits/rejected": -1.65625,
"logps/chosen": -215.0,
"logps/rejected": -260.0,
"loss": 0.0615,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.1875,
"rewards/margins": 3.875,
"rewards/rejected": -7.09375,
"step": 1700
},
{
"epoch": 2.050359712230216,
"grad_norm": 5.977598972558003,
"learning_rate": 1.759218125277654e-07,
"logits/chosen": -1.59375,
"logits/rejected": -1.578125,
"logps/chosen": -214.0,
"logps/rejected": -251.0,
"loss": 0.0588,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.703125,
"rewards/margins": 3.5625,
"rewards/rejected": -7.25,
"step": 1710
},
{
"epoch": 2.062350119904077,
"grad_norm": 6.11324671011277,
"learning_rate": 1.7370057752110174e-07,
"logits/chosen": -1.5859375,
"logits/rejected": -1.6171875,
"logps/chosen": -221.0,
"logps/rejected": -262.0,
"loss": 0.0604,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.546875,
"rewards/margins": 4.125,
"rewards/rejected": -7.65625,
"step": 1720
},
{
"epoch": 2.0743405275779376,
"grad_norm": 7.0419948282884555,
"learning_rate": 1.7147934251443802e-07,
"logits/chosen": -1.703125,
"logits/rejected": -1.6484375,
"logps/chosen": -218.0,
"logps/rejected": -250.0,
"loss": 0.0604,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.53125,
"rewards/margins": 3.90625,
"rewards/rejected": -7.4375,
"step": 1730
},
{
"epoch": 2.0863309352517985,
"grad_norm": 13.536498874343664,
"learning_rate": 1.6925810750777432e-07,
"logits/chosen": -1.6875,
"logits/rejected": -1.6171875,
"logps/chosen": -234.0,
"logps/rejected": -268.0,
"loss": 0.0559,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.890625,
"rewards/margins": 4.09375,
"rewards/rejected": -8.0,
"step": 1740
},
{
"epoch": 2.0983213429256593,
"grad_norm": 7.857771486014022,
"learning_rate": 1.6703687250111062e-07,
"logits/chosen": -1.6015625,
"logits/rejected": -1.5703125,
"logps/chosen": -228.0,
"logps/rejected": -260.0,
"loss": 0.0637,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.5,
"rewards/margins": 3.84375,
"rewards/rejected": -8.375,
"step": 1750
},
{
"epoch": 2.1103117505995206,
"grad_norm": 24.029270438668547,
"learning_rate": 1.648156374944469e-07,
"logits/chosen": -1.65625,
"logits/rejected": -1.625,
"logps/chosen": -227.0,
"logps/rejected": -272.0,
"loss": 0.0691,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -4.125,
"rewards/margins": 4.28125,
"rewards/rejected": -8.4375,
"step": 1760
},
{
"epoch": 2.1223021582733814,
"grad_norm": 9.575862100985665,
"learning_rate": 1.625944024877832e-07,
"logits/chosen": -1.640625,
"logits/rejected": -1.5703125,
"logps/chosen": -206.0,
"logps/rejected": -242.0,
"loss": 0.0761,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -3.734375,
"rewards/margins": 3.609375,
"rewards/rejected": -7.34375,
"step": 1770
},
{
"epoch": 2.1342925659472423,
"grad_norm": 8.445962784666682,
"learning_rate": 1.603731674811195e-07,
"logits/chosen": -1.671875,
"logits/rejected": -1.609375,
"logps/chosen": -229.0,
"logps/rejected": -258.0,
"loss": 0.0666,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -3.921875,
"rewards/margins": 3.953125,
"rewards/rejected": -7.875,
"step": 1780
},
{
"epoch": 2.146282973621103,
"grad_norm": 12.5096726751056,
"learning_rate": 1.581519324744558e-07,
"logits/chosen": -1.703125,
"logits/rejected": -1.625,
"logps/chosen": -229.0,
"logps/rejected": -264.0,
"loss": 0.0554,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.515625,
"rewards/margins": 3.875,
"rewards/rejected": -7.375,
"step": 1790
},
{
"epoch": 2.158273381294964,
"grad_norm": 7.424474262544135,
"learning_rate": 1.5593069746779208e-07,
"logits/chosen": -1.640625,
"logits/rejected": -1.609375,
"logps/chosen": -216.0,
"logps/rejected": -250.0,
"loss": 0.0665,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.15625,
"rewards/margins": 3.796875,
"rewards/rejected": -7.9375,
"step": 1800
},
{
"epoch": 2.1702637889688248,
"grad_norm": 7.886336061164498,
"learning_rate": 1.5370946246112836e-07,
"logits/chosen": -1.640625,
"logits/rejected": -1.640625,
"logps/chosen": -212.0,
"logps/rejected": -272.0,
"loss": 0.0565,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.25,
"rewards/margins": 4.34375,
"rewards/rejected": -8.5625,
"step": 1810
},
{
"epoch": 2.182254196642686,
"grad_norm": 5.804381015966981,
"learning_rate": 1.514882274544647e-07,
"logits/chosen": -1.6875,
"logits/rejected": -1.6484375,
"logps/chosen": -220.0,
"logps/rejected": -262.0,
"loss": 0.0475,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.03125,
"rewards/margins": 4.5625,
"rewards/rejected": -8.625,
"step": 1820
},
{
"epoch": 2.194244604316547,
"grad_norm": 8.523559366958416,
"learning_rate": 1.4926699244780097e-07,
"logits/chosen": -1.671875,
"logits/rejected": -1.6171875,
"logps/chosen": -215.0,
"logps/rejected": -262.0,
"loss": 0.0732,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -3.828125,
"rewards/margins": 4.21875,
"rewards/rejected": -8.0625,
"step": 1830
},
{
"epoch": 2.2062350119904077,
"grad_norm": 15.092361293159154,
"learning_rate": 1.4704575744113727e-07,
"logits/chosen": -1.6640625,
"logits/rejected": -1.625,
"logps/chosen": -227.0,
"logps/rejected": -258.0,
"loss": 0.0544,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.25,
"rewards/margins": 4.125,
"rewards/rejected": -8.375,
"step": 1840
},
{
"epoch": 2.2182254196642686,
"grad_norm": 7.809268443424338,
"learning_rate": 1.4482452243447357e-07,
"logits/chosen": -1.6171875,
"logits/rejected": -1.625,
"logps/chosen": -216.0,
"logps/rejected": -266.0,
"loss": 0.0463,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.0,
"rewards/margins": 4.625,
"rewards/rejected": -8.625,
"step": 1850
},
{
"epoch": 2.2302158273381294,
"grad_norm": 6.175474139623565,
"learning_rate": 1.4260328742780985e-07,
"logits/chosen": -1.6796875,
"logits/rejected": -1.59375,
"logps/chosen": -210.0,
"logps/rejected": -251.0,
"loss": 0.0524,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.8125,
"rewards/margins": 4.40625,
"rewards/rejected": -8.1875,
"step": 1860
},
{
"epoch": 2.2422062350119902,
"grad_norm": 8.016024927468784,
"learning_rate": 1.4038205242114615e-07,
"logits/chosen": -1.6328125,
"logits/rejected": -1.5703125,
"logps/chosen": -202.0,
"logps/rejected": -250.0,
"loss": 0.0476,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.671875,
"rewards/margins": 4.21875,
"rewards/rejected": -7.875,
"step": 1870
},
{
"epoch": 2.2541966426858515,
"grad_norm": 46.585404251146365,
"learning_rate": 1.3816081741448243e-07,
"logits/chosen": -1.65625,
"logits/rejected": -1.609375,
"logps/chosen": -214.0,
"logps/rejected": -243.0,
"loss": 0.0878,
"rewards/accuracies": 0.9624999761581421,
"rewards/chosen": -3.703125,
"rewards/margins": 4.125,
"rewards/rejected": -7.84375,
"step": 1880
},
{
"epoch": 2.2661870503597124,
"grad_norm": 14.496947844310064,
"learning_rate": 1.3593958240781876e-07,
"logits/chosen": -1.75,
"logits/rejected": -1.71875,
"logps/chosen": -244.0,
"logps/rejected": -272.0,
"loss": 0.0442,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.953125,
"rewards/margins": 4.59375,
"rewards/rejected": -8.5625,
"step": 1890
},
{
"epoch": 2.278177458033573,
"grad_norm": 10.300232011520219,
"learning_rate": 1.3371834740115504e-07,
"logits/chosen": -1.6875,
"logits/rejected": -1.65625,
"logps/chosen": -236.0,
"logps/rejected": -286.0,
"loss": 0.0516,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.34375,
"rewards/margins": 4.3125,
"rewards/rejected": -8.625,
"step": 1900
},
{
"epoch": 2.290167865707434,
"grad_norm": 8.877528274051476,
"learning_rate": 1.3149711239449131e-07,
"logits/chosen": -1.6640625,
"logits/rejected": -1.6328125,
"logps/chosen": -225.0,
"logps/rejected": -276.0,
"loss": 0.0473,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.71875,
"rewards/margins": 4.375,
"rewards/rejected": -9.125,
"step": 1910
},
{
"epoch": 2.302158273381295,
"grad_norm": 8.315723064599236,
"learning_rate": 1.2927587738782764e-07,
"logits/chosen": -1.6796875,
"logits/rejected": -1.6171875,
"logps/chosen": -218.0,
"logps/rejected": -262.0,
"loss": 0.0593,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.09375,
"rewards/margins": 4.28125,
"rewards/rejected": -8.375,
"step": 1920
},
{
"epoch": 2.3141486810551557,
"grad_norm": 13.642713487818822,
"learning_rate": 1.2705464238116392e-07,
"logits/chosen": -1.65625,
"logits/rejected": -1.6171875,
"logps/chosen": -209.0,
"logps/rejected": -260.0,
"loss": 0.0621,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.09375,
"rewards/margins": 4.59375,
"rewards/rejected": -8.6875,
"step": 1930
},
{
"epoch": 2.326139088729017,
"grad_norm": 13.131558338647796,
"learning_rate": 1.2483340737450022e-07,
"logits/chosen": -1.65625,
"logits/rejected": -1.609375,
"logps/chosen": -214.0,
"logps/rejected": -258.0,
"loss": 0.0602,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -3.84375,
"rewards/margins": 4.28125,
"rewards/rejected": -8.125,
"step": 1940
},
{
"epoch": 2.338129496402878,
"grad_norm": 10.187009587459894,
"learning_rate": 1.226121723678365e-07,
"logits/chosen": -1.6640625,
"logits/rejected": -1.6328125,
"logps/chosen": -206.0,
"logps/rejected": -258.0,
"loss": 0.05,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.671875,
"rewards/margins": 4.625,
"rewards/rejected": -8.3125,
"step": 1950
},
{
"epoch": 2.3501199040767387,
"grad_norm": 7.106372818158467,
"learning_rate": 1.203909373611728e-07,
"logits/chosen": -1.671875,
"logits/rejected": -1.6015625,
"logps/chosen": -213.0,
"logps/rejected": -249.0,
"loss": 0.0516,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.03125,
"rewards/margins": 4.21875,
"rewards/rejected": -8.25,
"step": 1960
},
{
"epoch": 2.3621103117505995,
"grad_norm": 6.0024755856013305,
"learning_rate": 1.1816970235450909e-07,
"logits/chosen": -1.6875,
"logits/rejected": -1.609375,
"logps/chosen": -224.0,
"logps/rejected": -280.0,
"loss": 0.045,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.1875,
"rewards/margins": 4.75,
"rewards/rejected": -8.9375,
"step": 1970
},
{
"epoch": 2.3741007194244603,
"grad_norm": 7.689477505076036,
"learning_rate": 1.159484673478454e-07,
"logits/chosen": -1.6796875,
"logits/rejected": -1.6953125,
"logps/chosen": -226.0,
"logps/rejected": -270.0,
"loss": 0.043,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.0625,
"rewards/margins": 4.40625,
"rewards/rejected": -8.4375,
"step": 1980
},
{
"epoch": 2.386091127098321,
"grad_norm": 7.812529725908209,
"learning_rate": 1.1372723234118169e-07,
"logits/chosen": -1.6640625,
"logits/rejected": -1.625,
"logps/chosen": -219.0,
"logps/rejected": -258.0,
"loss": 0.0501,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.25,
"rewards/margins": 4.125,
"rewards/rejected": -8.375,
"step": 1990
},
{
"epoch": 2.3980815347721824,
"grad_norm": 49.4364201724221,
"learning_rate": 1.1150599733451799e-07,
"logits/chosen": -1.7265625,
"logits/rejected": -1.671875,
"logps/chosen": -232.0,
"logps/rejected": -298.0,
"loss": 0.0593,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.1875,
"rewards/margins": 4.625,
"rewards/rejected": -8.8125,
"step": 2000
},
{
"epoch": 2.4100719424460433,
"grad_norm": 4.317275529656694,
"learning_rate": 1.0928476232785429e-07,
"logits/chosen": -1.671875,
"logits/rejected": -1.65625,
"logps/chosen": -220.0,
"logps/rejected": -266.0,
"loss": 0.0362,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.125,
"rewards/margins": 4.78125,
"rewards/rejected": -8.875,
"step": 2010
},
{
"epoch": 2.422062350119904,
"grad_norm": 7.089907087105454,
"learning_rate": 1.0706352732119057e-07,
"logits/chosen": -1.71875,
"logits/rejected": -1.6484375,
"logps/chosen": -214.0,
"logps/rejected": -256.0,
"loss": 0.0413,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -3.671875,
"rewards/margins": 4.4375,
"rewards/rejected": -8.125,
"step": 2020
},
{
"epoch": 2.434052757793765,
"grad_norm": 10.12450913020343,
"learning_rate": 1.0484229231452687e-07,
"logits/chosen": -1.7265625,
"logits/rejected": -1.65625,
"logps/chosen": -230.0,
"logps/rejected": -272.0,
"loss": 0.0598,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.65625,
"rewards/margins": 4.375,
"rewards/rejected": -9.0625,
"step": 2030
},
{
"epoch": 2.446043165467626,
"grad_norm": 10.848851517796744,
"learning_rate": 1.0262105730786316e-07,
"logits/chosen": -1.703125,
"logits/rejected": -1.7265625,
"logps/chosen": -213.0,
"logps/rejected": -268.0,
"loss": 0.0475,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.5,
"rewards/margins": 4.1875,
"rewards/rejected": -8.6875,
"step": 2040
},
{
"epoch": 2.4580335731414866,
"grad_norm": 3.5169019202687584,
"learning_rate": 1.0039982230119947e-07,
"logits/chosen": -1.6953125,
"logits/rejected": -1.703125,
"logps/chosen": -219.0,
"logps/rejected": -262.0,
"loss": 0.0752,
"rewards/accuracies": 0.949999988079071,
"rewards/chosen": -4.875,
"rewards/margins": 4.34375,
"rewards/rejected": -9.1875,
"step": 2050
},
{
"epoch": 2.470023980815348,
"grad_norm": 11.726493677946326,
"learning_rate": 9.817858729453577e-08,
"logits/chosen": -1.7109375,
"logits/rejected": -1.671875,
"logps/chosen": -220.0,
"logps/rejected": -274.0,
"loss": 0.054,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.21875,
"rewards/margins": 4.625,
"rewards/rejected": -8.875,
"step": 2060
},
{
"epoch": 2.4820143884892087,
"grad_norm": 10.3815257923276,
"learning_rate": 9.595735228787204e-08,
"logits/chosen": -1.6953125,
"logits/rejected": -1.65625,
"logps/chosen": -221.0,
"logps/rejected": -258.0,
"loss": 0.0545,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.65625,
"rewards/margins": 4.21875,
"rewards/rejected": -8.875,
"step": 2070
},
{
"epoch": 2.4940047961630696,
"grad_norm": 5.520164218604047,
"learning_rate": 9.373611728120835e-08,
"logits/chosen": -1.7109375,
"logits/rejected": -1.671875,
"logps/chosen": -219.0,
"logps/rejected": -255.0,
"loss": 0.0396,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.28125,
"rewards/margins": 4.09375,
"rewards/rejected": -8.375,
"step": 2080
},
{
"epoch": 2.5059952038369304,
"grad_norm": 8.376756986846386,
"learning_rate": 9.151488227454464e-08,
"logits/chosen": -1.6875,
"logits/rejected": -1.6484375,
"logps/chosen": -208.0,
"logps/rejected": -262.0,
"loss": 0.0427,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.0,
"rewards/margins": 4.75,
"rewards/rejected": -8.75,
"step": 2090
},
{
"epoch": 2.5179856115107913,
"grad_norm": 20.20395693776781,
"learning_rate": 8.929364726788094e-08,
"logits/chosen": -1.734375,
"logits/rejected": -1.6953125,
"logps/chosen": -226.0,
"logps/rejected": -274.0,
"loss": 0.0661,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.6875,
"rewards/margins": 4.5,
"rewards/rejected": -9.1875,
"step": 2100
},
{
"epoch": 2.5299760191846525,
"grad_norm": 30.373447856641558,
"learning_rate": 8.707241226121723e-08,
"logits/chosen": -1.671875,
"logits/rejected": -1.703125,
"logps/chosen": -220.0,
"logps/rejected": -284.0,
"loss": 0.0453,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.6875,
"rewards/margins": 4.59375,
"rewards/rejected": -9.3125,
"step": 2110
},
{
"epoch": 2.5419664268585134,
"grad_norm": 3.057799168034458,
"learning_rate": 8.485117725455352e-08,
"logits/chosen": -1.7109375,
"logits/rejected": -1.6796875,
"logps/chosen": -241.0,
"logps/rejected": -288.0,
"loss": 0.0326,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.84375,
"rewards/margins": 4.6875,
"rewards/rejected": -9.5,
"step": 2120
},
{
"epoch": 2.553956834532374,
"grad_norm": 8.47051827415506,
"learning_rate": 8.262994224788982e-08,
"logits/chosen": -1.734375,
"logits/rejected": -1.6796875,
"logps/chosen": -242.0,
"logps/rejected": -282.0,
"loss": 0.0523,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -5.09375,
"rewards/margins": 4.25,
"rewards/rejected": -9.3125,
"step": 2130
},
{
"epoch": 2.565947242206235,
"grad_norm": 3.8022915967693662,
"learning_rate": 8.040870724122611e-08,
"logits/chosen": -1.6796875,
"logits/rejected": -1.6640625,
"logps/chosen": -202.0,
"logps/rejected": -256.0,
"loss": 0.0353,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.953125,
"rewards/margins": 4.625,
"rewards/rejected": -8.5625,
"step": 2140
},
{
"epoch": 2.577937649880096,
"grad_norm": 22.652691529528166,
"learning_rate": 7.818747223456242e-08,
"logits/chosen": -1.703125,
"logits/rejected": -1.671875,
"logps/chosen": -231.0,
"logps/rejected": -270.0,
"loss": 0.0421,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.5625,
"rewards/margins": 4.84375,
"rewards/rejected": -9.4375,
"step": 2150
},
{
"epoch": 2.5899280575539567,
"grad_norm": 5.621319686196611,
"learning_rate": 7.596623722789871e-08,
"logits/chosen": -1.6328125,
"logits/rejected": -1.625,
"logps/chosen": -189.0,
"logps/rejected": -243.0,
"loss": 0.0383,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.125,
"rewards/margins": 4.625,
"rewards/rejected": -8.75,
"step": 2160
},
{
"epoch": 2.6019184652278176,
"grad_norm": 9.164491714974064,
"learning_rate": 7.3745002221235e-08,
"logits/chosen": -1.75,
"logits/rejected": -1.6796875,
"logps/chosen": -242.0,
"logps/rejected": -280.0,
"loss": 0.0414,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.53125,
"rewards/margins": 4.53125,
"rewards/rejected": -9.0625,
"step": 2170
},
{
"epoch": 2.6139088729016784,
"grad_norm": 5.372990484400878,
"learning_rate": 7.15237672145713e-08,
"logits/chosen": -1.71875,
"logits/rejected": -1.65625,
"logps/chosen": -212.0,
"logps/rejected": -258.0,
"loss": 0.0366,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.09375,
"rewards/margins": 5.3125,
"rewards/rejected": -9.375,
"step": 2180
},
{
"epoch": 2.6258992805755397,
"grad_norm": 10.746900341332498,
"learning_rate": 6.930253220790759e-08,
"logits/chosen": -1.75,
"logits/rejected": -1.6875,
"logps/chosen": -220.0,
"logps/rejected": -278.0,
"loss": 0.0452,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.65625,
"rewards/margins": 4.4375,
"rewards/rejected": -9.125,
"step": 2190
},
{
"epoch": 2.6378896882494005,
"grad_norm": 11.993659380624129,
"learning_rate": 6.70812972012439e-08,
"logits/chosen": -1.7421875,
"logits/rejected": -1.703125,
"logps/chosen": -212.0,
"logps/rejected": -262.0,
"loss": 0.0386,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.40625,
"rewards/margins": 4.9375,
"rewards/rejected": -9.3125,
"step": 2200
},
{
"epoch": 2.6498800959232613,
"grad_norm": 10.522371499574833,
"learning_rate": 6.486006219458018e-08,
"logits/chosen": -1.6640625,
"logits/rejected": -1.6328125,
"logps/chosen": -216.0,
"logps/rejected": -255.0,
"loss": 0.0462,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.4375,
"rewards/margins": 4.625,
"rewards/rejected": -9.0625,
"step": 2210
},
{
"epoch": 2.661870503597122,
"grad_norm": 12.225253766846958,
"learning_rate": 6.263882718791647e-08,
"logits/chosen": -1.6484375,
"logits/rejected": -1.6171875,
"logps/chosen": -228.0,
"logps/rejected": -268.0,
"loss": 0.0522,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.125,
"rewards/margins": 4.84375,
"rewards/rejected": -8.9375,
"step": 2220
},
{
"epoch": 2.6738609112709835,
"grad_norm": 12.629562732845173,
"learning_rate": 6.041759218125278e-08,
"logits/chosen": -1.671875,
"logits/rejected": -1.640625,
"logps/chosen": -221.0,
"logps/rejected": -262.0,
"loss": 0.0446,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.46875,
"rewards/margins": 4.40625,
"rewards/rejected": -8.875,
"step": 2230
},
{
"epoch": 2.6858513189448443,
"grad_norm": 8.807729780911718,
"learning_rate": 5.8196357174589066e-08,
"logits/chosen": -1.734375,
"logits/rejected": -1.7109375,
"logps/chosen": -241.0,
"logps/rejected": -280.0,
"loss": 0.0384,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.59375,
"rewards/margins": 4.625,
"rewards/rejected": -9.25,
"step": 2240
},
{
"epoch": 2.697841726618705,
"grad_norm": 2.884098452133717,
"learning_rate": 5.597512216792536e-08,
"logits/chosen": -1.734375,
"logits/rejected": -1.65625,
"logps/chosen": -224.0,
"logps/rejected": -262.0,
"loss": 0.036,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.09375,
"rewards/margins": 4.40625,
"rewards/rejected": -8.5,
"step": 2250
},
{
"epoch": 2.709832134292566,
"grad_norm": 7.527525000207272,
"learning_rate": 5.375388716126166e-08,
"logits/chosen": -1.703125,
"logits/rejected": -1.640625,
"logps/chosen": -212.0,
"logps/rejected": -249.0,
"loss": 0.0436,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.46875,
"rewards/margins": 4.65625,
"rewards/rejected": -9.125,
"step": 2260
},
{
"epoch": 2.721822541966427,
"grad_norm": 22.77210894040614,
"learning_rate": 5.1532652154597956e-08,
"logits/chosen": -1.703125,
"logits/rejected": -1.6796875,
"logps/chosen": -236.0,
"logps/rejected": -294.0,
"loss": 0.0441,
"rewards/accuracies": 1.0,
"rewards/chosen": -5.3125,
"rewards/margins": 5.21875,
"rewards/rejected": -10.5625,
"step": 2270
},
{
"epoch": 2.7338129496402876,
"grad_norm": 7.128969603906596,
"learning_rate": 4.9311417147934246e-08,
"logits/chosen": -1.703125,
"logits/rejected": -1.6796875,
"logps/chosen": -222.0,
"logps/rejected": -260.0,
"loss": 0.0474,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.625,
"rewards/margins": 4.46875,
"rewards/rejected": -9.0625,
"step": 2280
},
{
"epoch": 2.7458033573141485,
"grad_norm": 6.150942157607421,
"learning_rate": 4.709018214127054e-08,
"logits/chosen": -1.78125,
"logits/rejected": -1.71875,
"logps/chosen": -230.0,
"logps/rejected": -280.0,
"loss": 0.0352,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.5625,
"rewards/margins": 4.78125,
"rewards/rejected": -9.375,
"step": 2290
},
{
"epoch": 2.7577937649880093,
"grad_norm": 2.8164036629567843,
"learning_rate": 4.486894713460684e-08,
"logits/chosen": -1.7421875,
"logits/rejected": -1.6875,
"logps/chosen": -222.0,
"logps/rejected": -266.0,
"loss": 0.0347,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.6875,
"rewards/margins": 4.65625,
"rewards/rejected": -9.375,
"step": 2300
},
{
"epoch": 2.7697841726618706,
"grad_norm": 7.133912331859571,
"learning_rate": 4.2647712127943136e-08,
"logits/chosen": -1.765625,
"logits/rejected": -1.7578125,
"logps/chosen": -239.0,
"logps/rejected": -304.0,
"loss": 0.0356,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.75,
"rewards/margins": 4.90625,
"rewards/rejected": -9.6875,
"step": 2310
},
{
"epoch": 2.7817745803357314,
"grad_norm": 5.416214434119401,
"learning_rate": 4.042647712127943e-08,
"logits/chosen": -1.71875,
"logits/rejected": -1.6953125,
"logps/chosen": -230.0,
"logps/rejected": -278.0,
"loss": 0.0358,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -5.625,
"rewards/margins": 4.5625,
"rewards/rejected": -10.1875,
"step": 2320
},
{
"epoch": 2.7937649880095923,
"grad_norm": 7.05206840398353,
"learning_rate": 3.820524211461572e-08,
"logits/chosen": -1.6875,
"logits/rejected": -1.640625,
"logps/chosen": -214.0,
"logps/rejected": -272.0,
"loss": 0.0468,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -5.1875,
"rewards/margins": 4.40625,
"rewards/rejected": -9.5625,
"step": 2330
},
{
"epoch": 2.805755395683453,
"grad_norm": 6.8237218283267245,
"learning_rate": 3.598400710795202e-08,
"logits/chosen": -1.7109375,
"logits/rejected": -1.671875,
"logps/chosen": -239.0,
"logps/rejected": -296.0,
"loss": 0.0383,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.9375,
"rewards/margins": 5.25,
"rewards/rejected": -10.1875,
"step": 2340
},
{
"epoch": 2.8177458033573144,
"grad_norm": 17.473074202115587,
"learning_rate": 3.3762772101288315e-08,
"logits/chosen": -1.640625,
"logits/rejected": -1.640625,
"logps/chosen": -218.0,
"logps/rejected": -268.0,
"loss": 0.0388,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.84375,
"rewards/margins": 4.4375,
"rewards/rejected": -9.25,
"step": 2350
},
{
"epoch": 2.8297362110311752,
"grad_norm": 6.423408193266802,
"learning_rate": 3.1541537094624605e-08,
"logits/chosen": -1.734375,
"logits/rejected": -1.640625,
"logps/chosen": -233.0,
"logps/rejected": -264.0,
"loss": 0.0436,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.78125,
"rewards/margins": 4.84375,
"rewards/rejected": -9.625,
"step": 2360
},
{
"epoch": 2.841726618705036,
"grad_norm": 9.316314943177023,
"learning_rate": 2.9320302087960905e-08,
"logits/chosen": -1.6796875,
"logits/rejected": -1.65625,
"logps/chosen": -223.0,
"logps/rejected": -280.0,
"loss": 0.0494,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -5.25,
"rewards/margins": 4.875,
"rewards/rejected": -10.125,
"step": 2370
},
{
"epoch": 2.853717026378897,
"grad_norm": 4.563487715111162,
"learning_rate": 2.70990670812972e-08,
"logits/chosen": -1.6796875,
"logits/rejected": -1.6328125,
"logps/chosen": -215.0,
"logps/rejected": -264.0,
"loss": 0.0411,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -5.5625,
"rewards/margins": 4.71875,
"rewards/rejected": -10.25,
"step": 2380
},
{
"epoch": 2.8657074340527577,
"grad_norm": 7.265458467508151,
"learning_rate": 2.4877832074633495e-08,
"logits/chosen": -1.734375,
"logits/rejected": -1.6328125,
"logps/chosen": -219.0,
"logps/rejected": -266.0,
"loss": 0.035,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.96875,
"rewards/margins": 4.875,
"rewards/rejected": -9.8125,
"step": 2390
},
{
"epoch": 2.8776978417266186,
"grad_norm": 3.9055052792545744,
"learning_rate": 2.2656597067969788e-08,
"logits/chosen": -1.6875,
"logits/rejected": -1.7109375,
"logps/chosen": -217.0,
"logps/rejected": -272.0,
"loss": 0.0387,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.46875,
"rewards/margins": 5.0,
"rewards/rejected": -9.5,
"step": 2400
},
{
"epoch": 2.8896882494004794,
"grad_norm": 3.850000762426418,
"learning_rate": 2.0435362061306088e-08,
"logits/chosen": -1.703125,
"logits/rejected": -1.6875,
"logps/chosen": -222.0,
"logps/rejected": -260.0,
"loss": 0.0553,
"rewards/accuracies": 1.0,
"rewards/chosen": -5.1875,
"rewards/margins": 4.8125,
"rewards/rejected": -10.0,
"step": 2410
},
{
"epoch": 2.9016786570743403,
"grad_norm": 3.8530749601946486,
"learning_rate": 1.821412705464238e-08,
"logits/chosen": -1.765625,
"logits/rejected": -1.6796875,
"logps/chosen": -225.0,
"logps/rejected": -300.0,
"loss": 0.033,
"rewards/accuracies": 1.0,
"rewards/chosen": -5.4375,
"rewards/margins": 5.25,
"rewards/rejected": -10.6875,
"step": 2420
},
{
"epoch": 2.9136690647482015,
"grad_norm": 5.640559221099878,
"learning_rate": 1.5992892047978674e-08,
"logits/chosen": -1.6953125,
"logits/rejected": -1.671875,
"logps/chosen": -226.0,
"logps/rejected": -280.0,
"loss": 0.0343,
"rewards/accuracies": 1.0,
"rewards/chosen": -5.0,
"rewards/margins": 4.71875,
"rewards/rejected": -9.75,
"step": 2430
},
{
"epoch": 2.9256594724220624,
"grad_norm": 4.969640872073211,
"learning_rate": 1.377165704131497e-08,
"logits/chosen": -1.6640625,
"logits/rejected": -1.6484375,
"logps/chosen": -218.0,
"logps/rejected": -274.0,
"loss": 0.038,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.65625,
"rewards/margins": 5.0625,
"rewards/rejected": -9.6875,
"step": 2440
},
{
"epoch": 2.937649880095923,
"grad_norm": 5.866247247572104,
"learning_rate": 1.1550422034651266e-08,
"logits/chosen": -1.75,
"logits/rejected": -1.6484375,
"logps/chosen": -219.0,
"logps/rejected": -270.0,
"loss": 0.039,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.78125,
"rewards/margins": 4.78125,
"rewards/rejected": -9.5625,
"step": 2450
},
{
"epoch": 2.949640287769784,
"grad_norm": 4.3423386310633365,
"learning_rate": 9.32918702798756e-09,
"logits/chosen": -1.71875,
"logits/rejected": -1.65625,
"logps/chosen": -220.0,
"logps/rejected": -262.0,
"loss": 0.0506,
"rewards/accuracies": 1.0,
"rewards/chosen": -5.1875,
"rewards/margins": 4.21875,
"rewards/rejected": -9.4375,
"step": 2460
},
{
"epoch": 2.9616306954436453,
"grad_norm": 7.560339052869031,
"learning_rate": 7.107952021323855e-09,
"logits/chosen": -1.7265625,
"logits/rejected": -1.6796875,
"logps/chosen": -219.0,
"logps/rejected": -268.0,
"loss": 0.0335,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.96875,
"rewards/margins": 5.09375,
"rewards/rejected": -10.0625,
"step": 2470
},
{
"epoch": 2.973621103117506,
"grad_norm": 10.814073549670898,
"learning_rate": 4.886717014660151e-09,
"logits/chosen": -1.75,
"logits/rejected": -1.640625,
"logps/chosen": -232.0,
"logps/rejected": -260.0,
"loss": 0.0393,
"rewards/accuracies": 0.987500011920929,
"rewards/chosen": -4.875,
"rewards/margins": 4.78125,
"rewards/rejected": -9.6875,
"step": 2480
},
{
"epoch": 2.985611510791367,
"grad_norm": 4.9409390032818665,
"learning_rate": 2.665482007996446e-09,
"logits/chosen": -1.7265625,
"logits/rejected": -1.671875,
"logps/chosen": -228.0,
"logps/rejected": -260.0,
"loss": 0.0378,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -4.96875,
"rewards/margins": 4.28125,
"rewards/rejected": -9.25,
"step": 2490
},
{
"epoch": 2.997601918465228,
"grad_norm": 19.048106847839737,
"learning_rate": 4.4424700133274096e-10,
"logits/chosen": -1.71875,
"logits/rejected": -1.640625,
"logps/chosen": -238.0,
"logps/rejected": -274.0,
"loss": 0.0398,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.6875,
"rewards/margins": 4.9375,
"rewards/rejected": -9.625,
"step": 2500
},
{
"epoch": 3.0,
"eval_logits/chosen": -1.71875,
"eval_logits/rejected": -1.7421875,
"eval_logps/chosen": -240.0,
"eval_logps/rejected": -250.0,
"eval_loss": 0.9926745891571045,
"eval_rewards/accuracies": 0.6029411554336548,
"eval_rewards/chosen": -6.21875,
"eval_rewards/margins": 1.1953125,
"eval_rewards/rejected": -7.4375,
"eval_runtime": 28.9311,
"eval_samples_per_second": 13.861,
"eval_steps_per_second": 0.588,
"step": 2502
},
{
"epoch": 3.0,
"step": 2502,
"total_flos": 0.0,
"train_loss": 0.2617608135004791,
"train_runtime": 8206.5437,
"train_samples_per_second": 7.311,
"train_steps_per_second": 0.305
}
],
"logging_steps": 10,
"max_steps": 2502,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}