Model save
Browse files- README.md +1 -1
- all_results.json +4 -4
- train_results.json +4 -4
- trainer_state.json +255 -255
README.md
CHANGED
@@ -27,7 +27,7 @@ print(output["generated_text"])
|
|
27 |
|
28 |
## Training procedure
|
29 |
|
30 |
-
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/causalai/huggingface/runs/
|
31 |
|
32 |
|
33 |
This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
|
|
|
27 |
|
28 |
## Training procedure
|
29 |
|
30 |
+
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/causalai/huggingface/runs/mth6kytm)
|
31 |
|
32 |
|
33 |
This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
|
all_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"total_flos": 0.0,
|
3 |
-
"train_loss": 0.
|
4 |
-
"train_runtime":
|
5 |
"train_samples": 17056,
|
6 |
-
"train_samples_per_second":
|
7 |
-
"train_steps_per_second": 0.
|
8 |
}
|
|
|
1 |
{
|
2 |
"total_flos": 0.0,
|
3 |
+
"train_loss": 0.11666255951964093,
|
4 |
+
"train_runtime": 146747.14,
|
5 |
"train_samples": 17056,
|
6 |
+
"train_samples_per_second": 0.116,
|
7 |
+
"train_steps_per_second": 0.001
|
8 |
}
|
train_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"total_flos": 0.0,
|
3 |
-
"train_loss": 0.
|
4 |
-
"train_runtime":
|
5 |
"train_samples": 17056,
|
6 |
-
"train_samples_per_second":
|
7 |
-
"train_steps_per_second": 0.
|
8 |
}
|
|
|
1 |
{
|
2 |
"total_flos": 0.0,
|
3 |
+
"train_loss": 0.11666255951964093,
|
4 |
+
"train_runtime": 146747.14,
|
5 |
"train_samples": 17056,
|
6 |
+
"train_samples_per_second": 0.116,
|
7 |
+
"train_steps_per_second": 0.001
|
8 |
}
|
trainer_state.json
CHANGED
@@ -9,399 +9,399 @@
|
|
9 |
"is_world_process_zero": true,
|
10 |
"log_history": [
|
11 |
{
|
12 |
-
"completion_length":
|
13 |
"epoch": 0.0375234521575985,
|
14 |
-
"grad_norm":
|
15 |
-
"kl": 0.
|
16 |
"learning_rate": 7.1428571428571436e-06,
|
17 |
-
"loss": 0.
|
18 |
-
"reward": 0.
|
19 |
-
"reward_std": 0.
|
20 |
"rewards/accuracy_reward": 0.0158203125,
|
21 |
-
"rewards/format_reward": 0.
|
22 |
-
"rewards/relaxed_accuracy_reward": 0.
|
23 |
"step": 5
|
24 |
},
|
25 |
{
|
26 |
-
"completion_length":
|
27 |
"epoch": 0.075046904315197,
|
28 |
-
"grad_norm":
|
29 |
-
"kl":
|
30 |
"learning_rate": 1.4285714285714287e-05,
|
31 |
-
"loss": 0.
|
32 |
-
"reward": 0.
|
33 |
-
"reward_std": 0.
|
34 |
-
"rewards/accuracy_reward": 0.
|
35 |
-
"rewards/format_reward": 0.
|
36 |
-
"rewards/relaxed_accuracy_reward": 0.
|
37 |
"step": 10
|
38 |
},
|
39 |
{
|
40 |
-
"completion_length":
|
41 |
"epoch": 0.1125703564727955,
|
42 |
-
"grad_norm":
|
43 |
-
"kl":
|
44 |
"learning_rate": 1.9996515418688493e-05,
|
45 |
-
"loss": 0.
|
46 |
-
"reward": 0.
|
47 |
-
"reward_std": 0.
|
48 |
-
"rewards/accuracy_reward": 0.
|
49 |
-
"rewards/format_reward": 0.
|
50 |
-
"rewards/relaxed_accuracy_reward": 0.
|
51 |
"step": 15
|
52 |
},
|
53 |
{
|
54 |
-
"completion_length":
|
55 |
"epoch": 0.150093808630394,
|
56 |
-
"grad_norm":
|
57 |
-
"kl": 1.
|
58 |
"learning_rate": 1.9874809871741877e-05,
|
59 |
-
"loss": 0.
|
60 |
-
"reward":
|
61 |
-
"reward_std": 0.
|
62 |
-
"rewards/accuracy_reward": 0.
|
63 |
-
"rewards/format_reward": 0.
|
64 |
-
"rewards/relaxed_accuracy_reward": 0.
|
65 |
"step": 20
|
66 |
},
|
67 |
{
|
68 |
-
"completion_length":
|
69 |
"epoch": 0.18761726078799248,
|
70 |
-
"grad_norm":
|
71 |
-
"kl":
|
72 |
"learning_rate": 1.9581296124106682e-05,
|
73 |
-
"loss": 0.
|
74 |
-
"reward":
|
75 |
-
"reward_std": 0.
|
76 |
-
"rewards/accuracy_reward": 0.
|
77 |
-
"rewards/format_reward": 0.
|
78 |
-
"rewards/relaxed_accuracy_reward": 0.
|
79 |
"step": 25
|
80 |
},
|
81 |
{
|
82 |
-
"completion_length":
|
83 |
"epoch": 0.225140712945591,
|
84 |
-
"grad_norm":
|
85 |
-
"kl":
|
86 |
"learning_rate": 1.912108091398988e-05,
|
87 |
-
"loss": 0.
|
88 |
-
"reward":
|
89 |
-
"reward_std": 0.
|
90 |
-
"rewards/accuracy_reward": 0.
|
91 |
-
"rewards/format_reward": 0.
|
92 |
-
"rewards/relaxed_accuracy_reward": 0.
|
93 |
"step": 30
|
94 |
},
|
95 |
{
|
96 |
-
"completion_length":
|
97 |
"epoch": 0.2626641651031895,
|
98 |
-
"grad_norm":
|
99 |
-
"kl":
|
100 |
"learning_rate": 1.8502171357296144e-05,
|
101 |
-
"loss": 0.
|
102 |
-
"reward":
|
103 |
-
"reward_std": 0.
|
104 |
-
"rewards/accuracy_reward": 0.
|
105 |
-
"rewards/format_reward": 0.
|
106 |
-
"rewards/relaxed_accuracy_reward": 0.
|
107 |
"step": 35
|
108 |
},
|
109 |
{
|
110 |
-
"completion_length":
|
111 |
"epoch": 0.300187617260788,
|
112 |
-
"grad_norm":
|
113 |
-
"kl":
|
114 |
"learning_rate": 1.773533563475053e-05,
|
115 |
-
"loss": 0.
|
116 |
-
"reward":
|
117 |
-
"reward_std": 0.
|
118 |
-
"rewards/accuracy_reward": 0.
|
119 |
-
"rewards/format_reward": 0.
|
120 |
-
"rewards/relaxed_accuracy_reward": 0.
|
121 |
"step": 40
|
122 |
},
|
123 |
{
|
124 |
-
"completion_length":
|
125 |
"epoch": 0.33771106941838647,
|
126 |
-
"grad_norm":
|
127 |
-
"kl":
|
128 |
"learning_rate": 1.6833915640265485e-05,
|
129 |
-
"loss":
|
130 |
-
"reward":
|
131 |
-
"reward_std": 0.
|
132 |
-
"rewards/accuracy_reward": 0.
|
133 |
-
"rewards/format_reward": 0.
|
134 |
-
"rewards/relaxed_accuracy_reward": 0.
|
135 |
"step": 45
|
136 |
},
|
137 |
{
|
138 |
-
"completion_length":
|
139 |
"epoch": 0.37523452157598497,
|
140 |
-
"grad_norm":
|
141 |
-
"kl":
|
142 |
"learning_rate": 1.58135948502146e-05,
|
143 |
-
"loss": 0.
|
144 |
-
"reward":
|
145 |
-
"reward_std": 0.
|
146 |
-
"rewards/accuracy_reward": 0.
|
147 |
-
"rewards/format_reward": 0.
|
148 |
-
"rewards/relaxed_accuracy_reward": 0.
|
149 |
"step": 50
|
150 |
},
|
151 |
{
|
152 |
-
"completion_length":
|
153 |
"epoch": 0.41275797373358347,
|
154 |
-
"grad_norm":
|
155 |
-
"kl": 1.
|
156 |
"learning_rate": 1.4692125452370664e-05,
|
157 |
-
"loss": 0.
|
158 |
-
"reward":
|
159 |
-
"reward_std": 0.
|
160 |
-
"rewards/accuracy_reward": 0.
|
161 |
-
"rewards/format_reward": 0.
|
162 |
-
"rewards/relaxed_accuracy_reward": 0.
|
163 |
"step": 55
|
164 |
},
|
165 |
{
|
166 |
-
"completion_length":
|
167 |
"epoch": 0.450281425891182,
|
168 |
-
"grad_norm":
|
169 |
-
"kl":
|
170 |
"learning_rate": 1.348901948209167e-05,
|
171 |
-
"loss": 0.
|
172 |
-
"reward":
|
173 |
-
"reward_std": 0.
|
174 |
-
"rewards/accuracy_reward": 0.
|
175 |
-
"rewards/format_reward": 0.
|
176 |
-
"rewards/relaxed_accuracy_reward": 0.
|
177 |
"step": 60
|
178 |
},
|
179 |
{
|
180 |
-
"completion_length":
|
181 |
"epoch": 0.4878048780487805,
|
182 |
-
"grad_norm":
|
183 |
-
"kl":
|
184 |
"learning_rate": 1.2225209339563144e-05,
|
185 |
-
"loss": 0.
|
186 |
-
"reward":
|
187 |
-
"reward_std": 0.
|
188 |
-
"rewards/accuracy_reward": 0.
|
189 |
-
"rewards/format_reward": 0.
|
190 |
-
"rewards/relaxed_accuracy_reward": 0.
|
191 |
"step": 65
|
192 |
},
|
193 |
{
|
194 |
-
"completion_length":
|
195 |
"epoch": 0.525328330206379,
|
196 |
-
"grad_norm":
|
197 |
-
"kl":
|
198 |
"learning_rate": 1.092268359463302e-05,
|
199 |
-
"loss": 0.
|
200 |
-
"reward":
|
201 |
-
"reward_std": 0.
|
202 |
-
"rewards/accuracy_reward": 0.
|
203 |
-
"rewards/format_reward": 0.
|
204 |
-
"rewards/relaxed_accuracy_reward": 0.
|
205 |
"step": 70
|
206 |
},
|
207 |
{
|
208 |
-
"completion_length":
|
209 |
"epoch": 0.5628517823639775,
|
210 |
-
"grad_norm":
|
211 |
-
"kl":
|
212 |
"learning_rate": 9.604104415737309e-06,
|
213 |
-
"loss": 0.
|
214 |
-
"reward":
|
215 |
-
"reward_std": 0.
|
216 |
-
"rewards/accuracy_reward": 0.
|
217 |
-
"rewards/format_reward": 0.
|
218 |
-
"rewards/relaxed_accuracy_reward": 0.
|
219 |
"step": 75
|
220 |
},
|
221 |
{
|
222 |
-
"completion_length":
|
223 |
"epoch": 0.600375234521576,
|
224 |
-
"grad_norm":
|
225 |
-
"kl":
|
226 |
"learning_rate": 8.292413279130625e-06,
|
227 |
-
"loss": 0.
|
228 |
-
"reward":
|
229 |
-
"reward_std": 0.
|
230 |
-
"rewards/accuracy_reward": 0.
|
231 |
-
"rewards/format_reward": 0.
|
232 |
-
"rewards/relaxed_accuracy_reward": 0.
|
233 |
"step": 80
|
234 |
},
|
235 |
{
|
236 |
-
"completion_length":
|
237 |
"epoch": 0.6378986866791745,
|
238 |
-
"grad_norm":
|
239 |
-
"kl": 1.
|
240 |
"learning_rate": 7.010431818542298e-06,
|
241 |
-
"loss": 0.
|
242 |
-
"reward":
|
243 |
-
"reward_std": 0.
|
244 |
-
"rewards/accuracy_reward": 0.
|
245 |
-
"rewards/format_reward": 0.
|
246 |
-
"rewards/relaxed_accuracy_reward": 0.
|
247 |
"step": 85
|
248 |
},
|
249 |
{
|
250 |
-
"completion_length":
|
251 |
"epoch": 0.6754221388367729,
|
252 |
-
"grad_norm":
|
253 |
-
"kl":
|
254 |
"learning_rate": 5.780464759928623e-06,
|
255 |
-
"loss": 0.
|
256 |
-
"reward":
|
257 |
-
"reward_std": 0.
|
258 |
-
"rewards/accuracy_reward": 0.
|
259 |
-
"rewards/format_reward": 0.
|
260 |
-
"rewards/relaxed_accuracy_reward": 0.
|
261 |
"step": 90
|
262 |
},
|
263 |
{
|
264 |
-
"completion_length":
|
265 |
"epoch": 0.7129455909943715,
|
266 |
-
"grad_norm":
|
267 |
-
"kl":
|
268 |
"learning_rate": 4.623911849714226e-06,
|
269 |
-
"loss": 0.
|
270 |
-
"reward":
|
271 |
-
"reward_std": 0.
|
272 |
-
"rewards/accuracy_reward": 0.
|
273 |
-
"rewards/format_reward": 0.
|
274 |
-
"rewards/relaxed_accuracy_reward": 0.
|
275 |
"step": 95
|
276 |
},
|
277 |
{
|
278 |
-
"completion_length":
|
279 |
"epoch": 0.7504690431519699,
|
280 |
-
"grad_norm":
|
281 |
-
"kl":
|
282 |
"learning_rate": 3.560895528440844e-06,
|
283 |
-
"loss": 0.
|
284 |
-
"reward":
|
285 |
-
"reward_std": 0.
|
286 |
-
"rewards/accuracy_reward": 0.
|
287 |
-
"rewards/format_reward": 0.
|
288 |
-
"rewards/relaxed_accuracy_reward": 0.
|
289 |
"step": 100
|
290 |
},
|
291 |
{
|
292 |
"epoch": 0.7504690431519699,
|
293 |
-
"eval_completion_length":
|
294 |
-
"eval_kl":
|
295 |
-
"eval_loss": 0.
|
296 |
-
"eval_reward":
|
297 |
-
"eval_reward_std": 0.
|
298 |
-
"eval_rewards/accuracy_reward": 0.
|
299 |
-
"eval_rewards/format_reward": 0.
|
300 |
-
"eval_rewards/relaxed_accuracy_reward": 0.
|
301 |
-
"eval_runtime":
|
302 |
-
"eval_samples_per_second":
|
303 |
-
"eval_steps_per_second": 0.
|
304 |
"step": 100
|
305 |
},
|
306 |
{
|
307 |
-
"completion_length":
|
308 |
"epoch": 0.7879924953095685,
|
309 |
-
"grad_norm":
|
310 |
-
"kl": 1.
|
311 |
"learning_rate": 2.6099108277934105e-06,
|
312 |
-
"loss": 0.
|
313 |
-
"reward":
|
314 |
-
"reward_std": 0.
|
315 |
-
"rewards/accuracy_reward": 0.
|
316 |
-
"rewards/format_reward": 0.
|
317 |
-
"rewards/relaxed_accuracy_reward": 0.
|
318 |
"step": 105
|
319 |
},
|
320 |
{
|
321 |
-
"completion_length":
|
322 |
"epoch": 0.8255159474671669,
|
323 |
-
"grad_norm":
|
324 |
-
"kl":
|
325 |
"learning_rate": 1.7875035823168641e-06,
|
326 |
-
"loss": 0.
|
327 |
-
"reward":
|
328 |
-
"reward_std": 0.
|
329 |
-
"rewards/accuracy_reward": 0.
|
330 |
-
"rewards/format_reward": 0.
|
331 |
-
"rewards/relaxed_accuracy_reward": 0.
|
332 |
"step": 110
|
333 |
},
|
334 |
{
|
335 |
-
"completion_length":
|
336 |
"epoch": 0.8630393996247655,
|
337 |
-
"grad_norm":
|
338 |
-
"kl":
|
339 |
"learning_rate": 1.1079825545001887e-06,
|
340 |
-
"loss": 0.
|
341 |
-
"reward":
|
342 |
-
"reward_std": 0.
|
343 |
-
"rewards/accuracy_reward": 0.
|
344 |
-
"rewards/format_reward": 0.
|
345 |
-
"rewards/relaxed_accuracy_reward": 0.
|
346 |
"step": 115
|
347 |
},
|
348 |
{
|
349 |
-
"completion_length":
|
350 |
"epoch": 0.900562851782364,
|
351 |
-
"grad_norm":
|
352 |
-
"kl": 1.
|
353 |
"learning_rate": 5.831704818578842e-07,
|
354 |
-
"loss": 0.
|
355 |
-
"reward":
|
356 |
-
"reward_std": 0.
|
357 |
-
"rewards/accuracy_reward": 0.
|
358 |
-
"rewards/format_reward": 0.
|
359 |
-
"rewards/relaxed_accuracy_reward": 0.
|
360 |
"step": 120
|
361 |
},
|
362 |
{
|
363 |
-
"completion_length":
|
364 |
"epoch": 0.9380863039399625,
|
365 |
-
"grad_norm":
|
366 |
-
"kl": 1.
|
367 |
"learning_rate": 2.2219837744959284e-07,
|
368 |
-
"loss": 0.
|
369 |
-
"reward":
|
370 |
-
"reward_std": 0.
|
371 |
-
"rewards/accuracy_reward": 0.
|
372 |
-
"rewards/format_reward": 0.
|
373 |
-
"rewards/relaxed_accuracy_reward": 0.
|
374 |
"step": 125
|
375 |
},
|
376 |
{
|
377 |
-
"completion_length":
|
378 |
"epoch": 0.975609756097561,
|
379 |
-
"grad_norm":
|
380 |
-
"kl": 1.
|
381 |
"learning_rate": 3.134666272774034e-08,
|
382 |
-
"loss": 0.
|
383 |
-
"reward":
|
384 |
-
"reward_std": 0.
|
385 |
-
"rewards/accuracy_reward": 0.
|
386 |
-
"rewards/format_reward": 0.
|
387 |
-
"rewards/relaxed_accuracy_reward": 0.
|
388 |
"step": 130
|
389 |
},
|
390 |
{
|
391 |
-
"completion_length":
|
392 |
"epoch": 0.99812382739212,
|
393 |
-
"kl": 1.
|
394 |
-
"reward":
|
395 |
-
"reward_std": 0.
|
396 |
-
"rewards/accuracy_reward": 0.
|
397 |
-
"rewards/format_reward": 0.
|
398 |
-
"rewards/relaxed_accuracy_reward": 0.
|
399 |
"step": 133,
|
400 |
"total_flos": 0.0,
|
401 |
-
"train_loss": 0.
|
402 |
-
"train_runtime":
|
403 |
-
"train_samples_per_second":
|
404 |
-
"train_steps_per_second": 0.
|
405 |
}
|
406 |
],
|
407 |
"logging_steps": 5,
|
|
|
9 |
"is_world_process_zero": true,
|
10 |
"log_history": [
|
11 |
{
|
12 |
+
"completion_length": 33.946875,
|
13 |
"epoch": 0.0375234521575985,
|
14 |
+
"grad_norm": 1.5592833757400513,
|
15 |
+
"kl": 0.013537660241127014,
|
16 |
"learning_rate": 7.1428571428571436e-06,
|
17 |
+
"loss": 0.0031,
|
18 |
+
"reward": 0.03671875,
|
19 |
+
"reward_std": 0.08486471865326166,
|
20 |
"rewards/accuracy_reward": 0.0158203125,
|
21 |
+
"rewards/format_reward": 0.0033203125,
|
22 |
+
"rewards/relaxed_accuracy_reward": 0.017578125,
|
23 |
"step": 5
|
24 |
},
|
25 |
{
|
26 |
+
"completion_length": 7.9525390625,
|
27 |
"epoch": 0.075046904315197,
|
28 |
+
"grad_norm": 26.73110008239746,
|
29 |
+
"kl": 17.372037601470947,
|
30 |
"learning_rate": 1.4285714285714287e-05,
|
31 |
+
"loss": 0.7047,
|
32 |
+
"reward": 0.413671875,
|
33 |
+
"reward_std": 0.19759062808007002,
|
34 |
+
"rewards/accuracy_reward": 0.1939453125,
|
35 |
+
"rewards/format_reward": 0.0005859375,
|
36 |
+
"rewards/relaxed_accuracy_reward": 0.219140625,
|
37 |
"step": 10
|
38 |
},
|
39 |
{
|
40 |
+
"completion_length": 29.3978515625,
|
41 |
"epoch": 0.1125703564727955,
|
42 |
+
"grad_norm": 12.636091232299805,
|
43 |
+
"kl": 2.432843017578125,
|
44 |
"learning_rate": 1.9996515418688493e-05,
|
45 |
+
"loss": 0.0908,
|
46 |
+
"reward": 0.30703125,
|
47 |
+
"reward_std": 0.24456602307036518,
|
48 |
+
"rewards/accuracy_reward": 0.140625,
|
49 |
+
"rewards/format_reward": 0.0,
|
50 |
+
"rewards/relaxed_accuracy_reward": 0.16640625,
|
51 |
"step": 15
|
52 |
},
|
53 |
{
|
54 |
+
"completion_length": 27.5798828125,
|
55 |
"epoch": 0.150093808630394,
|
56 |
+
"grad_norm": 21.91870880126953,
|
57 |
+
"kl": 1.253253173828125,
|
58 |
"learning_rate": 1.9874809871741877e-05,
|
59 |
+
"loss": 0.0532,
|
60 |
+
"reward": 0.5099609375,
|
61 |
+
"reward_std": 0.2849292915314436,
|
62 |
+
"rewards/accuracy_reward": 0.22578125,
|
63 |
+
"rewards/format_reward": 0.0,
|
64 |
+
"rewards/relaxed_accuracy_reward": 0.2841796875,
|
65 |
"step": 20
|
66 |
},
|
67 |
{
|
68 |
+
"completion_length": 9.8216796875,
|
69 |
"epoch": 0.18761726078799248,
|
70 |
+
"grad_norm": 21.709964752197266,
|
71 |
+
"kl": 1.0984130859375,
|
72 |
"learning_rate": 1.9581296124106682e-05,
|
73 |
+
"loss": 0.0578,
|
74 |
+
"reward": 0.5490234375,
|
75 |
+
"reward_std": 0.32321499213576316,
|
76 |
+
"rewards/accuracy_reward": 0.2373046875,
|
77 |
+
"rewards/format_reward": 0.0,
|
78 |
+
"rewards/relaxed_accuracy_reward": 0.31171875,
|
79 |
"step": 25
|
80 |
},
|
81 |
{
|
82 |
+
"completion_length": 22.2578125,
|
83 |
"epoch": 0.225140712945591,
|
84 |
+
"grad_norm": 165.1543426513672,
|
85 |
+
"kl": 3.56005859375,
|
86 |
"learning_rate": 1.912108091398988e-05,
|
87 |
+
"loss": 0.1402,
|
88 |
+
"reward": 0.4880859375,
|
89 |
+
"reward_std": 0.3410706129856408,
|
90 |
+
"rewards/accuracy_reward": 0.222265625,
|
91 |
+
"rewards/format_reward": 0.0,
|
92 |
+
"rewards/relaxed_accuracy_reward": 0.2658203125,
|
93 |
"step": 30
|
94 |
},
|
95 |
{
|
96 |
+
"completion_length": 20.027734375,
|
97 |
"epoch": 0.2626641651031895,
|
98 |
+
"grad_norm": 1127.5472412109375,
|
99 |
+
"kl": 6.7017822265625,
|
100 |
"learning_rate": 1.8502171357296144e-05,
|
101 |
+
"loss": 0.2507,
|
102 |
+
"reward": 0.387890625,
|
103 |
+
"reward_std": 0.3576823682524264,
|
104 |
+
"rewards/accuracy_reward": 0.1626953125,
|
105 |
+
"rewards/format_reward": 0.0,
|
106 |
+
"rewards/relaxed_accuracy_reward": 0.2251953125,
|
107 |
"step": 35
|
108 |
},
|
109 |
{
|
110 |
+
"completion_length": 29.49453125,
|
111 |
"epoch": 0.300187617260788,
|
112 |
+
"grad_norm": 25.603010177612305,
|
113 |
+
"kl": 1.09813232421875,
|
114 |
"learning_rate": 1.773533563475053e-05,
|
115 |
+
"loss": 0.0428,
|
116 |
+
"reward": 0.5158203125,
|
117 |
+
"reward_std": 0.36268206988461316,
|
118 |
+
"rewards/accuracy_reward": 0.22578125,
|
119 |
+
"rewards/format_reward": 0.0,
|
120 |
+
"rewards/relaxed_accuracy_reward": 0.2900390625,
|
121 |
"step": 40
|
122 |
},
|
123 |
{
|
124 |
+
"completion_length": 17.1138671875,
|
125 |
"epoch": 0.33771106941838647,
|
126 |
+
"grad_norm": 6.592025279998779,
|
127 |
+
"kl": 1.22408447265625,
|
128 |
"learning_rate": 1.6833915640265485e-05,
|
129 |
+
"loss": 0.0536,
|
130 |
+
"reward": 0.4765625,
|
131 |
+
"reward_std": 0.2620124928187579,
|
132 |
+
"rewards/accuracy_reward": 0.2068359375,
|
133 |
+
"rewards/format_reward": 0.0,
|
134 |
+
"rewards/relaxed_accuracy_reward": 0.2697265625,
|
135 |
"step": 45
|
136 |
},
|
137 |
{
|
138 |
+
"completion_length": 35.35703125,
|
139 |
"epoch": 0.37523452157598497,
|
140 |
+
"grad_norm": 3.753960132598877,
|
141 |
+
"kl": 1.41783447265625,
|
142 |
"learning_rate": 1.58135948502146e-05,
|
143 |
+
"loss": 0.0564,
|
144 |
+
"reward": 0.5171875,
|
145 |
+
"reward_std": 0.3441387979779392,
|
146 |
+
"rewards/accuracy_reward": 0.232421875,
|
147 |
+
"rewards/format_reward": 0.0,
|
148 |
+
"rewards/relaxed_accuracy_reward": 0.284765625,
|
149 |
"step": 50
|
150 |
},
|
151 |
{
|
152 |
+
"completion_length": 19.3244140625,
|
153 |
"epoch": 0.41275797373358347,
|
154 |
+
"grad_norm": 8.08942985534668,
|
155 |
+
"kl": 1.07467041015625,
|
156 |
"learning_rate": 1.4692125452370664e-05,
|
157 |
+
"loss": 0.038,
|
158 |
+
"reward": 0.5447265625,
|
159 |
+
"reward_std": 0.3162791552487761,
|
160 |
+
"rewards/accuracy_reward": 0.2421875,
|
161 |
+
"rewards/format_reward": 0.0,
|
162 |
+
"rewards/relaxed_accuracy_reward": 0.3025390625,
|
163 |
"step": 55
|
164 |
},
|
165 |
{
|
166 |
+
"completion_length": 13.803515625,
|
167 |
"epoch": 0.450281425891182,
|
168 |
+
"grad_norm": 6688.68994140625,
|
169 |
+
"kl": 11.382958984375,
|
170 |
"learning_rate": 1.348901948209167e-05,
|
171 |
+
"loss": 0.4542,
|
172 |
+
"reward": 0.5509765625,
|
173 |
+
"reward_std": 0.2770004874095321,
|
174 |
+
"rewards/accuracy_reward": 0.24453125,
|
175 |
+
"rewards/format_reward": 0.0,
|
176 |
+
"rewards/relaxed_accuracy_reward": 0.3064453125,
|
177 |
"step": 60
|
178 |
},
|
179 |
{
|
180 |
+
"completion_length": 10.150390625,
|
181 |
"epoch": 0.4878048780487805,
|
182 |
+
"grad_norm": 22.93476104736328,
|
183 |
+
"kl": 2.70247802734375,
|
184 |
"learning_rate": 1.2225209339563144e-05,
|
185 |
+
"loss": 0.1131,
|
186 |
+
"reward": 0.58359375,
|
187 |
+
"reward_std": 0.277045093011111,
|
188 |
+
"rewards/accuracy_reward": 0.2546875,
|
189 |
+
"rewards/format_reward": 0.0,
|
190 |
+
"rewards/relaxed_accuracy_reward": 0.32890625,
|
191 |
"step": 65
|
192 |
},
|
193 |
{
|
194 |
+
"completion_length": 9.9724609375,
|
195 |
"epoch": 0.525328330206379,
|
196 |
+
"grad_norm": 4.137301445007324,
|
197 |
+
"kl": 1.7162109375,
|
198 |
"learning_rate": 1.092268359463302e-05,
|
199 |
+
"loss": 0.0629,
|
200 |
+
"reward": 0.524609375,
|
201 |
+
"reward_std": 0.2948215680196881,
|
202 |
+
"rewards/accuracy_reward": 0.2169921875,
|
203 |
+
"rewards/format_reward": 0.0,
|
204 |
+
"rewards/relaxed_accuracy_reward": 0.3076171875,
|
205 |
"step": 70
|
206 |
},
|
207 |
{
|
208 |
+
"completion_length": 8.46484375,
|
209 |
"epoch": 0.5628517823639775,
|
210 |
+
"grad_norm": 5.097509860992432,
|
211 |
+
"kl": 2.0114501953125,
|
212 |
"learning_rate": 9.604104415737309e-06,
|
213 |
+
"loss": 0.0958,
|
214 |
+
"reward": 0.5404296875,
|
215 |
+
"reward_std": 0.3218729373533279,
|
216 |
+
"rewards/accuracy_reward": 0.2349609375,
|
217 |
+
"rewards/format_reward": 0.0,
|
218 |
+
"rewards/relaxed_accuracy_reward": 0.30546875,
|
219 |
"step": 75
|
220 |
},
|
221 |
{
|
222 |
+
"completion_length": 8.421875,
|
223 |
"epoch": 0.600375234521576,
|
224 |
+
"grad_norm": 6.521908283233643,
|
225 |
+
"kl": 2.06651611328125,
|
226 |
"learning_rate": 8.292413279130625e-06,
|
227 |
+
"loss": 0.0886,
|
228 |
+
"reward": 0.528125,
|
229 |
+
"reward_std": 0.32329283356666566,
|
230 |
+
"rewards/accuracy_reward": 0.222265625,
|
231 |
+
"rewards/format_reward": 0.0,
|
232 |
+
"rewards/relaxed_accuracy_reward": 0.305859375,
|
233 |
"step": 80
|
234 |
},
|
235 |
{
|
236 |
+
"completion_length": 8.27578125,
|
237 |
"epoch": 0.6378986866791745,
|
238 |
+
"grad_norm": 3.548377513885498,
|
239 |
+
"kl": 1.679052734375,
|
240 |
"learning_rate": 7.010431818542298e-06,
|
241 |
+
"loss": 0.0482,
|
242 |
+
"reward": 0.4205078125,
|
243 |
+
"reward_std": 0.3025111163035035,
|
244 |
+
"rewards/accuracy_reward": 0.1734375,
|
245 |
+
"rewards/format_reward": 0.0,
|
246 |
+
"rewards/relaxed_accuracy_reward": 0.2470703125,
|
247 |
"step": 85
|
248 |
},
|
249 |
{
|
250 |
+
"completion_length": 8.621484375,
|
251 |
"epoch": 0.6754221388367729,
|
252 |
+
"grad_norm": 11.26955509185791,
|
253 |
+
"kl": 2.084521484375,
|
254 |
"learning_rate": 5.780464759928623e-06,
|
255 |
+
"loss": 0.097,
|
256 |
+
"reward": 0.566015625,
|
257 |
+
"reward_std": 0.31781149725429714,
|
258 |
+
"rewards/accuracy_reward": 0.239453125,
|
259 |
+
"rewards/format_reward": 0.0,
|
260 |
+
"rewards/relaxed_accuracy_reward": 0.3265625,
|
261 |
"step": 90
|
262 |
},
|
263 |
{
|
264 |
+
"completion_length": 8.6640625,
|
265 |
"epoch": 0.7129455909943715,
|
266 |
+
"grad_norm": 15.46524715423584,
|
267 |
+
"kl": 2.5925048828125,
|
268 |
"learning_rate": 4.623911849714226e-06,
|
269 |
+
"loss": 0.0919,
|
270 |
+
"reward": 0.546484375,
|
271 |
+
"reward_std": 0.26226845681667327,
|
272 |
+
"rewards/accuracy_reward": 0.22421875,
|
273 |
+
"rewards/format_reward": 0.0,
|
274 |
+
"rewards/relaxed_accuracy_reward": 0.322265625,
|
275 |
"step": 95
|
276 |
},
|
277 |
{
|
278 |
+
"completion_length": 8.1521484375,
|
279 |
"epoch": 0.7504690431519699,
|
280 |
+
"grad_norm": 8.370095252990723,
|
281 |
+
"kl": 2.80860595703125,
|
282 |
"learning_rate": 3.560895528440844e-06,
|
283 |
+
"loss": 0.0876,
|
284 |
+
"reward": 0.6302734375,
|
285 |
+
"reward_std": 0.27545339791104195,
|
286 |
+
"rewards/accuracy_reward": 0.275,
|
287 |
+
"rewards/format_reward": 0.0,
|
288 |
+
"rewards/relaxed_accuracy_reward": 0.3552734375,
|
289 |
"step": 100
|
290 |
},
|
291 |
{
|
292 |
"epoch": 0.7504690431519699,
|
293 |
+
"eval_completion_length": 8.064242160278745,
|
294 |
+
"eval_kl": 1.6110899390243902,
|
295 |
+
"eval_loss": 0.05133385583758354,
|
296 |
+
"eval_reward": 0.5677264808362369,
|
297 |
+
"eval_reward_std": 0.37672981305716347,
|
298 |
+
"eval_rewards/accuracy_reward": 0.24401132404181183,
|
299 |
+
"eval_rewards/format_reward": 0.0,
|
300 |
+
"eval_rewards/relaxed_accuracy_reward": 0.3237151567944251,
|
301 |
+
"eval_runtime": 419.9905,
|
302 |
+
"eval_samples_per_second": 2.726,
|
303 |
+
"eval_steps_per_second": 0.683,
|
304 |
"step": 100
|
305 |
},
|
306 |
{
|
307 |
+
"completion_length": 8.2787109375,
|
308 |
"epoch": 0.7879924953095685,
|
309 |
+
"grad_norm": 5.566633224487305,
|
310 |
+
"kl": 1.6791015625,
|
311 |
"learning_rate": 2.6099108277934105e-06,
|
312 |
+
"loss": 0.0743,
|
313 |
+
"reward": 0.580859375,
|
314 |
+
"reward_std": 0.3424753251951188,
|
315 |
+
"rewards/accuracy_reward": 0.253125,
|
316 |
+
"rewards/format_reward": 0.0,
|
317 |
+
"rewards/relaxed_accuracy_reward": 0.327734375,
|
318 |
"step": 105
|
319 |
},
|
320 |
{
|
321 |
+
"completion_length": 8.3333984375,
|
322 |
"epoch": 0.8255159474671669,
|
323 |
+
"grad_norm": 4.05443000793457,
|
324 |
+
"kl": 2.09364013671875,
|
325 |
"learning_rate": 1.7875035823168641e-06,
|
326 |
+
"loss": 0.0748,
|
327 |
+
"reward": 0.5283203125,
|
328 |
+
"reward_std": 0.3331008433829993,
|
329 |
+
"rewards/accuracy_reward": 0.2318359375,
|
330 |
+
"rewards/format_reward": 0.0,
|
331 |
+
"rewards/relaxed_accuracy_reward": 0.296484375,
|
332 |
"step": 110
|
333 |
},
|
334 |
{
|
335 |
+
"completion_length": 8.5595703125,
|
336 |
"epoch": 0.8630393996247655,
|
337 |
+
"grad_norm": 20.012693405151367,
|
338 |
+
"kl": 2.143017578125,
|
339 |
"learning_rate": 1.1079825545001887e-06,
|
340 |
+
"loss": 0.0681,
|
341 |
+
"reward": 0.5373046875,
|
342 |
+
"reward_std": 0.3182933186646551,
|
343 |
+
"rewards/accuracy_reward": 0.23046875,
|
344 |
+
"rewards/format_reward": 0.0,
|
345 |
+
"rewards/relaxed_accuracy_reward": 0.3068359375,
|
346 |
"step": 115
|
347 |
},
|
348 |
{
|
349 |
+
"completion_length": 8.526171875,
|
350 |
"epoch": 0.900562851782364,
|
351 |
+
"grad_norm": 12.22225570678711,
|
352 |
+
"kl": 1.8458984375,
|
353 |
"learning_rate": 5.831704818578842e-07,
|
354 |
+
"loss": 0.0716,
|
355 |
+
"reward": 0.5498046875,
|
356 |
+
"reward_std": 0.31468736389651897,
|
357 |
+
"rewards/accuracy_reward": 0.237890625,
|
358 |
+
"rewards/format_reward": 0.0,
|
359 |
+
"rewards/relaxed_accuracy_reward": 0.3119140625,
|
360 |
"step": 120
|
361 |
},
|
362 |
{
|
363 |
+
"completion_length": 8.767578125,
|
364 |
"epoch": 0.9380863039399625,
|
365 |
+
"grad_norm": 4.848618507385254,
|
366 |
+
"kl": 1.8501708984375,
|
367 |
"learning_rate": 2.2219837744959284e-07,
|
368 |
+
"loss": 0.0649,
|
369 |
+
"reward": 0.5712890625,
|
370 |
+
"reward_std": 0.33388876002281903,
|
371 |
+
"rewards/accuracy_reward": 0.242578125,
|
372 |
+
"rewards/format_reward": 0.0,
|
373 |
+
"rewards/relaxed_accuracy_reward": 0.3287109375,
|
374 |
"step": 125
|
375 |
},
|
376 |
{
|
377 |
+
"completion_length": 8.5228515625,
|
378 |
"epoch": 0.975609756097561,
|
379 |
+
"grad_norm": 5.429755210876465,
|
380 |
+
"kl": 1.84400634765625,
|
381 |
"learning_rate": 3.134666272774034e-08,
|
382 |
+
"loss": 0.0876,
|
383 |
+
"reward": 0.564453125,
|
384 |
+
"reward_std": 0.31874394970946013,
|
385 |
+
"rewards/accuracy_reward": 0.2439453125,
|
386 |
+
"rewards/format_reward": 0.0,
|
387 |
+
"rewards/relaxed_accuracy_reward": 0.3205078125,
|
388 |
"step": 130
|
389 |
},
|
390 |
{
|
391 |
+
"completion_length": 8.5595703125,
|
392 |
"epoch": 0.99812382739212,
|
393 |
+
"kl": 1.7513020833333333,
|
394 |
+
"reward": 0.55078125,
|
395 |
+
"reward_std": 0.3167417396325618,
|
396 |
+
"rewards/accuracy_reward": 0.2275390625,
|
397 |
+
"rewards/format_reward": 0.0,
|
398 |
+
"rewards/relaxed_accuracy_reward": 0.3232421875,
|
399 |
"step": 133,
|
400 |
"total_flos": 0.0,
|
401 |
+
"train_loss": 0.11666255951964093,
|
402 |
+
"train_runtime": 146747.14,
|
403 |
+
"train_samples_per_second": 0.116,
|
404 |
+
"train_steps_per_second": 0.001
|
405 |
}
|
406 |
],
|
407 |
"logging_steps": 5,
|