File size: 27,019 Bytes
daaec58
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 3.0,
  "eval_steps": 500,
  "global_step": 249,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.060240963855421686,
      "grad_norm": 6.8344698068850365,
      "learning_rate": 6.25e-06,
      "logits/chosen": 0.20106494426727295,
      "logits/rejected": 0.19740548729896545,
      "logps/chosen": -7.359566688537598,
      "logps/rejected": -7.794281959533691,
      "loss": 0.7984,
      "rewards/accuracies": 0.30000001192092896,
      "rewards/chosen": 0.0005805142573080957,
      "rewards/margins": 0.00020884368859697133,
      "rewards/rejected": 0.0003716700884979218,
      "step": 5
    },
    {
      "epoch": 0.12048192771084337,
      "grad_norm": 11.077793932918919,
      "learning_rate": 9.998300813454981e-06,
      "logits/chosen": 0.14798372983932495,
      "logits/rejected": 0.14764083921909332,
      "logps/chosen": -9.343093872070312,
      "logps/rejected": -10.072733879089355,
      "loss": 0.804,
      "rewards/accuracies": 0.6500000357627869,
      "rewards/chosen": -0.12536723911762238,
      "rewards/margins": 0.08952564746141434,
      "rewards/rejected": -0.21489287912845612,
      "step": 10
    },
    {
      "epoch": 0.18072289156626506,
      "grad_norm": 22.64666641244041,
      "learning_rate": 9.979198225579968e-06,
      "logits/chosen": 0.09950681030750275,
      "logits/rejected": 0.10036235302686691,
      "logps/chosen": -15.521926879882812,
      "logps/rejected": -17.505037307739258,
      "loss": 0.9697,
      "rewards/accuracies": 0.5375000238418579,
      "rewards/chosen": -0.8507131338119507,
      "rewards/margins": 0.12971031665802002,
      "rewards/rejected": -0.9804234504699707,
      "step": 15
    },
    {
      "epoch": 0.24096385542168675,
      "grad_norm": 15.494932567768581,
      "learning_rate": 9.938950460359912e-06,
      "logits/chosen": 0.19120505452156067,
      "logits/rejected": 0.19118723273277283,
      "logps/chosen": -16.326614379882812,
      "logps/rejected": -21.430246353149414,
      "loss": 0.9114,
      "rewards/accuracies": 0.5625000596046448,
      "rewards/chosen": -0.8851631879806519,
      "rewards/margins": 0.45752018690109253,
      "rewards/rejected": -1.3426833152770996,
      "step": 20
    },
    {
      "epoch": 0.30120481927710846,
      "grad_norm": 7.705967374303585,
      "learning_rate": 9.877728438110645e-06,
      "logits/chosen": 0.2934933602809906,
      "logits/rejected": 0.2933184504508972,
      "logps/chosen": -16.846477508544922,
      "logps/rejected": -20.12131690979004,
      "loss": 0.9228,
      "rewards/accuracies": 0.543749988079071,
      "rewards/chosen": -0.9210986495018005,
      "rewards/margins": 0.357336163520813,
      "rewards/rejected": -1.2784347534179688,
      "step": 25
    },
    {
      "epoch": 0.3614457831325301,
      "grad_norm": 6.557578354068986,
      "learning_rate": 9.795792150593739e-06,
      "logits/chosen": 0.39821481704711914,
      "logits/rejected": 0.3969055414199829,
      "logps/chosen": -16.84841537475586,
      "logps/rejected": -20.08253288269043,
      "loss": 0.9177,
      "rewards/accuracies": 0.5625,
      "rewards/chosen": -0.9185025691986084,
      "rewards/margins": 0.3549797832965851,
      "rewards/rejected": -1.273482322692871,
      "step": 30
    },
    {
      "epoch": 0.42168674698795183,
      "grad_norm": 8.954398134887585,
      "learning_rate": 9.693489556908641e-06,
      "logits/chosen": 0.4319503605365753,
      "logits/rejected": 0.43394067883491516,
      "logps/chosen": -16.256404876708984,
      "logps/rejected": -19.0562686920166,
      "loss": 0.9048,
      "rewards/accuracies": 0.6187500357627869,
      "rewards/chosen": -0.8165277242660522,
      "rewards/margins": 0.362092524766922,
      "rewards/rejected": -1.1786202192306519,
      "step": 35
    },
    {
      "epoch": 0.4819277108433735,
      "grad_norm": 5.470944347532639,
      "learning_rate": 9.571255105813632e-06,
      "logits/chosen": 0.4045424163341522,
      "logits/rejected": 0.40858352184295654,
      "logps/chosen": -13.871018409729004,
      "logps/rejected": -18.067955017089844,
      "loss": 0.8794,
      "rewards/accuracies": 0.606249988079071,
      "rewards/chosen": -0.7279474139213562,
      "rewards/margins": 0.30724459886550903,
      "rewards/rejected": -1.0351920127868652,
      "step": 40
    },
    {
      "epoch": 0.5421686746987951,
      "grad_norm": 11.381498960351852,
      "learning_rate": 9.429607890750863e-06,
      "logits/chosen": 0.4048815369606018,
      "logits/rejected": 0.4054288864135742,
      "logps/chosen": -14.333499908447266,
      "logps/rejected": -17.505741119384766,
      "loss": 0.9043,
      "rewards/accuracies": 0.53125,
      "rewards/chosen": -0.6680309772491455,
      "rewards/margins": 0.30727389454841614,
      "rewards/rejected": -0.9753048419952393,
      "step": 45
    },
    {
      "epoch": 0.6024096385542169,
      "grad_norm": 7.883892084991124,
      "learning_rate": 9.269149445410545e-06,
      "logits/chosen": 0.4062591791152954,
      "logits/rejected": 0.4050535261631012,
      "logps/chosen": -12.784080505371094,
      "logps/rejected": -18.091772079467773,
      "loss": 0.822,
      "rewards/accuracies": 0.612500011920929,
      "rewards/chosen": -0.6254960298538208,
      "rewards/margins": 0.4347843527793884,
      "rewards/rejected": -1.060280442237854,
      "step": 50
    },
    {
      "epoch": 0.6626506024096386,
      "grad_norm": 6.279250929917772,
      "learning_rate": 9.09056118919587e-06,
      "logits/chosen": 0.33784574270248413,
      "logits/rejected": 0.3434804677963257,
      "logps/chosen": -12.477110862731934,
      "logps/rejected": -20.124866485595703,
      "loss": 0.7625,
      "rewards/accuracies": 0.6375000476837158,
      "rewards/chosen": -0.6189417839050293,
      "rewards/margins": 0.6243987083435059,
      "rewards/rejected": -1.2433404922485352,
      "step": 55
    },
    {
      "epoch": 0.7228915662650602,
      "grad_norm": 8.046915696828538,
      "learning_rate": 8.894601533437e-06,
      "logits/chosen": 0.3401602506637573,
      "logits/rejected": 0.34562352299690247,
      "logps/chosen": -14.957855224609375,
      "logps/rejected": -19.155860900878906,
      "loss": 0.931,
      "rewards/accuracies": 0.53125,
      "rewards/chosen": -0.7567940950393677,
      "rewards/margins": 0.40305179357528687,
      "rewards/rejected": -1.1598458290100098,
      "step": 60
    },
    {
      "epoch": 0.7831325301204819,
      "grad_norm": 7.80904642662881,
      "learning_rate": 8.682102660643196e-06,
      "logits/chosen": 0.4684681296348572,
      "logits/rejected": 0.4649001955986023,
      "logps/chosen": -13.015806198120117,
      "logps/rejected": -16.524856567382812,
      "loss": 0.8736,
      "rewards/accuracies": 0.574999988079071,
      "rewards/chosen": -0.6119887232780457,
      "rewards/margins": 0.3549143671989441,
      "rewards/rejected": -0.9669030904769897,
      "step": 65
    },
    {
      "epoch": 0.8433734939759037,
      "grad_norm": 5.157128287254778,
      "learning_rate": 8.453966990470656e-06,
      "logits/chosen": 0.5199801325798035,
      "logits/rejected": 0.5173701047897339,
      "logps/chosen": -14.613012313842773,
      "logps/rejected": -19.30719757080078,
      "loss": 0.8357,
      "rewards/accuracies": 0.6375000476837158,
      "rewards/chosen": -0.712582528591156,
      "rewards/margins": 0.5169816017150879,
      "rewards/rejected": -1.2295641899108887,
      "step": 70
    },
    {
      "epoch": 0.9036144578313253,
      "grad_norm": 7.350972551804275,
      "learning_rate": 8.211163347414005e-06,
      "logits/chosen": 0.5550334453582764,
      "logits/rejected": 0.5541264414787292,
      "logps/chosen": -14.093545913696289,
      "logps/rejected": -18.224023818969727,
      "loss": 0.8537,
      "rewards/accuracies": 0.625,
      "rewards/chosen": -0.7159973978996277,
      "rewards/margins": 0.4007648527622223,
      "rewards/rejected": -1.1167622804641724,
      "step": 75
    },
    {
      "epoch": 0.963855421686747,
      "grad_norm": 6.599463284958903,
      "learning_rate": 7.95472284649615e-06,
      "logits/chosen": 0.5369991064071655,
      "logits/rejected": 0.5401480197906494,
      "logps/chosen": -14.460978507995605,
      "logps/rejected": -19.939844131469727,
      "loss": 0.89,
      "rewards/accuracies": 0.5375000238418579,
      "rewards/chosen": -0.7761189937591553,
      "rewards/margins": 0.39927852153778076,
      "rewards/rejected": -1.1753976345062256,
      "step": 80
    },
    {
      "epoch": 1.0240963855421688,
      "grad_norm": 16.015900049544836,
      "learning_rate": 7.685734514428767e-06,
      "logits/chosen": 0.42177876830101013,
      "logits/rejected": 0.42529550194740295,
      "logps/chosen": -14.51510238647461,
      "logps/rejected": -17.642459869384766,
      "loss": 0.815,
      "rewards/accuracies": 0.6187499761581421,
      "rewards/chosen": -0.6715702414512634,
      "rewards/margins": 0.41218772530555725,
      "rewards/rejected": -1.0837578773498535,
      "step": 85
    },
    {
      "epoch": 1.0843373493975903,
      "grad_norm": 4.065830749870838,
      "learning_rate": 7.405340664838994e-06,
      "logits/chosen": 0.23701924085617065,
      "logits/rejected": 0.24219843745231628,
      "logps/chosen": -11.591026306152344,
      "logps/rejected": -18.730134963989258,
      "loss": 0.6683,
      "rewards/accuracies": 0.7437500357627869,
      "rewards/chosen": -0.41484612226486206,
      "rewards/margins": 0.7188076376914978,
      "rewards/rejected": -1.1336537599563599,
      "step": 90
    },
    {
      "epoch": 1.144578313253012,
      "grad_norm": 6.737624906291123,
      "learning_rate": 7.114732047202433e-06,
      "logits/chosen": 0.06327857077121735,
      "logits/rejected": 0.06292692571878433,
      "logps/chosen": -11.42073917388916,
      "logps/rejected": -19.14717674255371,
      "loss": 0.6966,
      "rewards/accuracies": 0.7124999761581421,
      "rewards/chosen": -0.4310266673564911,
      "rewards/margins": 0.7305406332015991,
      "rewards/rejected": -1.1615674495697021,
      "step": 95
    },
    {
      "epoch": 1.2048192771084336,
      "grad_norm": 7.082800328690027,
      "learning_rate": 6.815142790083473e-06,
      "logits/chosen": -0.01345985010266304,
      "logits/rejected": -0.005416684318333864,
      "logps/chosen": -12.335067749023438,
      "logps/rejected": -20.66636848449707,
      "loss": 0.7393,
      "rewards/accuracies": 0.6875,
      "rewards/chosen": -0.5101935863494873,
      "rewards/margins": 0.8145975470542908,
      "rewards/rejected": -1.3247911930084229,
      "step": 100
    },
    {
      "epoch": 1.2650602409638554,
      "grad_norm": 5.036833862784185,
      "learning_rate": 6.507845160157476e-06,
      "logits/chosen": -0.03185756877064705,
      "logits/rejected": -0.026032937690615654,
      "logps/chosen": -12.31982421875,
      "logps/rejected": -19.825225830078125,
      "loss": 0.6988,
      "rewards/accuracies": 0.737500011920929,
      "rewards/chosen": -0.5740019083023071,
      "rewards/margins": 0.696465790271759,
      "rewards/rejected": -1.270467758178711,
      "step": 105
    },
    {
      "epoch": 1.3253012048192772,
      "grad_norm": 5.560017374017693,
      "learning_rate": 6.1941441592717564e-06,
      "logits/chosen": -0.028415212407708168,
      "logits/rejected": -0.03128795325756073,
      "logps/chosen": -12.783760070800781,
      "logps/rejected": -21.475841522216797,
      "loss": 0.7107,
      "rewards/accuracies": 0.6875000596046448,
      "rewards/chosen": -0.6131666302680969,
      "rewards/margins": 0.7542582750320435,
      "rewards/rejected": -1.3674249649047852,
      "step": 110
    },
    {
      "epoch": 1.3855421686746987,
      "grad_norm": 4.679243672883931,
      "learning_rate": 5.875371982489959e-06,
      "logits/chosen": -0.03955061361193657,
      "logits/rejected": -0.03935273364186287,
      "logps/chosen": -12.692768096923828,
      "logps/rejected": -20.340290069580078,
      "loss": 0.7095,
      "rewards/accuracies": 0.625,
      "rewards/chosen": -0.5668330788612366,
      "rewards/margins": 0.7566565275192261,
      "rewards/rejected": -1.3234896659851074,
      "step": 115
    },
    {
      "epoch": 1.4457831325301205,
      "grad_norm": 4.303668638983521,
      "learning_rate": 5.55288236065495e-06,
      "logits/chosen": -0.03517518565058708,
      "logits/rejected": -0.03254169970750809,
      "logps/chosen": -12.00464153289795,
      "logps/rejected": -20.043167114257812,
      "loss": 0.6784,
      "rewards/accuracies": 0.6937500238418579,
      "rewards/chosen": -0.45237845182418823,
      "rewards/margins": 0.8140364289283752,
      "rewards/rejected": -1.2664148807525635,
      "step": 120
    },
    {
      "epoch": 1.5060240963855422,
      "grad_norm": 9.89407814357214,
      "learning_rate": 5.228044811495632e-06,
      "logits/chosen": 0.011906255967915058,
      "logits/rejected": 0.012696187943220139,
      "logps/chosen": -11.616506576538086,
      "logps/rejected": -20.238014221191406,
      "loss": 0.7106,
      "rewards/accuracies": 0.6875,
      "rewards/chosen": -0.44120335578918457,
      "rewards/margins": 0.8242735266685486,
      "rewards/rejected": -1.2654768228530884,
      "step": 125
    },
    {
      "epoch": 1.5662650602409638,
      "grad_norm": 4.549821099257562,
      "learning_rate": 4.9022388236915306e-06,
      "logits/chosen": 0.04570060223340988,
      "logits/rejected": 0.04913484305143356,
      "logps/chosen": -10.814414978027344,
      "logps/rejected": -20.540760040283203,
      "loss": 0.6321,
      "rewards/accuracies": 0.7749999761581421,
      "rewards/chosen": -0.38592851161956787,
      "rewards/margins": 0.9144089818000793,
      "rewards/rejected": -1.300337553024292,
      "step": 130
    },
    {
      "epoch": 1.6265060240963856,
      "grad_norm": 6.159969465605162,
      "learning_rate": 4.57684799859372e-06,
      "logits/chosen": 0.051878318190574646,
      "logits/rejected": 0.049009501934051514,
      "logps/chosen": -10.995405197143555,
      "logps/rejected": -20.4885196685791,
      "loss": 0.6566,
      "rewards/accuracies": 0.7062500715255737,
      "rewards/chosen": -0.399277925491333,
      "rewards/margins": 0.9130607843399048,
      "rewards/rejected": -1.3123387098312378,
      "step": 135
    },
    {
      "epoch": 1.6867469879518073,
      "grad_norm": 7.807608973663333,
      "learning_rate": 4.253254174480462e-06,
      "logits/chosen": 0.06293821334838867,
      "logits/rejected": 0.06775335967540741,
      "logps/chosen": -12.86220932006836,
      "logps/rejected": -24.761363983154297,
      "loss": 0.6637,
      "rewards/accuracies": 0.7437500357627869,
      "rewards/chosen": -0.5232489705085754,
      "rewards/margins": 1.1066060066223145,
      "rewards/rejected": -1.6298549175262451,
      "step": 140
    },
    {
      "epoch": 1.7469879518072289,
      "grad_norm": 6.450758798029886,
      "learning_rate": 3.932831558300074e-06,
      "logits/chosen": 0.03402797505259514,
      "logits/rejected": 0.03250021114945412,
      "logps/chosen": -12.925631523132324,
      "logps/rejected": -19.254789352416992,
      "loss": 0.7336,
      "rewards/accuracies": 0.71875,
      "rewards/chosen": -0.49226850271224976,
      "rewards/margins": 0.6960192322731018,
      "rewards/rejected": -1.1882877349853516,
      "step": 145
    },
    {
      "epoch": 1.8072289156626506,
      "grad_norm": 4.3607756550439625,
      "learning_rate": 3.6169408898217973e-06,
      "logits/chosen": 0.039087243378162384,
      "logits/rejected": 0.04352856054902077,
      "logps/chosen": -12.372940063476562,
      "logps/rejected": -20.964872360229492,
      "loss": 0.7091,
      "rewards/accuracies": 0.6875,
      "rewards/chosen": -0.5094107389450073,
      "rewards/margins": 0.7860168218612671,
      "rewards/rejected": -1.295427680015564,
      "step": 150
    },
    {
      "epoch": 1.8674698795180724,
      "grad_norm": 4.665085160330557,
      "learning_rate": 3.306923662977789e-06,
      "logits/chosen": 0.06090332567691803,
      "logits/rejected": 0.06559460610151291,
      "logps/chosen": -12.070850372314453,
      "logps/rejected": -21.903215408325195,
      "loss": 0.6518,
      "rewards/accuracies": 0.7062500715255737,
      "rewards/chosen": -0.46340861916542053,
      "rewards/margins": 1.006162405014038,
      "rewards/rejected": -1.4695709943771362,
      "step": 155
    },
    {
      "epoch": 1.927710843373494,
      "grad_norm": 6.373734799601269,
      "learning_rate": 3.0040964289364618e-06,
      "logits/chosen": 0.10437053442001343,
      "logits/rejected": 0.09873737394809723,
      "logps/chosen": -11.04707145690918,
      "logps/rejected": -20.31425666809082,
      "loss": 0.6583,
      "rewards/accuracies": 0.71875,
      "rewards/chosen": -0.3955325484275818,
      "rewards/margins": 0.8207127451896667,
      "rewards/rejected": -1.2162452936172485,
      "step": 160
    },
    {
      "epoch": 1.9879518072289155,
      "grad_norm": 6.76579535839464,
      "learning_rate": 2.7097452051003375e-06,
      "logits/chosen": 0.12389400601387024,
      "logits/rejected": 0.12251937389373779,
      "logps/chosen": -10.690954208374023,
      "logps/rejected": -17.01201057434082,
      "loss": 0.7035,
      "rewards/accuracies": 0.6625000238418579,
      "rewards/chosen": -0.3639286160469055,
      "rewards/margins": 0.6530975103378296,
      "rewards/rejected": -1.0170261859893799,
      "step": 165
    },
    {
      "epoch": 2.0481927710843375,
      "grad_norm": 3.0286043203118043,
      "learning_rate": 2.4251200137717545e-06,
      "logits/chosen": 0.12429633736610413,
      "logits/rejected": 0.13369517028331757,
      "logps/chosen": -9.702492713928223,
      "logps/rejected": -18.70101547241211,
      "loss": 0.5816,
      "rewards/accuracies": 0.7687499523162842,
      "rewards/chosen": -0.19249771535396576,
      "rewards/margins": 0.9518150091171265,
      "rewards/rejected": -1.144312858581543,
      "step": 170
    },
    {
      "epoch": 2.108433734939759,
      "grad_norm": 3.4650296160280996,
      "learning_rate": 2.151429573679084e-06,
      "logits/chosen": 0.09276320040225983,
      "logits/rejected": 0.08381576836109161,
      "logps/chosen": -7.954525470733643,
      "logps/rejected": -18.878183364868164,
      "loss": 0.5543,
      "rewards/accuracies": 0.7937500476837158,
      "rewards/chosen": -0.13666492700576782,
      "rewards/margins": 1.0228486061096191,
      "rewards/rejected": -1.1595135927200317,
      "step": 175
    },
    {
      "epoch": 2.1686746987951806,
      "grad_norm": 4.009672492571936,
      "learning_rate": 1.8898361669069497e-06,
      "logits/chosen": 0.012567025609314442,
      "logits/rejected": 0.013043450191617012,
      "logps/chosen": -8.962043762207031,
      "logps/rejected": -20.484146118164062,
      "loss": 0.5452,
      "rewards/accuracies": 0.7937500476837158,
      "rewards/chosen": -0.16434553265571594,
      "rewards/margins": 1.1564074754714966,
      "rewards/rejected": -1.3207528591156006,
      "step": 180
    },
    {
      "epoch": 2.2289156626506026,
      "grad_norm": 4.442304788111209,
      "learning_rate": 1.6414507030291249e-06,
      "logits/chosen": -0.06221061572432518,
      "logits/rejected": -0.06934291869401932,
      "logps/chosen": -9.080927848815918,
      "logps/rejected": -23.444576263427734,
      "loss": 0.5333,
      "rewards/accuracies": 0.831250011920929,
      "rewards/chosen": -0.18415439128875732,
      "rewards/margins": 1.3432217836380005,
      "rewards/rejected": -1.5273760557174683,
      "step": 185
    },
    {
      "epoch": 2.289156626506024,
      "grad_norm": 4.2271531800043585,
      "learning_rate": 1.4073280014052077e-06,
      "logits/chosen": -0.14234301447868347,
      "logits/rejected": -0.13763678073883057,
      "logps/chosen": -10.012588500976562,
      "logps/rejected": -22.59408187866211,
      "loss": 0.5441,
      "rewards/accuracies": 0.8187500238418579,
      "rewards/chosen": -0.22738967835903168,
      "rewards/margins": 1.3349870443344116,
      "rewards/rejected": -1.562376618385315,
      "step": 190
    },
    {
      "epoch": 2.3493975903614457,
      "grad_norm": 3.8404210034307105,
      "learning_rate": 1.1884623116758121e-06,
      "logits/chosen": -0.17886070907115936,
      "logits/rejected": -0.17571109533309937,
      "logps/chosen": -9.892772674560547,
      "logps/rejected": -26.796541213989258,
      "loss": 0.4638,
      "rewards/accuracies": 0.862500011920929,
      "rewards/chosen": -0.21987608075141907,
      "rewards/margins": 1.6867713928222656,
      "rewards/rejected": -1.9066474437713623,
      "step": 195
    },
    {
      "epoch": 2.4096385542168672,
      "grad_norm": 4.953359294501781,
      "learning_rate": 9.857830914793827e-07,
      "logits/chosen": -0.20311155915260315,
      "logits/rejected": -0.19799628853797913,
      "logps/chosen": -10.871113777160645,
      "logps/rejected": -29.490276336669922,
      "loss": 0.4925,
      "rewards/accuracies": 0.8374999761581421,
      "rewards/chosen": -0.3226281404495239,
      "rewards/margins": 1.7889926433563232,
      "rewards/rejected": -2.1116209030151367,
      "step": 200
    },
    {
      "epoch": 2.4698795180722892,
      "grad_norm": 4.64517618032467,
      "learning_rate": 8.001510593213946e-07,
      "logits/chosen": -0.22599855065345764,
      "logits/rejected": -0.22137105464935303,
      "logps/chosen": -8.583234786987305,
      "logps/rejected": -25.918903350830078,
      "loss": 0.4905,
      "rewards/accuracies": 0.862500011920929,
      "rewards/chosen": -0.14613361656665802,
      "rewards/margins": 1.7042248249053955,
      "rewards/rejected": -1.850358247756958,
      "step": 205
    },
    {
      "epoch": 2.5301204819277108,
      "grad_norm": 3.152112838442371,
      "learning_rate": 6.323545393582847e-07,
      "logits/chosen": -0.22711718082427979,
      "logits/rejected": -0.2298433780670166,
      "logps/chosen": -9.491654396057129,
      "logps/rejected": -22.80984115600586,
      "loss": 0.5301,
      "rewards/accuracies": 0.8187500238418579,
      "rewards/chosen": -0.1725078970193863,
      "rewards/margins": 1.4063202142715454,
      "rewards/rejected": -1.57882821559906,
      "step": 210
    },
    {
      "epoch": 2.5903614457831328,
      "grad_norm": 4.4630924013044115,
      "learning_rate": 4.831061136186787e-07,
      "logits/chosen": -0.23690445721149445,
      "logits/rejected": -0.23807355761528015,
      "logps/chosen": -8.89923095703125,
      "logps/rejected": -23.76407814025879,
      "loss": 0.5269,
      "rewards/accuracies": 0.8250000476837158,
      "rewards/chosen": -0.21625632047653198,
      "rewards/margins": 1.421187400817871,
      "rewards/rejected": -1.6374436616897583,
      "step": 215
    },
    {
      "epoch": 2.6506024096385543,
      "grad_norm": 5.625801336718983,
      "learning_rate": 3.53039595878959e-07,
      "logits/chosen": -0.24186240136623383,
      "logits/rejected": -0.24021676182746887,
      "logps/chosen": -10.780243873596191,
      "logps/rejected": -24.856246948242188,
      "loss": 0.6027,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": -0.37535130977630615,
      "rewards/margins": 1.3886429071426392,
      "rewards/rejected": -1.7639942169189453,
      "step": 220
    },
    {
      "epoch": 2.710843373493976,
      "grad_norm": 5.374747811868763,
      "learning_rate": 2.4270734004424643e-07,
      "logits/chosen": -0.24513426423072815,
      "logits/rejected": -0.24106302857398987,
      "logps/chosen": -9.546283721923828,
      "logps/rejected": -26.042221069335938,
      "loss": 0.4833,
      "rewards/accuracies": 0.8625000715255737,
      "rewards/chosen": -0.1909545361995697,
      "rewards/margins": 1.6568984985351562,
      "rewards/rejected": -1.8478529453277588,
      "step": 225
    },
    {
      "epoch": 2.7710843373493974,
      "grad_norm": 4.71023027186897,
      "learning_rate": 1.5257789446526172e-07,
      "logits/chosen": -0.24886594712734222,
      "logits/rejected": -0.24283945560455322,
      "logps/chosen": -9.447517395019531,
      "logps/rejected": -25.348175048828125,
      "loss": 0.521,
      "rewards/accuracies": 0.8374999761581421,
      "rewards/chosen": -0.2334991991519928,
      "rewards/margins": 1.5295912027359009,
      "rewards/rejected": -1.7630903720855713,
      "step": 230
    },
    {
      "epoch": 2.8313253012048194,
      "grad_norm": 3.618285262250259,
      "learning_rate": 8.303401215251583e-08,
      "logits/chosen": -0.23923178017139435,
      "logits/rejected": -0.2397070974111557,
      "logps/chosen": -9.20031452178955,
      "logps/rejected": -28.882890701293945,
      "loss": 0.4402,
      "rewards/accuracies": 0.887499988079071,
      "rewards/chosen": -0.24343208968639374,
      "rewards/margins": 1.849422812461853,
      "rewards/rejected": -2.0928549766540527,
      "step": 235
    },
    {
      "epoch": 2.891566265060241,
      "grad_norm": 4.639163710743134,
      "learning_rate": 3.437102533785541e-08,
      "logits/chosen": -0.24758636951446533,
      "logits/rejected": -0.2467891424894333,
      "logps/chosen": -7.994501113891602,
      "logps/rejected": -24.78835678100586,
      "loss": 0.4808,
      "rewards/accuracies": 0.8187500238418579,
      "rewards/chosen": -0.12487827241420746,
      "rewards/margins": 1.5749868154525757,
      "rewards/rejected": -1.6998651027679443,
      "step": 240
    },
    {
      "epoch": 2.9518072289156625,
      "grad_norm": 3.9125816690429254,
      "learning_rate": 6.7955912861095155e-09,
      "logits/chosen": -0.2559281587600708,
      "logits/rejected": -0.253335177898407,
      "logps/chosen": -8.652589797973633,
      "logps/rejected": -21.75301170349121,
      "loss": 0.4927,
      "rewards/accuracies": 0.84375,
      "rewards/chosen": -0.1405271738767624,
      "rewards/margins": 1.3281803131103516,
      "rewards/rejected": -1.46870756149292,
      "step": 245
    },
    {
      "epoch": 3.0,
      "step": 249,
      "total_flos": 11469721436160.0,
      "train_loss": 0.693384567896525,
      "train_runtime": 1185.9878,
      "train_samples_per_second": 6.716,
      "train_steps_per_second": 0.21
    }
  ],
  "logging_steps": 5,
  "max_steps": 249,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 3,
  "save_steps": 500,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 11469721436160.0,
  "train_batch_size": 2,
  "trial_name": null,
  "trial_params": null
}