File size: 38,951 Bytes
0bf4bb9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 1.9946666666666668,
  "eval_steps": 500,
  "global_step": 374,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.02666666666666667,
      "grad_norm": 91.80353546142578,
      "learning_rate": 1.3157894736842104e-08,
      "logits/chosen": 0.05448655039072037,
      "logits/rejected": 0.061774831265211105,
      "logps/chosen": -70.88352966308594,
      "logps/rejected": -68.78348541259766,
      "loss": 0.6945,
      "rewards/accuracies": 0.3499999940395355,
      "rewards/chosen": -0.007239150814712048,
      "rewards/margins": 0.00464630126953125,
      "rewards/rejected": -0.011885452084243298,
      "step": 5
    },
    {
      "epoch": 0.05333333333333334,
      "grad_norm": 98.0750732421875,
      "learning_rate": 2.6315789473684208e-08,
      "logits/chosen": -0.09558521211147308,
      "logits/rejected": -0.09429871290922165,
      "logps/chosen": -47.07535934448242,
      "logps/rejected": -57.48785400390625,
      "loss": 0.6855,
      "rewards/accuracies": 0.5,
      "rewards/chosen": 0.04929858446121216,
      "rewards/margins": 0.017685014754533768,
      "rewards/rejected": 0.03161356970667839,
      "step": 10
    },
    {
      "epoch": 0.08,
      "grad_norm": 86.46255493164062,
      "learning_rate": 3.947368421052631e-08,
      "logits/chosen": -0.0949799194931984,
      "logits/rejected": -0.08726556599140167,
      "logps/chosen": -58.88152313232422,
      "logps/rejected": -69.49443054199219,
      "loss": 0.6504,
      "rewards/accuracies": 0.6499999761581421,
      "rewards/chosen": 0.033757805824279785,
      "rewards/margins": 0.1223212480545044,
      "rewards/rejected": -0.08856344223022461,
      "step": 15
    },
    {
      "epoch": 0.10666666666666667,
      "grad_norm": 85.12960052490234,
      "learning_rate": 5.2631578947368416e-08,
      "logits/chosen": 0.04131064563989639,
      "logits/rejected": 0.047180645167827606,
      "logps/chosen": -64.10095977783203,
      "logps/rejected": -71.49341583251953,
      "loss": 0.622,
      "rewards/accuracies": 0.75,
      "rewards/chosen": 0.08759395778179169,
      "rewards/margins": 0.22742386162281036,
      "rewards/rejected": -0.13982990384101868,
      "step": 20
    },
    {
      "epoch": 0.13333333333333333,
      "grad_norm": 72.46548461914062,
      "learning_rate": 6.578947368421053e-08,
      "logits/chosen": 0.07103381305932999,
      "logits/rejected": 0.06991696357727051,
      "logps/chosen": -71.91819763183594,
      "logps/rejected": -76.61006927490234,
      "loss": 0.6107,
      "rewards/accuracies": 0.6000000238418579,
      "rewards/chosen": 0.14151427149772644,
      "rewards/margins": 0.2075047492980957,
      "rewards/rejected": -0.06599047034978867,
      "step": 25
    },
    {
      "epoch": 0.16,
      "grad_norm": 71.43992614746094,
      "learning_rate": 7.894736842105262e-08,
      "logits/chosen": -0.0016497105825692415,
      "logits/rejected": -0.003721952438354492,
      "logps/chosen": -64.13134765625,
      "logps/rejected": -59.88336181640625,
      "loss": 0.5604,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.14888896048069,
      "rewards/margins": 0.5770025253295898,
      "rewards/rejected": -0.4281136095523834,
      "step": 30
    },
    {
      "epoch": 0.18666666666666668,
      "grad_norm": 86.11872100830078,
      "learning_rate": 9.210526315789473e-08,
      "logits/chosen": 0.09134344756603241,
      "logits/rejected": 0.08688181638717651,
      "logps/chosen": -74.20964050292969,
      "logps/rejected": -66.22145080566406,
      "loss": 0.5322,
      "rewards/accuracies": 0.6000000238418579,
      "rewards/chosen": 0.3988303542137146,
      "rewards/margins": 0.3464100956916809,
      "rewards/rejected": 0.052420247346162796,
      "step": 35
    },
    {
      "epoch": 0.21333333333333335,
      "grad_norm": 83.05272674560547,
      "learning_rate": 1e-07,
      "logits/chosen": -0.051578253507614136,
      "logits/rejected": -0.04844808578491211,
      "logps/chosen": -91.24781036376953,
      "logps/rejected": -73.64341735839844,
      "loss": 0.5272,
      "rewards/accuracies": 0.6499999761581421,
      "rewards/chosen": 0.11290506273508072,
      "rewards/margins": 0.25997716188430786,
      "rewards/rejected": -0.14707207679748535,
      "step": 40
    },
    {
      "epoch": 0.24,
      "grad_norm": 74.43838500976562,
      "learning_rate": 1e-07,
      "logits/chosen": -0.06837300211191177,
      "logits/rejected": -0.06631087511777878,
      "logps/chosen": -60.6038818359375,
      "logps/rejected": -62.58821487426758,
      "loss": 0.5138,
      "rewards/accuracies": 0.8500000238418579,
      "rewards/chosen": 0.5276230573654175,
      "rewards/margins": 0.8340091705322266,
      "rewards/rejected": -0.30638617277145386,
      "step": 45
    },
    {
      "epoch": 0.26666666666666666,
      "grad_norm": 62.49097442626953,
      "learning_rate": 1e-07,
      "logits/chosen": -0.00017045289860107005,
      "logits/rejected": -0.0001672551006777212,
      "logps/chosen": -82.781005859375,
      "logps/rejected": -82.09158325195312,
      "loss": 0.4674,
      "rewards/accuracies": 0.6499999761581421,
      "rewards/chosen": 0.34503522515296936,
      "rewards/margins": 1.0920366048812866,
      "rewards/rejected": -0.7470014691352844,
      "step": 50
    },
    {
      "epoch": 0.29333333333333333,
      "grad_norm": 80.10330200195312,
      "learning_rate": 1e-07,
      "logits/chosen": -0.02159346453845501,
      "logits/rejected": -0.023507962003350258,
      "logps/chosen": -66.51358032226562,
      "logps/rejected": -65.78565979003906,
      "loss": 0.5481,
      "rewards/accuracies": 0.6499999761581421,
      "rewards/chosen": 0.3966673016548157,
      "rewards/margins": 0.8023613095283508,
      "rewards/rejected": -0.4056939482688904,
      "step": 55
    },
    {
      "epoch": 0.32,
      "grad_norm": 67.92163848876953,
      "learning_rate": 1e-07,
      "logits/chosen": -0.02508384920656681,
      "logits/rejected": -0.023196673020720482,
      "logps/chosen": -69.94526672363281,
      "logps/rejected": -54.190528869628906,
      "loss": 0.5126,
      "rewards/accuracies": 0.6000000238418579,
      "rewards/chosen": 0.5231302976608276,
      "rewards/margins": 0.935653805732727,
      "rewards/rejected": -0.41252344846725464,
      "step": 60
    },
    {
      "epoch": 0.3466666666666667,
      "grad_norm": 62.79459762573242,
      "learning_rate": 1e-07,
      "logits/chosen": -0.14604972302913666,
      "logits/rejected": -0.1421947032213211,
      "logps/chosen": -71.80470275878906,
      "logps/rejected": -62.09037399291992,
      "loss": 0.5224,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": 0.6033689379692078,
      "rewards/margins": 0.9922472238540649,
      "rewards/rejected": -0.3888782560825348,
      "step": 65
    },
    {
      "epoch": 0.37333333333333335,
      "grad_norm": 63.457252502441406,
      "learning_rate": 1e-07,
      "logits/chosen": 0.06079145520925522,
      "logits/rejected": 0.07840003073215485,
      "logps/chosen": -71.81758117675781,
      "logps/rejected": -85.1531982421875,
      "loss": 0.39,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": 0.24788689613342285,
      "rewards/margins": 1.4833877086639404,
      "rewards/rejected": -1.2355008125305176,
      "step": 70
    },
    {
      "epoch": 0.4,
      "grad_norm": 69.04164123535156,
      "learning_rate": 1e-07,
      "logits/chosen": 0.030724067240953445,
      "logits/rejected": 0.037750810384750366,
      "logps/chosen": -63.71375274658203,
      "logps/rejected": -74.11785888671875,
      "loss": 0.501,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.5482760667800903,
      "rewards/margins": 1.5258257389068604,
      "rewards/rejected": -0.9775495529174805,
      "step": 75
    },
    {
      "epoch": 0.4266666666666667,
      "grad_norm": 68.25962829589844,
      "learning_rate": 1e-07,
      "logits/chosen": -0.0079464977607131,
      "logits/rejected": 0.0006744593265466392,
      "logps/chosen": -44.04719924926758,
      "logps/rejected": -63.342002868652344,
      "loss": 0.441,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": 0.5909361839294434,
      "rewards/margins": 0.9477832913398743,
      "rewards/rejected": -0.3568470776081085,
      "step": 80
    },
    {
      "epoch": 0.4533333333333333,
      "grad_norm": 61.337249755859375,
      "learning_rate": 1e-07,
      "logits/chosen": -0.13302002847194672,
      "logits/rejected": -0.12650710344314575,
      "logps/chosen": -76.5062255859375,
      "logps/rejected": -77.22346496582031,
      "loss": 0.3969,
      "rewards/accuracies": 0.6000000238418579,
      "rewards/chosen": 0.7510318160057068,
      "rewards/margins": 1.1422964334487915,
      "rewards/rejected": -0.3912646174430847,
      "step": 85
    },
    {
      "epoch": 0.48,
      "grad_norm": 60.21578598022461,
      "learning_rate": 1e-07,
      "logits/chosen": 0.08266867697238922,
      "logits/rejected": 0.08583595603704453,
      "logps/chosen": -85.27580261230469,
      "logps/rejected": -83.88575744628906,
      "loss": 0.4887,
      "rewards/accuracies": 0.6499999761581421,
      "rewards/chosen": 0.2144974172115326,
      "rewards/margins": 1.155656337738037,
      "rewards/rejected": -0.9411588907241821,
      "step": 90
    },
    {
      "epoch": 0.5066666666666667,
      "grad_norm": 78.39936828613281,
      "learning_rate": 1e-07,
      "logits/chosen": -0.018161656334996223,
      "logits/rejected": -0.012231660075485706,
      "logps/chosen": -72.49486541748047,
      "logps/rejected": -81.66197967529297,
      "loss": 0.4794,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": 0.487576961517334,
      "rewards/margins": 1.1082786321640015,
      "rewards/rejected": -0.6207017302513123,
      "step": 95
    },
    {
      "epoch": 0.5333333333333333,
      "grad_norm": 72.66243743896484,
      "learning_rate": 1e-07,
      "logits/chosen": -0.11775938421487808,
      "logits/rejected": -0.10979805141687393,
      "logps/chosen": -37.96864318847656,
      "logps/rejected": -41.957908630371094,
      "loss": 0.4535,
      "rewards/accuracies": 0.75,
      "rewards/chosen": 0.5172302722930908,
      "rewards/margins": 0.8645200729370117,
      "rewards/rejected": -0.3472897410392761,
      "step": 100
    },
    {
      "epoch": 0.56,
      "grad_norm": 84.33699798583984,
      "learning_rate": 1e-07,
      "logits/chosen": -0.07096003741025925,
      "logits/rejected": -0.06229569762945175,
      "logps/chosen": -79.24551391601562,
      "logps/rejected": -73.00336456298828,
      "loss": 0.4727,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.48409804701805115,
      "rewards/margins": 1.5594733953475952,
      "rewards/rejected": -1.0753753185272217,
      "step": 105
    },
    {
      "epoch": 0.5866666666666667,
      "grad_norm": 75.8474349975586,
      "learning_rate": 1e-07,
      "logits/chosen": 0.002287261188030243,
      "logits/rejected": 0.011448122560977936,
      "logps/chosen": -36.34912872314453,
      "logps/rejected": -46.311866760253906,
      "loss": 0.5285,
      "rewards/accuracies": 0.6000000238418579,
      "rewards/chosen": 0.3635416328907013,
      "rewards/margins": 0.3584732413291931,
      "rewards/rejected": 0.005068361759185791,
      "step": 110
    },
    {
      "epoch": 0.6133333333333333,
      "grad_norm": 75.15863800048828,
      "learning_rate": 1e-07,
      "logits/chosen": -0.07114384323358536,
      "logits/rejected": -0.06986474245786667,
      "logps/chosen": -53.876953125,
      "logps/rejected": -52.810569763183594,
      "loss": 0.4579,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": 0.7901820540428162,
      "rewards/margins": 1.1014235019683838,
      "rewards/rejected": -0.3112414479255676,
      "step": 115
    },
    {
      "epoch": 0.64,
      "grad_norm": 64.49634552001953,
      "learning_rate": 1e-07,
      "logits/chosen": -0.023414045572280884,
      "logits/rejected": -0.02103838510811329,
      "logps/chosen": -58.588645935058594,
      "logps/rejected": -54.12468719482422,
      "loss": 0.4412,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": 0.6539210081100464,
      "rewards/margins": 1.1925280094146729,
      "rewards/rejected": -0.5386068224906921,
      "step": 120
    },
    {
      "epoch": 0.6666666666666666,
      "grad_norm": 77.18801879882812,
      "learning_rate": 1e-07,
      "logits/chosen": 0.07651327550411224,
      "logits/rejected": 0.09625453501939774,
      "logps/chosen": -86.07091522216797,
      "logps/rejected": -89.9097671508789,
      "loss": 0.449,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": 0.31949272751808167,
      "rewards/margins": 1.6389261484146118,
      "rewards/rejected": -1.3194334506988525,
      "step": 125
    },
    {
      "epoch": 0.6933333333333334,
      "grad_norm": 56.2886848449707,
      "learning_rate": 1e-07,
      "logits/chosen": 0.029146382585167885,
      "logits/rejected": 0.03658589348196983,
      "logps/chosen": -73.04118347167969,
      "logps/rejected": -76.2184829711914,
      "loss": 0.4175,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": 0.45680832862854004,
      "rewards/margins": 1.049865484237671,
      "rewards/rejected": -0.5930570960044861,
      "step": 130
    },
    {
      "epoch": 0.72,
      "grad_norm": 77.33464813232422,
      "learning_rate": 1e-07,
      "logits/chosen": -0.075675830245018,
      "logits/rejected": -0.0697128027677536,
      "logps/chosen": -76.72476959228516,
      "logps/rejected": -83.47200012207031,
      "loss": 0.4749,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": 0.5277665853500366,
      "rewards/margins": 1.3413928747177124,
      "rewards/rejected": -0.8136262893676758,
      "step": 135
    },
    {
      "epoch": 0.7466666666666667,
      "grad_norm": 86.7103271484375,
      "learning_rate": 1e-07,
      "logits/chosen": -0.06388665735721588,
      "logits/rejected": -0.055876873433589935,
      "logps/chosen": -48.2696647644043,
      "logps/rejected": -43.421836853027344,
      "loss": 0.4583,
      "rewards/accuracies": 0.699999988079071,
      "rewards/chosen": 0.27933669090270996,
      "rewards/margins": 0.9728415608406067,
      "rewards/rejected": -0.6935049295425415,
      "step": 140
    },
    {
      "epoch": 0.7733333333333333,
      "grad_norm": 77.15120697021484,
      "learning_rate": 1e-07,
      "logits/chosen": -0.04482016712427139,
      "logits/rejected": -0.0454244539141655,
      "logps/chosen": -58.773719787597656,
      "logps/rejected": -54.565284729003906,
      "loss": 0.4598,
      "rewards/accuracies": 0.550000011920929,
      "rewards/chosen": 0.33584028482437134,
      "rewards/margins": 0.4276786744594574,
      "rewards/rejected": -0.09183841943740845,
      "step": 145
    },
    {
      "epoch": 0.8,
      "grad_norm": 62.76329803466797,
      "learning_rate": 1e-07,
      "logits/chosen": -0.046587713062763214,
      "logits/rejected": -0.04582952335476875,
      "logps/chosen": -77.17513275146484,
      "logps/rejected": -81.3776626586914,
      "loss": 0.414,
      "rewards/accuracies": 0.75,
      "rewards/chosen": -0.00032804012880660594,
      "rewards/margins": 0.8608261346817017,
      "rewards/rejected": -0.8611541986465454,
      "step": 150
    },
    {
      "epoch": 0.8266666666666667,
      "grad_norm": 48.6246337890625,
      "learning_rate": 1e-07,
      "logits/chosen": -0.0023877639323472977,
      "logits/rejected": 0.000560196116566658,
      "logps/chosen": -61.36391067504883,
      "logps/rejected": -74.22142028808594,
      "loss": 0.3678,
      "rewards/accuracies": 0.75,
      "rewards/chosen": 0.5382565855979919,
      "rewards/margins": 1.6945278644561768,
      "rewards/rejected": -1.15627121925354,
      "step": 155
    },
    {
      "epoch": 0.8533333333333334,
      "grad_norm": 66.40213775634766,
      "learning_rate": 1e-07,
      "logits/chosen": -0.07235555350780487,
      "logits/rejected": -0.06712132692337036,
      "logps/chosen": -47.608489990234375,
      "logps/rejected": -58.56159591674805,
      "loss": 0.4747,
      "rewards/accuracies": 0.6499999761581421,
      "rewards/chosen": 0.3097238540649414,
      "rewards/margins": 0.753086268901825,
      "rewards/rejected": -0.44336241483688354,
      "step": 160
    },
    {
      "epoch": 0.88,
      "grad_norm": 56.45853805541992,
      "learning_rate": 1e-07,
      "logits/chosen": 0.007908957079052925,
      "logits/rejected": 0.008793281391263008,
      "logps/chosen": -56.72149658203125,
      "logps/rejected": -59.84600067138672,
      "loss": 0.4839,
      "rewards/accuracies": 0.550000011920929,
      "rewards/chosen": 0.4165565073490143,
      "rewards/margins": 0.6458317637443542,
      "rewards/rejected": -0.22927527129650116,
      "step": 165
    },
    {
      "epoch": 0.9066666666666666,
      "grad_norm": 56.173946380615234,
      "learning_rate": 1e-07,
      "logits/chosen": 0.030985862016677856,
      "logits/rejected": 0.03558924049139023,
      "logps/chosen": -62.272987365722656,
      "logps/rejected": -69.59276580810547,
      "loss": 0.4139,
      "rewards/accuracies": 0.8500000238418579,
      "rewards/chosen": 0.3212319016456604,
      "rewards/margins": 1.2905486822128296,
      "rewards/rejected": -0.9693166613578796,
      "step": 170
    },
    {
      "epoch": 0.9333333333333333,
      "grad_norm": 78.22785949707031,
      "learning_rate": 1e-07,
      "logits/chosen": 0.0006909176590852439,
      "logits/rejected": 0.002517472254112363,
      "logps/chosen": -59.8375129699707,
      "logps/rejected": -58.26896286010742,
      "loss": 0.3862,
      "rewards/accuracies": 0.8500000238418579,
      "rewards/chosen": 0.49182993173599243,
      "rewards/margins": 1.2024425268173218,
      "rewards/rejected": -0.7106126546859741,
      "step": 175
    },
    {
      "epoch": 0.96,
      "grad_norm": 50.518157958984375,
      "learning_rate": 1e-07,
      "logits/chosen": 0.03367740660905838,
      "logits/rejected": 0.04943201318383217,
      "logps/chosen": -73.77796173095703,
      "logps/rejected": -83.64482879638672,
      "loss": 0.4253,
      "rewards/accuracies": 0.8500000238418579,
      "rewards/chosen": 0.5700420141220093,
      "rewards/margins": 2.157716751098633,
      "rewards/rejected": -1.587674856185913,
      "step": 180
    },
    {
      "epoch": 0.9866666666666667,
      "grad_norm": 57.97651290893555,
      "learning_rate": 1e-07,
      "logits/chosen": -0.07821191847324371,
      "logits/rejected": -0.07423903793096542,
      "logps/chosen": -63.325286865234375,
      "logps/rejected": -49.606910705566406,
      "loss": 0.4184,
      "rewards/accuracies": 0.75,
      "rewards/chosen": 0.7217522859573364,
      "rewards/margins": 1.4272280931472778,
      "rewards/rejected": -0.7054757475852966,
      "step": 185
    },
    {
      "epoch": 1.0133333333333334,
      "grad_norm": 42.711490631103516,
      "learning_rate": 1e-07,
      "logits/chosen": -0.011517315171658993,
      "logits/rejected": -0.016577688977122307,
      "logps/chosen": -63.11487579345703,
      "logps/rejected": -71.26173400878906,
      "loss": 0.3845,
      "rewards/accuracies": 0.8500000238418579,
      "rewards/chosen": 0.48871931433677673,
      "rewards/margins": 1.1433701515197754,
      "rewards/rejected": -0.654650866985321,
      "step": 190
    },
    {
      "epoch": 1.04,
      "grad_norm": 36.31159973144531,
      "learning_rate": 1e-07,
      "logits/chosen": -0.006033201701939106,
      "logits/rejected": 0.016330499202013016,
      "logps/chosen": -57.083030700683594,
      "logps/rejected": -71.14306640625,
      "loss": 0.2752,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": 0.7944145202636719,
      "rewards/margins": 1.1578031778335571,
      "rewards/rejected": -0.3633885979652405,
      "step": 195
    },
    {
      "epoch": 1.0666666666666667,
      "grad_norm": 32.363685607910156,
      "learning_rate": 1e-07,
      "logits/chosen": -0.09754471480846405,
      "logits/rejected": -0.09675069153308868,
      "logps/chosen": -50.0403938293457,
      "logps/rejected": -50.053855895996094,
      "loss": 0.2892,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": 0.6043117642402649,
      "rewards/margins": 1.4791052341461182,
      "rewards/rejected": -0.8747934103012085,
      "step": 200
    },
    {
      "epoch": 1.0933333333333333,
      "grad_norm": 49.40591812133789,
      "learning_rate": 1e-07,
      "logits/chosen": 0.03340379148721695,
      "logits/rejected": 0.03613832965493202,
      "logps/chosen": -75.98091888427734,
      "logps/rejected": -89.6574935913086,
      "loss": 0.2736,
      "rewards/accuracies": 0.8500000238418579,
      "rewards/chosen": 0.5458163022994995,
      "rewards/margins": 1.8934462070465088,
      "rewards/rejected": -1.3476299047470093,
      "step": 205
    },
    {
      "epoch": 1.12,
      "grad_norm": 33.99880599975586,
      "learning_rate": 1e-07,
      "logits/chosen": -0.014881300739943981,
      "logits/rejected": -0.008952580392360687,
      "logps/chosen": -61.28364944458008,
      "logps/rejected": -67.7909164428711,
      "loss": 0.2599,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.6829001903533936,
      "rewards/margins": 2.180811643600464,
      "rewards/rejected": -1.4979116916656494,
      "step": 210
    },
    {
      "epoch": 1.1466666666666667,
      "grad_norm": 32.734886169433594,
      "learning_rate": 1e-07,
      "logits/chosen": -0.1162935271859169,
      "logits/rejected": -0.1172286868095398,
      "logps/chosen": -72.66918182373047,
      "logps/rejected": -81.00222778320312,
      "loss": 0.2414,
      "rewards/accuracies": 1.0,
      "rewards/chosen": 0.747806191444397,
      "rewards/margins": 2.400660276412964,
      "rewards/rejected": -1.652854323387146,
      "step": 215
    },
    {
      "epoch": 1.1733333333333333,
      "grad_norm": 45.963077545166016,
      "learning_rate": 1e-07,
      "logits/chosen": 0.03322611004114151,
      "logits/rejected": 0.03860603645443916,
      "logps/chosen": -67.83540344238281,
      "logps/rejected": -77.8966293334961,
      "loss": 0.262,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": 0.487581342458725,
      "rewards/margins": 2.018979072570801,
      "rewards/rejected": -1.531397819519043,
      "step": 220
    },
    {
      "epoch": 1.2,
      "grad_norm": 31.000837326049805,
      "learning_rate": 1e-07,
      "logits/chosen": -0.06304231286048889,
      "logits/rejected": -0.06262607872486115,
      "logps/chosen": -72.00182342529297,
      "logps/rejected": -69.18010711669922,
      "loss": 0.3098,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": -0.16077418625354767,
      "rewards/margins": 1.9396165609359741,
      "rewards/rejected": -2.100390911102295,
      "step": 225
    },
    {
      "epoch": 1.2266666666666666,
      "grad_norm": 41.64716339111328,
      "learning_rate": 1e-07,
      "logits/chosen": -0.06013431400060654,
      "logits/rejected": -0.05363202840089798,
      "logps/chosen": -71.9935531616211,
      "logps/rejected": -77.51136779785156,
      "loss": 0.276,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.6929202675819397,
      "rewards/margins": 2.5861668586730957,
      "rewards/rejected": -1.8932468891143799,
      "step": 230
    },
    {
      "epoch": 1.2533333333333334,
      "grad_norm": 39.7431526184082,
      "learning_rate": 1e-07,
      "logits/chosen": -0.0543329119682312,
      "logits/rejected": -0.04811491817235947,
      "logps/chosen": -44.972129821777344,
      "logps/rejected": -48.955535888671875,
      "loss": 0.3028,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": 1.212233066558838,
      "rewards/margins": 2.2813668251037598,
      "rewards/rejected": -1.069133996963501,
      "step": 235
    },
    {
      "epoch": 1.28,
      "grad_norm": 36.94087600708008,
      "learning_rate": 1e-07,
      "logits/chosen": -0.03034238889813423,
      "logits/rejected": -0.02829553559422493,
      "logps/chosen": -45.940433502197266,
      "logps/rejected": -47.90568161010742,
      "loss": 0.2795,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": 0.9621788263320923,
      "rewards/margins": 1.6789710521697998,
      "rewards/rejected": -0.7167922258377075,
      "step": 240
    },
    {
      "epoch": 1.3066666666666666,
      "grad_norm": 34.41165542602539,
      "learning_rate": 1e-07,
      "logits/chosen": 0.08339103311300278,
      "logits/rejected": 0.10816546529531479,
      "logps/chosen": -73.56194305419922,
      "logps/rejected": -73.25684356689453,
      "loss": 0.2708,
      "rewards/accuracies": 1.0,
      "rewards/chosen": 0.37739673256874084,
      "rewards/margins": 2.098477840423584,
      "rewards/rejected": -1.7210811376571655,
      "step": 245
    },
    {
      "epoch": 1.3333333333333333,
      "grad_norm": 31.31505584716797,
      "learning_rate": 1e-07,
      "logits/chosen": 0.04144307225942612,
      "logits/rejected": 0.04325007274746895,
      "logps/chosen": -71.2078628540039,
      "logps/rejected": -80.70985412597656,
      "loss": 0.274,
      "rewards/accuracies": 0.8500000238418579,
      "rewards/chosen": 0.5666126012802124,
      "rewards/margins": 1.788511872291565,
      "rewards/rejected": -1.2218992710113525,
      "step": 250
    },
    {
      "epoch": 1.3599999999999999,
      "grad_norm": 35.27589416503906,
      "learning_rate": 1e-07,
      "logits/chosen": 0.09748046845197678,
      "logits/rejected": 0.08802054077386856,
      "logps/chosen": -70.91510772705078,
      "logps/rejected": -70.45256805419922,
      "loss": 0.2802,
      "rewards/accuracies": 0.699999988079071,
      "rewards/chosen": 0.5191822648048401,
      "rewards/margins": 2.1395535469055176,
      "rewards/rejected": -1.6203714609146118,
      "step": 255
    },
    {
      "epoch": 1.3866666666666667,
      "grad_norm": 36.5883674621582,
      "learning_rate": 1e-07,
      "logits/chosen": -0.028733888640999794,
      "logits/rejected": -0.022805647924542427,
      "logps/chosen": -68.54586029052734,
      "logps/rejected": -66.59730529785156,
      "loss": 0.3011,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": 0.6113399267196655,
      "rewards/margins": 2.464463472366333,
      "rewards/rejected": -1.853123664855957,
      "step": 260
    },
    {
      "epoch": 1.4133333333333333,
      "grad_norm": 21.833608627319336,
      "learning_rate": 1e-07,
      "logits/chosen": 0.007651590742170811,
      "logits/rejected": 0.011075135320425034,
      "logps/chosen": -70.35153198242188,
      "logps/rejected": -72.46214294433594,
      "loss": 0.2587,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.6863104701042175,
      "rewards/margins": 1.9516360759735107,
      "rewards/rejected": -1.2653255462646484,
      "step": 265
    },
    {
      "epoch": 1.44,
      "grad_norm": 42.52511215209961,
      "learning_rate": 1e-07,
      "logits/chosen": -0.019661933183670044,
      "logits/rejected": -0.020269040018320084,
      "logps/chosen": -81.03693389892578,
      "logps/rejected": -75.59123229980469,
      "loss": 0.263,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": 1.1058200597763062,
      "rewards/margins": 2.4359288215637207,
      "rewards/rejected": -1.3301087617874146,
      "step": 270
    },
    {
      "epoch": 1.4666666666666668,
      "grad_norm": 54.02488327026367,
      "learning_rate": 1e-07,
      "logits/chosen": -0.045877061784267426,
      "logits/rejected": -0.05564676970243454,
      "logps/chosen": -80.84842681884766,
      "logps/rejected": -79.160888671875,
      "loss": 0.3026,
      "rewards/accuracies": 0.8500000238418579,
      "rewards/chosen": 0.76087886095047,
      "rewards/margins": 1.8281453847885132,
      "rewards/rejected": -1.067266583442688,
      "step": 275
    },
    {
      "epoch": 1.4933333333333334,
      "grad_norm": 50.83435821533203,
      "learning_rate": 1e-07,
      "logits/chosen": -0.09148342907428741,
      "logits/rejected": -0.08500328660011292,
      "logps/chosen": -41.23069763183594,
      "logps/rejected": -40.55739212036133,
      "loss": 0.3747,
      "rewards/accuracies": 0.8500000238418579,
      "rewards/chosen": 1.123250126838684,
      "rewards/margins": 1.7874069213867188,
      "rewards/rejected": -0.6641567945480347,
      "step": 280
    },
    {
      "epoch": 1.52,
      "grad_norm": 49.45879364013672,
      "learning_rate": 1e-07,
      "logits/chosen": -0.001988143427297473,
      "logits/rejected": -0.006540440954267979,
      "logps/chosen": -50.0777587890625,
      "logps/rejected": -53.56150436401367,
      "loss": 0.3404,
      "rewards/accuracies": 0.75,
      "rewards/chosen": 0.7993362545967102,
      "rewards/margins": 1.6362117528915405,
      "rewards/rejected": -0.8368755578994751,
      "step": 285
    },
    {
      "epoch": 1.5466666666666666,
      "grad_norm": 47.30949020385742,
      "learning_rate": 1e-07,
      "logits/chosen": 0.02390037290751934,
      "logits/rejected": 0.01714068278670311,
      "logps/chosen": -59.122962951660156,
      "logps/rejected": -60.09889602661133,
      "loss": 0.2568,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": 1.1686420440673828,
      "rewards/margins": 2.2462868690490723,
      "rewards/rejected": -1.0776447057724,
      "step": 290
    },
    {
      "epoch": 1.5733333333333333,
      "grad_norm": 34.02714538574219,
      "learning_rate": 1e-07,
      "logits/chosen": 0.0011189490323886275,
      "logits/rejected": 0.0036900490522384644,
      "logps/chosen": -56.07664108276367,
      "logps/rejected": -59.93505859375,
      "loss": 0.2744,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": 1.0613322257995605,
      "rewards/margins": 1.9388986825942993,
      "rewards/rejected": -0.8775663375854492,
      "step": 295
    },
    {
      "epoch": 1.6,
      "grad_norm": 52.804840087890625,
      "learning_rate": 1e-07,
      "logits/chosen": -0.017051326110959053,
      "logits/rejected": -0.011620084755122662,
      "logps/chosen": -53.503883361816406,
      "logps/rejected": -56.5186882019043,
      "loss": 0.3073,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": 0.9534000158309937,
      "rewards/margins": 1.7106940746307373,
      "rewards/rejected": -0.7572939395904541,
      "step": 300
    },
    {
      "epoch": 1.6266666666666667,
      "grad_norm": 38.153648376464844,
      "learning_rate": 1e-07,
      "logits/chosen": 0.00943007878959179,
      "logits/rejected": 0.028043877333402634,
      "logps/chosen": -66.48413848876953,
      "logps/rejected": -105.61421966552734,
      "loss": 0.2633,
      "rewards/accuracies": 0.8500000238418579,
      "rewards/chosen": 0.8428307771682739,
      "rewards/margins": 2.7967689037323,
      "rewards/rejected": -1.9539384841918945,
      "step": 305
    },
    {
      "epoch": 1.6533333333333333,
      "grad_norm": 61.87904739379883,
      "learning_rate": 1e-07,
      "logits/chosen": -0.03582911938428879,
      "logits/rejected": -0.0231150072067976,
      "logps/chosen": -59.17612838745117,
      "logps/rejected": -71.96540832519531,
      "loss": 0.2505,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.9612646102905273,
      "rewards/margins": 2.2090392112731934,
      "rewards/rejected": -1.247774600982666,
      "step": 310
    },
    {
      "epoch": 1.6800000000000002,
      "grad_norm": 55.832542419433594,
      "learning_rate": 1e-07,
      "logits/chosen": -0.044271357357501984,
      "logits/rejected": -0.021111857146024704,
      "logps/chosen": -69.70079040527344,
      "logps/rejected": -89.76176452636719,
      "loss": 0.2382,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.37344738841056824,
      "rewards/margins": 2.854713201522827,
      "rewards/rejected": -2.4812657833099365,
      "step": 315
    },
    {
      "epoch": 1.7066666666666666,
      "grad_norm": 47.42922592163086,
      "learning_rate": 1e-07,
      "logits/chosen": -0.07435096800327301,
      "logits/rejected": -0.074483223259449,
      "logps/chosen": -77.44403076171875,
      "logps/rejected": -71.1507797241211,
      "loss": 0.3022,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.5832604169845581,
      "rewards/margins": 2.0570945739746094,
      "rewards/rejected": -1.4738342761993408,
      "step": 320
    },
    {
      "epoch": 1.7333333333333334,
      "grad_norm": 44.8173713684082,
      "learning_rate": 1e-07,
      "logits/chosen": -0.001697111176326871,
      "logits/rejected": 0.004212519619613886,
      "logps/chosen": -73.09126281738281,
      "logps/rejected": -77.27495574951172,
      "loss": 0.2864,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": 0.5175535678863525,
      "rewards/margins": 1.9960486888885498,
      "rewards/rejected": -1.4784951210021973,
      "step": 325
    },
    {
      "epoch": 1.76,
      "grad_norm": 52.862125396728516,
      "learning_rate": 1e-07,
      "logits/chosen": -0.05788411572575569,
      "logits/rejected": -0.054354745894670486,
      "logps/chosen": -77.21604919433594,
      "logps/rejected": -90.43736267089844,
      "loss": 0.2593,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.5059728026390076,
      "rewards/margins": 2.594764232635498,
      "rewards/rejected": -2.088791847229004,
      "step": 330
    },
    {
      "epoch": 1.7866666666666666,
      "grad_norm": 37.6073112487793,
      "learning_rate": 1e-07,
      "logits/chosen": -0.029447251930832863,
      "logits/rejected": -0.02990039810538292,
      "logps/chosen": -61.29267501831055,
      "logps/rejected": -83.2472915649414,
      "loss": 0.1939,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.3546075224876404,
      "rewards/margins": 2.5164620876312256,
      "rewards/rejected": -2.1618547439575195,
      "step": 335
    },
    {
      "epoch": 1.8133333333333335,
      "grad_norm": 49.868560791015625,
      "learning_rate": 1e-07,
      "logits/chosen": 0.016156326979398727,
      "logits/rejected": 0.03281542658805847,
      "logps/chosen": -81.39087677001953,
      "logps/rejected": -81.13011169433594,
      "loss": 0.2456,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.4369918704032898,
      "rewards/margins": 2.3583126068115234,
      "rewards/rejected": -1.9213206768035889,
      "step": 340
    },
    {
      "epoch": 1.8399999999999999,
      "grad_norm": 41.97678756713867,
      "learning_rate": 1e-07,
      "logits/chosen": -0.021811150014400482,
      "logits/rejected": -0.02089584432542324,
      "logps/chosen": -69.17388916015625,
      "logps/rejected": -58.414764404296875,
      "loss": 0.3057,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": 1.2403881549835205,
      "rewards/margins": 2.0268137454986572,
      "rewards/rejected": -0.7864255905151367,
      "step": 345
    },
    {
      "epoch": 1.8666666666666667,
      "grad_norm": 46.311893463134766,
      "learning_rate": 1e-07,
      "logits/chosen": -0.02446429245173931,
      "logits/rejected": -0.014963751658797264,
      "logps/chosen": -36.05630874633789,
      "logps/rejected": -43.71437454223633,
      "loss": 0.2696,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 1.1039693355560303,
      "rewards/margins": 2.3277952671051025,
      "rewards/rejected": -1.2238258123397827,
      "step": 350
    },
    {
      "epoch": 1.8933333333333333,
      "grad_norm": 44.26567459106445,
      "learning_rate": 1e-07,
      "logits/chosen": -0.04746484011411667,
      "logits/rejected": -0.04588810354471207,
      "logps/chosen": -67.33465576171875,
      "logps/rejected": -73.4018325805664,
      "loss": 0.2557,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.6922093629837036,
      "rewards/margins": 1.8044688701629639,
      "rewards/rejected": -1.1122596263885498,
      "step": 355
    },
    {
      "epoch": 1.92,
      "grad_norm": 29.275634765625,
      "learning_rate": 1e-07,
      "logits/chosen": -0.00010519176430534571,
      "logits/rejected": 0.01266922615468502,
      "logps/chosen": -63.38561248779297,
      "logps/rejected": -76.10896301269531,
      "loss": 0.2347,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": 0.5989769697189331,
      "rewards/margins": 2.0385549068450928,
      "rewards/rejected": -1.4395779371261597,
      "step": 360
    },
    {
      "epoch": 1.9466666666666668,
      "grad_norm": 38.70104217529297,
      "learning_rate": 1e-07,
      "logits/chosen": -0.07840217649936676,
      "logits/rejected": -0.06996472924947739,
      "logps/chosen": -77.53556060791016,
      "logps/rejected": -79.41651916503906,
      "loss": 0.23,
      "rewards/accuracies": 0.8999999761581421,
      "rewards/chosen": 0.3816269040107727,
      "rewards/margins": 2.385305166244507,
      "rewards/rejected": -2.003678321838379,
      "step": 365
    },
    {
      "epoch": 1.9733333333333334,
      "grad_norm": 45.61483383178711,
      "learning_rate": 1e-07,
      "logits/chosen": -0.030093077570199966,
      "logits/rejected": -0.03412201628088951,
      "logps/chosen": -65.24625396728516,
      "logps/rejected": -68.85826873779297,
      "loss": 0.3033,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": 1.0214712619781494,
      "rewards/margins": 1.9241225719451904,
      "rewards/rejected": -0.902651309967041,
      "step": 370
    },
    {
      "epoch": 1.9946666666666668,
      "step": 374,
      "total_flos": 0.0,
      "train_loss": 0.3853599507222201,
      "train_runtime": 1020.3175,
      "train_samples_per_second": 11.759,
      "train_steps_per_second": 0.367
    }
  ],
  "logging_steps": 5,
  "max_steps": 374,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 2,
  "save_steps": 500,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": false,
        "should_training_stop": false
      },
      "attributes": {}
    }
  },
  "total_flos": 0.0,
  "train_batch_size": 2,
  "trial_name": null,
  "trial_params": null
}