hugodk-sch commited on
Commit
a1d636a
1 Parent(s): f4e6563

Model save

Browse files
Files changed (5) hide show
  1. README.md +13 -16
  2. adapter_model.safetensors +1 -1
  3. all_results.json +2 -15
  4. train_results.json +2 -2
  5. trainer_state.json +376 -376
README.md CHANGED
@@ -1,13 +1,10 @@
1
  ---
2
  library_name: peft
3
  tags:
4
- - alignment-handbook
5
  - trl
6
  - dpo
7
  - generated_from_trainer
8
  base_model: NbAiLab/nb-gpt-j-6B-v2
9
- datasets:
10
- - hugodk-sch/aftonposten_title_prefs
11
  model-index:
12
  - name: aftonposten-6b-align-scan
13
  results: []
@@ -18,17 +15,17 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # aftonposten-6b-align-scan
20
 
21
- This model is a fine-tuned version of [data/ap-gpt-j-6b-sft-qlora-04-08](https://huggingface.co/data/ap-gpt-j-6b-sft-qlora-04-08) on the hugodk-sch/aftonposten_title_prefs dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.6926
24
- - Rewards/chosen: -0.0279
25
- - Rewards/rejected: -0.0401
26
- - Rewards/accuracies: 0.5212
27
- - Rewards/margins: 0.0122
28
- - Logps/rejected: -37.5969
29
- - Logps/chosen: -34.0903
30
- - Logits/rejected: -2.2228
31
- - Logits/chosen: -2.2276
32
 
33
  ## Model description
34
 
@@ -63,9 +60,9 @@ The following hyperparameters were used during training:
63
 
64
  | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
65
  |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
66
- | 0.6634 | 0.26 | 100 | 0.6931 | 0.0028 | -0.0041 | 0.5216 | 0.0069 | -37.5249 | -34.0290 | -2.2276 | -2.2324 |
67
- | 0.6329 | 0.52 | 200 | 0.6905 | -0.0127 | -0.0275 | 0.5274 | 0.0148 | -37.5716 | -34.0600 | -2.2255 | -2.2304 |
68
- | 0.5742 | 0.78 | 300 | 0.6920 | -0.0273 | -0.0406 | 0.5278 | 0.0133 | -37.5978 | -34.0891 | -2.2231 | -2.2279 |
69
 
70
 
71
  ### Framework versions
 
1
  ---
2
  library_name: peft
3
  tags:
 
4
  - trl
5
  - dpo
6
  - generated_from_trainer
7
  base_model: NbAiLab/nb-gpt-j-6B-v2
 
 
8
  model-index:
9
  - name: aftonposten-6b-align-scan
10
  results: []
 
15
 
16
  # aftonposten-6b-align-scan
17
 
18
+ This model is a fine-tuned version of [NbAiLab/nb-gpt-j-6B-v2](https://huggingface.co/NbAiLab/nb-gpt-j-6B-v2) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.4944
21
+ - Rewards/chosen: 0.1419
22
+ - Rewards/rejected: 0.1189
23
+ - Rewards/accuracies: 0.5307
24
+ - Rewards/margins: 0.0230
25
+ - Logps/rejected: -37.2789
26
+ - Logps/chosen: -33.7508
27
+ - Logits/rejected: -2.2285
28
+ - Logits/chosen: -2.2333
29
 
30
  ## Model description
31
 
 
60
 
61
  | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
62
  |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
63
+ | 0.4794 | 0.26 | 100 | 0.4971 | 0.0824 | 0.0699 | 0.5573 | 0.0124 | -37.3767 | -33.8699 | -2.2327 | -2.2375 |
64
+ | 0.4561 | 0.52 | 200 | 0.4959 | 0.1292 | 0.1118 | 0.5424 | 0.0174 | -37.2930 | -33.7761 | -2.2294 | -2.2342 |
65
+ | 0.4202 | 0.78 | 300 | 0.4944 | 0.1419 | 0.1189 | 0.5307 | 0.0230 | -37.2789 | -33.7508 | -2.2285 | -2.2333 |
66
 
67
 
68
  ### Framework versions
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa79feb981d64bcd2e8b9a25f6997ab0b5ebf680edce8b9fc3c1946b2c548361
3
  size 176183216
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:238cb301e0590031e64dec1ea6ceea3ef26d10efb357a1a485256c5bc0082a4f
3
  size 176183216
all_results.json CHANGED
@@ -1,20 +1,7 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_logits/chosen": -2.22763729095459,
4
- "eval_logits/rejected": -2.222792387008667,
5
- "eval_logps/chosen": -34.090328216552734,
6
- "eval_logps/rejected": -37.59687423706055,
7
- "eval_loss": 0.6925650835037231,
8
- "eval_rewards/accuracies": 0.5211793780326843,
9
- "eval_rewards/chosen": -0.02788795717060566,
10
- "eval_rewards/margins": 0.01224011555314064,
11
- "eval_rewards/rejected": -0.04012807086110115,
12
- "eval_runtime": 145.6508,
13
- "eval_samples": 343,
14
- "eval_samples_per_second": 2.355,
15
- "eval_steps_per_second": 0.295,
16
- "train_loss": 0.64145151051608,
17
- "train_runtime": 3249.8987,
18
  "train_samples": 3079,
19
  "train_samples_per_second": 0.947,
20
  "train_steps_per_second": 0.118
 
1
  {
2
  "epoch": 1.0,
3
+ "train_loss": 0.4687722819192069,
4
+ "train_runtime": 3250.758,
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "train_samples": 3079,
6
  "train_samples_per_second": 0.947,
7
  "train_steps_per_second": 0.118
train_results.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "epoch": 1.0,
3
- "train_loss": 0.64145151051608,
4
- "train_runtime": 3249.8987,
5
  "train_samples": 3079,
6
  "train_samples_per_second": 0.947,
7
  "train_steps_per_second": 0.118
 
1
  {
2
  "epoch": 1.0,
3
+ "train_loss": 0.4687722819192069,
4
+ "train_runtime": 3250.758,
5
  "train_samples": 3079,
6
  "train_samples_per_second": 0.947,
7
  "train_steps_per_second": 0.118
trainer_state.json CHANGED
@@ -15,7 +15,7 @@
15
  "logits/rejected": -1.7377450466156006,
16
  "logps/chosen": -29.553977966308594,
17
  "logps/rejected": -42.813133239746094,
18
- "loss": 0.6931,
19
  "rewards/accuracies": 0.0,
20
  "rewards/chosen": 0.0,
21
  "rewards/margins": 0.0,
@@ -25,589 +25,589 @@
25
  {
26
  "epoch": 0.03,
27
  "learning_rate": 1.282051282051282e-06,
28
- "logits/chosen": -1.8664319515228271,
29
- "logits/rejected": -1.8707623481750488,
30
- "logps/chosen": -36.98527526855469,
31
- "logps/rejected": -33.654090881347656,
32
- "loss": 0.6829,
33
- "rewards/accuracies": 0.5416666865348816,
34
- "rewards/chosen": 0.010662304237484932,
35
- "rewards/margins": 0.02267039567232132,
36
- "rewards/rejected": -0.012008090503513813,
37
  "step": 10
38
  },
39
  {
40
  "epoch": 0.05,
41
  "learning_rate": 2.564102564102564e-06,
42
- "logits/chosen": -1.9980642795562744,
43
- "logits/rejected": -2.0007288455963135,
44
- "logps/chosen": -29.634414672851562,
45
- "logps/rejected": -29.0543270111084,
46
- "loss": 0.6943,
47
- "rewards/accuracies": 0.5874999761581421,
48
- "rewards/chosen": 0.0038894296158105135,
49
- "rewards/margins": -0.000667938613332808,
50
- "rewards/rejected": 0.004557368345558643,
51
  "step": 20
52
  },
53
  {
54
  "epoch": 0.08,
55
  "learning_rate": 3.846153846153847e-06,
56
- "logits/chosen": -1.9210376739501953,
57
- "logits/rejected": -1.9183601140975952,
58
- "logps/chosen": -31.391239166259766,
59
- "logps/rejected": -33.24319076538086,
60
- "loss": 0.6828,
61
- "rewards/accuracies": 0.6000000238418579,
62
- "rewards/chosen": 0.012426799163222313,
63
- "rewards/margins": 0.024792592972517014,
64
- "rewards/rejected": -0.0123657938092947,
65
  "step": 30
66
  },
67
  {
68
  "epoch": 0.1,
69
  "learning_rate": 4.999896948438434e-06,
70
- "logits/chosen": -2.017646312713623,
71
- "logits/rejected": -2.0089142322540283,
72
- "logps/chosen": -32.557518005371094,
73
- "logps/rejected": -32.51502227783203,
74
- "loss": 0.6894,
75
- "rewards/accuracies": 0.4749999940395355,
76
- "rewards/chosen": 0.009644975885748863,
77
- "rewards/margins": 0.010818523354828358,
78
- "rewards/rejected": -0.0011735468870028853,
79
  "step": 40
80
  },
81
  {
82
  "epoch": 0.13,
83
  "learning_rate": 4.987541037542187e-06,
84
- "logits/chosen": -1.8620023727416992,
85
- "logits/rejected": -1.8512481451034546,
86
- "logps/chosen": -33.577735900878906,
87
- "logps/rejected": -35.46040344238281,
88
- "loss": 0.6982,
89
- "rewards/accuracies": 0.5,
90
- "rewards/chosen": -0.010127579793334007,
91
- "rewards/margins": -0.0060965316370129585,
92
- "rewards/rejected": -0.004031048621982336,
93
  "step": 50
94
  },
95
  {
96
  "epoch": 0.16,
97
  "learning_rate": 4.954691471941119e-06,
98
- "logits/chosen": -1.9402154684066772,
99
- "logits/rejected": -1.9421701431274414,
100
- "logps/chosen": -32.552555084228516,
101
- "logps/rejected": -33.22978973388672,
102
- "loss": 0.6723,
103
- "rewards/accuracies": 0.6000000238418579,
104
- "rewards/chosen": 0.023814614862203598,
105
- "rewards/margins": 0.053018856793642044,
106
- "rewards/rejected": -0.029204240068793297,
107
  "step": 60
108
  },
109
  {
110
  "epoch": 0.18,
111
  "learning_rate": 4.901618883413549e-06,
112
- "logits/chosen": -2.071889877319336,
113
- "logits/rejected": -2.0768685340881348,
114
- "logps/chosen": -33.997718811035156,
115
- "logps/rejected": -36.63623809814453,
116
- "loss": 0.6836,
117
- "rewards/accuracies": 0.5375000238418579,
118
- "rewards/chosen": -0.004616844467818737,
119
- "rewards/margins": 0.02906452678143978,
120
- "rewards/rejected": -0.033681370317935944,
121
  "step": 70
122
  },
123
  {
124
  "epoch": 0.21,
125
  "learning_rate": 4.828760511501322e-06,
126
- "logits/chosen": -1.9326432943344116,
127
- "logits/rejected": -1.935786247253418,
128
- "logps/chosen": -34.30440902709961,
129
- "logps/rejected": -34.659637451171875,
130
- "loss": 0.6521,
131
- "rewards/accuracies": 0.625,
132
- "rewards/chosen": 0.06169893592596054,
133
- "rewards/margins": 0.09531383961439133,
134
- "rewards/rejected": -0.03361489623785019,
135
  "step": 80
136
  },
137
  {
138
  "epoch": 0.23,
139
  "learning_rate": 4.7367166013034295e-06,
140
- "logits/chosen": -1.9409675598144531,
141
- "logits/rejected": -1.9454774856567383,
142
- "logps/chosen": -32.3830680847168,
143
- "logps/rejected": -32.33238983154297,
144
- "loss": 0.6838,
145
- "rewards/accuracies": 0.5874999761581421,
146
- "rewards/chosen": 0.042176127433776855,
147
- "rewards/margins": 0.028917592018842697,
148
- "rewards/rejected": 0.013258534483611584,
149
  "step": 90
150
  },
151
  {
152
  "epoch": 0.26,
153
  "learning_rate": 4.626245458345211e-06,
154
- "logits/chosen": -2.038121461868286,
155
- "logits/rejected": -2.036132574081421,
156
- "logps/chosen": -32.12568664550781,
157
- "logps/rejected": -31.2890567779541,
158
- "loss": 0.6634,
159
- "rewards/accuracies": 0.5874999761581421,
160
- "rewards/chosen": 0.05757413059473038,
161
- "rewards/margins": 0.06899620592594147,
162
- "rewards/rejected": -0.011422084644436836,
163
  "step": 100
164
  },
165
  {
166
  "epoch": 0.26,
167
- "eval_logits/chosen": -2.232415199279785,
168
- "eval_logits/rejected": -2.2275755405426025,
169
- "eval_logps/chosen": -34.029048919677734,
170
- "eval_logps/rejected": -37.52485275268555,
171
- "eval_loss": 0.693107545375824,
172
- "eval_rewards/accuracies": 0.5215947031974792,
173
- "eval_rewards/chosen": 0.0027513643726706505,
174
- "eval_rewards/margins": 0.006868092343211174,
175
- "eval_rewards/rejected": -0.0041167279705405235,
176
- "eval_runtime": 145.7484,
177
- "eval_samples_per_second": 2.353,
178
  "eval_steps_per_second": 0.295,
179
  "step": 100
180
  },
181
  {
182
  "epoch": 0.29,
183
  "learning_rate": 4.498257201263691e-06,
184
- "logits/chosen": -1.9926633834838867,
185
- "logits/rejected": -1.9902803897857666,
186
- "logps/chosen": -33.11687088012695,
187
- "logps/rejected": -34.01213836669922,
188
- "loss": 0.6814,
189
- "rewards/accuracies": 0.5874999761581421,
190
- "rewards/chosen": 0.06347335875034332,
191
- "rewards/margins": 0.054856397211551666,
192
- "rewards/rejected": 0.008616959676146507,
193
  "step": 110
194
  },
195
  {
196
  "epoch": 0.31,
197
  "learning_rate": 4.353806263777678e-06,
198
- "logits/chosen": -2.0042788982391357,
199
- "logits/rejected": -1.9959495067596436,
200
- "logps/chosen": -32.306739807128906,
201
- "logps/rejected": -32.13039779663086,
202
- "loss": 0.6734,
203
- "rewards/accuracies": 0.574999988079071,
204
- "rewards/chosen": 0.06908417493104935,
205
- "rewards/margins": 0.05353052541613579,
206
- "rewards/rejected": 0.015553650446236134,
207
  "step": 120
208
  },
209
  {
210
  "epoch": 0.34,
211
  "learning_rate": 4.1940827077152755e-06,
212
- "logits/chosen": -2.0326714515686035,
213
- "logits/rejected": -2.0247092247009277,
214
- "logps/chosen": -30.308746337890625,
215
- "logps/rejected": -32.05224609375,
216
- "loss": 0.6637,
217
- "rewards/accuracies": 0.5874999761581421,
218
- "rewards/chosen": 0.08348459005355835,
219
- "rewards/margins": 0.08310474455356598,
220
- "rewards/rejected": 0.00037985146627761424,
221
  "step": 130
222
  },
223
  {
224
  "epoch": 0.36,
225
  "learning_rate": 4.0204024186666215e-06,
226
- "logits/chosen": -1.9627164602279663,
227
- "logits/rejected": -1.9729163646697998,
228
- "logps/chosen": -31.189788818359375,
229
- "logps/rejected": -32.54594421386719,
230
- "loss": 0.6424,
231
- "rewards/accuracies": 0.612500011920929,
232
- "rewards/chosen": 0.1149359717965126,
233
- "rewards/margins": 0.12046756595373154,
234
- "rewards/rejected": -0.005531603004783392,
235
  "step": 140
236
  },
237
  {
238
  "epoch": 0.39,
239
  "learning_rate": 3.834196265035119e-06,
240
- "logits/chosen": -1.8740726709365845,
241
- "logits/rejected": -1.875239372253418,
242
- "logps/chosen": -33.88011932373047,
243
- "logps/rejected": -34.779319763183594,
244
- "loss": 0.6271,
245
- "rewards/accuracies": 0.6000000238418579,
246
- "rewards/chosen": 0.15628577768802643,
247
- "rewards/margins": 0.1673184335231781,
248
- "rewards/rejected": -0.011032682843506336,
249
  "step": 150
250
  },
251
  {
252
  "epoch": 0.42,
253
  "learning_rate": 3.636998309800573e-06,
254
- "logits/chosen": -1.9257261753082275,
255
- "logits/rejected": -1.922323226928711,
256
- "logps/chosen": -35.9793586730957,
257
- "logps/rejected": -32.714969635009766,
258
- "loss": 0.6539,
259
- "rewards/accuracies": 0.612500011920929,
260
- "rewards/chosen": 0.09615939855575562,
261
- "rewards/margins": 0.09226818382740021,
262
- "rewards/rejected": 0.003891219152137637,
263
  "step": 160
264
  },
265
  {
266
  "epoch": 0.44,
267
  "learning_rate": 3.4304331721118078e-06,
268
- "logits/chosen": -2.027190923690796,
269
- "logits/rejected": -2.019850254058838,
270
- "logps/chosen": -33.4937629699707,
271
- "logps/rejected": -31.404333114624023,
272
- "loss": 0.6193,
273
  "rewards/accuracies": 0.6875,
274
- "rewards/chosen": 0.15204860270023346,
275
- "rewards/margins": 0.17829009890556335,
276
- "rewards/rejected": -0.026241496205329895,
277
  "step": 170
278
  },
279
  {
280
  "epoch": 0.47,
281
  "learning_rate": 3.2162026428305436e-06,
282
- "logits/chosen": -2.033841609954834,
283
- "logits/rejected": -2.039079427719116,
284
- "logps/chosen": -32.22673797607422,
285
- "logps/rejected": -32.453857421875,
286
- "loss": 0.632,
287
- "rewards/accuracies": 0.75,
288
- "rewards/chosen": 0.1590258777141571,
289
- "rewards/margins": 0.1433834582567215,
290
- "rewards/rejected": 0.015642408281564713,
291
  "step": 180
292
  },
293
  {
294
  "epoch": 0.49,
295
  "learning_rate": 2.996071664294641e-06,
296
- "logits/chosen": -2.03458833694458,
297
- "logits/rejected": -2.031813144683838,
298
- "logps/chosen": -31.249963760375977,
299
- "logps/rejected": -31.329097747802734,
300
- "loss": 0.6439,
301
- "rewards/accuracies": 0.5874999761581421,
302
- "rewards/chosen": 0.11949291080236435,
303
- "rewards/margins": 0.12493407726287842,
304
- "rewards/rejected": -0.005441152956336737,
305
  "step": 190
306
  },
307
  {
308
  "epoch": 0.52,
309
  "learning_rate": 2.7718537898066833e-06,
310
- "logits/chosen": -1.9052807092666626,
311
- "logits/rejected": -1.9099184274673462,
312
- "logps/chosen": -31.314193725585938,
313
- "logps/rejected": -32.81206512451172,
314
- "loss": 0.6329,
315
- "rewards/accuracies": 0.7124999761581421,
316
- "rewards/chosen": 0.14604374766349792,
317
- "rewards/margins": 0.15660937130451202,
318
- "rewards/rejected": -0.010565629228949547,
319
  "step": 200
320
  },
321
  {
322
  "epoch": 0.52,
323
- "eval_logits/chosen": -2.230377674102783,
324
- "eval_logits/rejected": -2.225529193878174,
325
- "eval_logps/chosen": -34.05998992919922,
326
- "eval_logps/rejected": -37.57161331176758,
327
- "eval_loss": 0.6905081868171692,
328
- "eval_rewards/accuracies": 0.5274086594581604,
329
- "eval_rewards/chosen": -0.012718739919364452,
330
- "eval_rewards/margins": 0.014778696931898594,
331
- "eval_rewards/rejected": -0.027497438713908195,
332
- "eval_runtime": 145.701,
333
- "eval_samples_per_second": 2.354,
334
  "eval_steps_per_second": 0.295,
335
  "step": 200
336
  },
337
  {
338
  "epoch": 0.55,
339
  "learning_rate": 2.5453962426402006e-06,
340
- "logits/chosen": -2.017439603805542,
341
- "logits/rejected": -2.0280823707580566,
342
- "logps/chosen": -31.72454261779785,
343
- "logps/rejected": -33.935951232910156,
344
- "loss": 0.6269,
345
- "rewards/accuracies": 0.6625000238418579,
346
- "rewards/chosen": 0.12337962538003922,
347
- "rewards/margins": 0.16000542044639587,
348
- "rewards/rejected": -0.036625780165195465,
349
  "step": 210
350
  },
351
  {
352
  "epoch": 0.57,
353
  "learning_rate": 2.3185646976551794e-06,
354
- "logits/chosen": -1.909328818321228,
355
- "logits/rejected": -1.924088716506958,
356
- "logps/chosen": -29.841415405273438,
357
- "logps/rejected": -31.60904884338379,
358
- "loss": 0.6274,
359
- "rewards/accuracies": 0.675000011920929,
360
- "rewards/chosen": 0.13506175577640533,
361
- "rewards/margins": 0.16049805283546448,
362
- "rewards/rejected": -0.02543630823493004,
363
  "step": 220
364
  },
365
  {
366
  "epoch": 0.6,
367
  "learning_rate": 2.0932279108998323e-06,
368
- "logits/chosen": -1.966265320777893,
369
- "logits/rejected": -1.970245361328125,
370
- "logps/chosen": -33.091209411621094,
371
- "logps/rejected": -31.639759063720703,
372
- "loss": 0.6126,
373
- "rewards/accuracies": 0.7250000238418579,
374
- "rewards/chosen": 0.16312837600708008,
375
- "rewards/margins": 0.20835788547992706,
376
- "rewards/rejected": -0.045229505747556686,
377
  "step": 230
378
  },
379
  {
380
  "epoch": 0.62,
381
  "learning_rate": 1.8712423238279358e-06,
382
- "logits/chosen": -1.9648067951202393,
383
- "logits/rejected": -1.9429614543914795,
384
- "logps/chosen": -33.82001495361328,
385
- "logps/rejected": -35.11749267578125,
386
- "loss": 0.5941,
387
- "rewards/accuracies": 0.7250000238418579,
388
- "rewards/chosen": 0.16686691343784332,
389
- "rewards/margins": 0.25021862983703613,
390
- "rewards/rejected": -0.08335171639919281,
391
  "step": 240
392
  },
393
  {
394
  "epoch": 0.65,
395
  "learning_rate": 1.6544367689701824e-06,
396
- "logits/chosen": -2.005873441696167,
397
- "logits/rejected": -2.002545118331909,
398
- "logps/chosen": -32.70961380004883,
399
- "logps/rejected": -36.252098083496094,
400
- "loss": 0.6448,
401
- "rewards/accuracies": 0.637499988079071,
402
- "rewards/chosen": 0.10772605985403061,
403
- "rewards/margins": 0.12131496518850327,
404
- "rewards/rejected": -0.013588905334472656,
405
  "step": 250
406
  },
407
  {
408
  "epoch": 0.68,
409
  "learning_rate": 1.4445974030621963e-06,
410
- "logits/chosen": -1.8737099170684814,
411
- "logits/rejected": -1.8713098764419556,
412
- "logps/chosen": -33.96501922607422,
413
- "logps/rejected": -35.54829025268555,
414
- "loss": 0.6377,
415
- "rewards/accuracies": 0.7250000238418579,
416
- "rewards/chosen": 0.11169042438268661,
417
- "rewards/margins": 0.13352498412132263,
418
- "rewards/rejected": -0.02183455601334572,
419
  "step": 260
420
  },
421
  {
422
  "epoch": 0.7,
423
  "learning_rate": 1.243452991757889e-06,
424
- "logits/chosen": -1.858642578125,
425
- "logits/rejected": -1.8562240600585938,
426
- "logps/chosen": -34.18030548095703,
427
- "logps/rejected": -31.82675552368164,
428
- "loss": 0.6384,
429
- "rewards/accuracies": 0.6625000238418579,
430
- "rewards/chosen": 0.1086462140083313,
431
- "rewards/margins": 0.1363571435213089,
432
- "rewards/rejected": -0.027710938826203346,
433
  "step": 270
434
  },
435
  {
436
  "epoch": 0.73,
437
  "learning_rate": 1.0526606671603523e-06,
438
- "logits/chosen": -1.9618957042694092,
439
- "logits/rejected": -1.9513881206512451,
440
- "logps/chosen": -35.000816345214844,
441
- "logps/rejected": -31.879558563232422,
442
- "loss": 0.6107,
443
- "rewards/accuracies": 0.737500011920929,
444
- "rewards/chosen": 0.17791253328323364,
445
- "rewards/margins": 0.1971966028213501,
446
- "rewards/rejected": -0.019284065812826157,
447
  "step": 280
448
  },
449
  {
450
  "epoch": 0.75,
451
  "learning_rate": 8.737922755071455e-07,
452
- "logits/chosen": -2.0570178031921387,
453
- "logits/rejected": -2.04209566116333,
454
- "logps/chosen": -30.695226669311523,
455
- "logps/rejected": -32.64103317260742,
456
- "loss": 0.654,
457
- "rewards/accuracies": 0.5874999761581421,
458
- "rewards/chosen": 0.11444780975580215,
459
- "rewards/margins": 0.10904743522405624,
460
- "rewards/rejected": 0.005400371737778187,
461
  "step": 290
462
  },
463
  {
464
  "epoch": 0.78,
465
  "learning_rate": 7.08321427484816e-07,
466
- "logits/chosen": -1.9285688400268555,
467
- "logits/rejected": -1.9260343313217163,
468
- "logps/chosen": -32.38969039916992,
469
- "logps/rejected": -30.898773193359375,
470
- "loss": 0.5742,
471
- "rewards/accuracies": 0.762499988079071,
472
- "rewards/chosen": 0.2687075138092041,
473
- "rewards/margins": 0.3096885085105896,
474
- "rewards/rejected": -0.04098101332783699,
475
  "step": 300
476
  },
477
  {
478
  "epoch": 0.78,
479
- "eval_logits/chosen": -2.2279160022735596,
480
- "eval_logits/rejected": -2.223081350326538,
481
- "eval_logps/chosen": -34.089088439941406,
482
- "eval_logps/rejected": -37.59783935546875,
483
- "eval_loss": 0.6919631958007812,
484
- "eval_rewards/accuracies": 0.5278239250183105,
485
- "eval_rewards/chosen": -0.027267219498753548,
486
- "eval_rewards/margins": 0.013343668542802334,
487
- "eval_rewards/rejected": -0.040610890835523605,
488
- "eval_runtime": 145.7459,
489
- "eval_samples_per_second": 2.353,
490
  "eval_steps_per_second": 0.295,
491
  "step": 300
492
  },
493
  {
494
  "epoch": 0.81,
495
  "learning_rate": 5.576113578589035e-07,
496
- "logits/chosen": -1.9126602411270142,
497
- "logits/rejected": -1.9093825817108154,
498
- "logps/chosen": -31.319168090820312,
499
- "logps/rejected": -33.805519104003906,
500
- "loss": 0.624,
501
- "rewards/accuracies": 0.7749999761581421,
502
- "rewards/chosen": 0.1420917958021164,
503
- "rewards/margins": 0.1737762689590454,
504
- "rewards/rejected": -0.03168448060750961,
505
  "step": 310
506
  },
507
  {
508
  "epoch": 0.83,
509
  "learning_rate": 4.229036944380913e-07,
510
- "logits/chosen": -1.9627044200897217,
511
- "logits/rejected": -1.9504749774932861,
512
- "logps/chosen": -34.31007385253906,
513
- "logps/rejected": -33.66672134399414,
514
- "loss": 0.6084,
515
  "rewards/accuracies": 0.7124999761581421,
516
- "rewards/chosen": 0.1430002748966217,
517
- "rewards/margins": 0.20694026350975037,
518
- "rewards/rejected": -0.06393997371196747,
519
  "step": 320
520
  },
521
  {
522
  "epoch": 0.86,
523
  "learning_rate": 3.053082288996112e-07,
524
- "logits/chosen": -1.9980976581573486,
525
- "logits/rejected": -1.9966537952423096,
526
- "logps/chosen": -33.16533660888672,
527
- "logps/rejected": -32.55678939819336,
528
- "loss": 0.6136,
529
- "rewards/accuracies": 0.7250000238418579,
530
- "rewards/chosen": 0.1602279245853424,
531
- "rewards/margins": 0.19883206486701965,
532
- "rewards/rejected": -0.03860412910580635,
533
  "step": 330
534
  },
535
  {
536
  "epoch": 0.88,
537
  "learning_rate": 2.0579377374915805e-07,
538
- "logits/chosen": -2.0848796367645264,
539
- "logits/rejected": -2.069186210632324,
540
- "logps/chosen": -33.787841796875,
541
- "logps/rejected": -33.07987976074219,
542
- "loss": 0.6229,
543
- "rewards/accuracies": 0.7250000238418579,
544
- "rewards/chosen": 0.19577431678771973,
545
- "rewards/margins": 0.17669571936130524,
546
- "rewards/rejected": 0.01907859742641449,
547
  "step": 340
548
  },
549
  {
550
  "epoch": 0.91,
551
  "learning_rate": 1.2518018074041684e-07,
552
- "logits/chosen": -1.957275390625,
553
- "logits/rejected": -1.9564218521118164,
554
- "logps/chosen": -32.81622314453125,
555
- "logps/rejected": -32.52650833129883,
556
- "loss": 0.5923,
557
- "rewards/accuracies": 0.7124999761581421,
558
- "rewards/chosen": 0.23293733596801758,
559
- "rewards/margins": 0.2636169195175171,
560
- "rewards/rejected": -0.030679568648338318,
561
  "step": 350
562
  },
563
  {
564
  "epoch": 0.94,
565
  "learning_rate": 6.41315865106129e-08,
566
- "logits/chosen": -1.9124408960342407,
567
- "logits/rejected": -1.92275071144104,
568
- "logps/chosen": -31.859888076782227,
569
- "logps/rejected": -35.33869934082031,
570
- "loss": 0.6119,
571
- "rewards/accuracies": 0.699999988079071,
572
- "rewards/chosen": 0.17240020632743835,
573
- "rewards/margins": 0.19736871123313904,
574
- "rewards/rejected": -0.024968529120087624,
575
  "step": 360
576
  },
577
  {
578
  "epoch": 0.96,
579
  "learning_rate": 2.3150941078050325e-08,
580
- "logits/chosen": -2.0522782802581787,
581
- "logits/rejected": -2.045797824859619,
582
- "logps/chosen": -33.34915542602539,
583
- "logps/rejected": -29.27215576171875,
584
- "loss": 0.6194,
585
- "rewards/accuracies": 0.75,
586
- "rewards/chosen": 0.15140748023986816,
587
- "rewards/margins": 0.1742721050977707,
588
- "rewards/rejected": -0.022864630445837975,
589
  "step": 370
590
  },
591
  {
592
  "epoch": 0.99,
593
  "learning_rate": 2.575864278703266e-09,
594
- "logits/chosen": -1.9120715856552124,
595
- "logits/rejected": -1.9142844676971436,
596
- "logps/chosen": -33.86906051635742,
597
- "logps/rejected": -30.961559295654297,
598
- "loss": 0.5996,
599
- "rewards/accuracies": 0.7250000238418579,
600
- "rewards/chosen": 0.1849948763847351,
601
- "rewards/margins": 0.22946183383464813,
602
- "rewards/rejected": -0.04446694999933243,
603
  "step": 380
604
  },
605
  {
606
  "epoch": 1.0,
607
  "step": 385,
608
  "total_flos": 0.0,
609
- "train_loss": 0.64145151051608,
610
- "train_runtime": 3249.8987,
611
  "train_samples_per_second": 0.947,
612
  "train_steps_per_second": 0.118
613
  }
 
15
  "logits/rejected": -1.7377450466156006,
16
  "logps/chosen": -29.553977966308594,
17
  "logps/rejected": -42.813133239746094,
18
+ "loss": 0.5,
19
  "rewards/accuracies": 0.0,
20
  "rewards/chosen": 0.0,
21
  "rewards/margins": 0.0,
 
25
  {
26
  "epoch": 0.03,
27
  "learning_rate": 1.282051282051282e-06,
28
+ "logits/chosen": -1.8663586378097534,
29
+ "logits/rejected": -1.8706679344177246,
30
+ "logps/chosen": -36.9964485168457,
31
+ "logps/rejected": -33.65947723388672,
32
+ "loss": 0.4966,
33
+ "rewards/accuracies": 0.5277777910232544,
34
+ "rewards/chosen": 0.005075507797300816,
35
+ "rewards/margins": 0.019778331741690636,
36
+ "rewards/rejected": -0.014702823013067245,
37
  "step": 10
38
  },
39
  {
40
  "epoch": 0.05,
41
  "learning_rate": 2.564102564102564e-06,
42
+ "logits/chosen": -1.9970680475234985,
43
+ "logits/rejected": -1.9997154474258423,
44
+ "logps/chosen": -29.64749526977539,
45
+ "logps/rejected": -29.048025131225586,
46
+ "loss": 0.5018,
47
+ "rewards/accuracies": 0.42500001192092896,
48
+ "rewards/chosen": -0.0026510744355618954,
49
+ "rewards/margins": -0.010360640473663807,
50
+ "rewards/rejected": 0.007709565572440624,
51
  "step": 20
52
  },
53
  {
54
  "epoch": 0.08,
55
  "learning_rate": 3.846153846153847e-06,
56
+ "logits/chosen": -1.9203827381134033,
57
+ "logits/rejected": -1.9176925420761108,
58
+ "logps/chosen": -31.42234230041504,
59
+ "logps/rejected": -33.24127960205078,
60
+ "loss": 0.4984,
61
+ "rewards/accuracies": 0.512499988079071,
62
+ "rewards/chosen": -0.003123724367469549,
63
+ "rewards/margins": 0.008287688717246056,
64
+ "rewards/rejected": -0.011411413550376892,
65
  "step": 30
66
  },
67
  {
68
  "epoch": 0.1,
69
  "learning_rate": 4.999896948438434e-06,
70
+ "logits/chosen": -2.018051862716675,
71
+ "logits/rejected": -2.009334087371826,
72
+ "logps/chosen": -32.55129623413086,
73
+ "logps/rejected": -32.50330352783203,
74
+ "loss": 0.4982,
75
+ "rewards/accuracies": 0.550000011920929,
76
+ "rewards/chosen": 0.012754792347550392,
77
+ "rewards/margins": 0.008071592077612877,
78
+ "rewards/rejected": 0.004683199338614941,
79
  "step": 40
80
  },
81
  {
82
  "epoch": 0.13,
83
  "learning_rate": 4.987541037542187e-06,
84
+ "logits/chosen": -1.8634856939315796,
85
+ "logits/rejected": -1.8527206182479858,
86
+ "logps/chosen": -33.50724411010742,
87
+ "logps/rejected": -35.390602111816406,
88
+ "loss": 0.5016,
89
+ "rewards/accuracies": 0.48750001192092896,
90
+ "rewards/chosen": 0.02511655166745186,
91
+ "rewards/margins": -0.005753959529101849,
92
+ "rewards/rejected": 0.030870508402585983,
93
  "step": 50
94
  },
95
  {
96
  "epoch": 0.16,
97
  "learning_rate": 4.954691471941119e-06,
98
+ "logits/chosen": -1.9438356161117554,
99
+ "logits/rejected": -1.94576096534729,
100
+ "logps/chosen": -32.481632232666016,
101
+ "logps/rejected": -33.15100860595703,
102
+ "loss": 0.4894,
103
+ "rewards/accuracies": 0.574999988079071,
104
+ "rewards/chosen": 0.05927763134241104,
105
+ "rewards/margins": 0.049092620611190796,
106
+ "rewards/rejected": 0.010185008868575096,
107
  "step": 60
108
  },
109
  {
110
  "epoch": 0.18,
111
  "learning_rate": 4.901618883413549e-06,
112
+ "logits/chosen": -2.0757923126220703,
113
+ "logits/rejected": -2.080766439437866,
114
+ "logps/chosen": -33.89708709716797,
115
+ "logps/rejected": -36.524818420410156,
116
+ "loss": 0.4939,
117
+ "rewards/accuracies": 0.574999988079071,
118
+ "rewards/chosen": 0.04569761082530022,
119
+ "rewards/margins": 0.023670893162488937,
120
+ "rewards/rejected": 0.02202671766281128,
121
  "step": 70
122
  },
123
  {
124
  "epoch": 0.21,
125
  "learning_rate": 4.828760511501322e-06,
126
+ "logits/chosen": -1.936668038368225,
127
+ "logits/rejected": -1.9397681951522827,
128
+ "logps/chosen": -34.20936965942383,
129
+ "logps/rejected": -34.525596618652344,
130
+ "loss": 0.4817,
131
+ "rewards/accuracies": 0.6000000238418579,
132
+ "rewards/chosen": 0.10922203958034515,
133
+ "rewards/margins": 0.07581819593906403,
134
+ "rewards/rejected": 0.033403851091861725,
135
  "step": 80
136
  },
137
  {
138
  "epoch": 0.23,
139
  "learning_rate": 4.7367166013034295e-06,
140
+ "logits/chosen": -1.9462896585464478,
141
+ "logits/rejected": -1.9508006572723389,
142
+ "logps/chosen": -32.27099609375,
143
+ "logps/rejected": -32.275699615478516,
144
+ "loss": 0.4865,
145
+ "rewards/accuracies": 0.612500011920929,
146
+ "rewards/chosen": 0.0982138067483902,
147
+ "rewards/margins": 0.05660901591181755,
148
+ "rewards/rejected": 0.04160478338599205,
149
  "step": 90
150
  },
151
  {
152
  "epoch": 0.26,
153
  "learning_rate": 4.626245458345211e-06,
154
+ "logits/chosen": -2.043682336807251,
155
+ "logits/rejected": -2.04168438911438,
156
+ "logps/chosen": -31.95809555053711,
157
+ "logps/rejected": -31.16133689880371,
158
+ "loss": 0.4794,
159
+ "rewards/accuracies": 0.6499999761581421,
160
+ "rewards/chosen": 0.14136961102485657,
161
+ "rewards/margins": 0.08893296122550964,
162
+ "rewards/rejected": 0.05243664234876633,
163
  "step": 100
164
  },
165
  {
166
  "epoch": 0.26,
167
+ "eval_logits/chosen": -2.2374911308288574,
168
+ "eval_logits/rejected": -2.232652425765991,
169
+ "eval_logps/chosen": -33.869850158691406,
170
+ "eval_logps/rejected": -37.37673568725586,
171
+ "eval_loss": 0.49711015820503235,
172
+ "eval_rewards/accuracies": 0.5573089718818665,
173
+ "eval_rewards/chosen": 0.08235026895999908,
174
+ "eval_rewards/margins": 0.012408134527504444,
175
+ "eval_rewards/rejected": 0.06994213908910751,
176
+ "eval_runtime": 145.9739,
177
+ "eval_samples_per_second": 2.35,
178
  "eval_steps_per_second": 0.295,
179
  "step": 100
180
  },
181
  {
182
  "epoch": 0.29,
183
  "learning_rate": 4.498257201263691e-06,
184
+ "logits/chosen": -1.9995167255401611,
185
+ "logits/rejected": -1.9971492290496826,
186
+ "logps/chosen": -32.9360237121582,
187
+ "logps/rejected": -33.857337951660156,
188
+ "loss": 0.4819,
189
+ "rewards/accuracies": 0.574999988079071,
190
+ "rewards/chosen": 0.15389741957187653,
191
+ "rewards/margins": 0.06788130104541779,
192
+ "rewards/rejected": 0.08601613342761993,
193
  "step": 110
194
  },
195
  {
196
  "epoch": 0.31,
197
  "learning_rate": 4.353806263777678e-06,
198
+ "logits/chosen": -2.0095458030700684,
199
+ "logits/rejected": -2.001213788986206,
200
+ "logps/chosen": -32.16087341308594,
201
+ "logps/rejected": -31.98464584350586,
202
+ "loss": 0.4876,
203
+ "rewards/accuracies": 0.5874999761581421,
204
+ "rewards/chosen": 0.142018124461174,
205
+ "rewards/margins": 0.05358927324414253,
206
+ "rewards/rejected": 0.08842884749174118,
207
  "step": 120
208
  },
209
  {
210
  "epoch": 0.34,
211
  "learning_rate": 4.1940827077152755e-06,
212
+ "logits/chosen": -2.0384786128997803,
213
+ "logits/rejected": -2.030527114868164,
214
+ "logps/chosen": -30.18499755859375,
215
+ "logps/rejected": -31.905630111694336,
216
+ "loss": 0.4841,
217
+ "rewards/accuracies": 0.625,
218
+ "rewards/chosen": 0.14535793662071228,
219
+ "rewards/margins": 0.07167014479637146,
220
+ "rewards/rejected": 0.07368779182434082,
221
  "step": 130
222
  },
223
  {
224
  "epoch": 0.36,
225
  "learning_rate": 4.0204024186666215e-06,
226
+ "logits/chosen": -1.9675172567367554,
227
+ "logits/rejected": -1.97771418094635,
228
+ "logps/chosen": -31.068078994750977,
229
+ "logps/rejected": -32.39047622680664,
230
+ "loss": 0.4757,
231
+ "rewards/accuracies": 0.6000000238418579,
232
+ "rewards/chosen": 0.17579205334186554,
233
+ "rewards/margins": 0.10358880460262299,
234
+ "rewards/rejected": 0.07220325618982315,
235
  "step": 140
236
  },
237
  {
238
  "epoch": 0.39,
239
  "learning_rate": 3.834196265035119e-06,
240
+ "logits/chosen": -1.8799912929534912,
241
+ "logits/rejected": -1.8811372518539429,
242
+ "logps/chosen": -33.688819885253906,
243
+ "logps/rejected": -34.5561637878418,
244
+ "loss": 0.4651,
245
+ "rewards/accuracies": 0.574999988079071,
246
+ "rewards/chosen": 0.2519363760948181,
247
+ "rewards/margins": 0.15139077603816986,
248
+ "rewards/rejected": 0.10054560005664825,
249
  "step": 150
250
  },
251
  {
252
  "epoch": 0.42,
253
  "learning_rate": 3.636998309800573e-06,
254
+ "logits/chosen": -1.9306682348251343,
255
+ "logits/rejected": -1.9272987842559814,
256
+ "logps/chosen": -35.7833251953125,
257
+ "logps/rejected": -32.48335266113281,
258
+ "loss": 0.4818,
259
+ "rewards/accuracies": 0.625,
260
+ "rewards/chosen": 0.19417627155780792,
261
+ "rewards/margins": 0.07447630167007446,
262
+ "rewards/rejected": 0.11969996988773346,
263
  "step": 160
264
  },
265
  {
266
  "epoch": 0.44,
267
  "learning_rate": 3.4304331721118078e-06,
268
+ "logits/chosen": -2.0319628715515137,
269
+ "logits/rejected": -2.0246424674987793,
270
+ "logps/chosen": -33.24143981933594,
271
+ "logps/rejected": -31.190576553344727,
272
+ "loss": 0.4536,
273
  "rewards/accuracies": 0.6875,
274
+ "rewards/chosen": 0.2782108783721924,
275
+ "rewards/margins": 0.19757375121116638,
276
+ "rewards/rejected": 0.0806371346116066,
277
  "step": 170
278
  },
279
  {
280
  "epoch": 0.47,
281
  "learning_rate": 3.2162026428305436e-06,
282
+ "logits/chosen": -2.039504289627075,
283
+ "logits/rejected": -2.044722080230713,
284
+ "logps/chosen": -31.97749900817871,
285
+ "logps/rejected": -32.195125579833984,
286
+ "loss": 0.4662,
287
+ "rewards/accuracies": 0.7124999761581421,
288
+ "rewards/chosen": 0.28364285826683044,
289
+ "rewards/margins": 0.1386357843875885,
290
+ "rewards/rejected": 0.14500707387924194,
291
  "step": 180
292
  },
293
  {
294
  "epoch": 0.49,
295
  "learning_rate": 2.996071664294641e-06,
296
+ "logits/chosen": -2.0397489070892334,
297
+ "logits/rejected": -2.037031412124634,
298
+ "logps/chosen": -31.067768096923828,
299
+ "logps/rejected": -31.097219467163086,
300
+ "loss": 0.4756,
301
+ "rewards/accuracies": 0.625,
302
+ "rewards/chosen": 0.21059219539165497,
303
+ "rewards/margins": 0.10009355843067169,
304
+ "rewards/rejected": 0.11049864441156387,
305
  "step": 190
306
  },
307
  {
308
  "epoch": 0.52,
309
  "learning_rate": 2.7718537898066833e-06,
310
+ "logits/chosen": -1.910517930984497,
311
+ "logits/rejected": -1.915186882019043,
312
+ "logps/chosen": -31.059677124023438,
313
+ "logps/rejected": -32.61454391479492,
314
+ "loss": 0.4561,
315
+ "rewards/accuracies": 0.699999988079071,
316
+ "rewards/chosen": 0.27330341935157776,
317
+ "rewards/margins": 0.1851106435060501,
318
+ "rewards/rejected": 0.08819273114204407,
319
  "step": 200
320
  },
321
  {
322
  "epoch": 0.52,
323
+ "eval_logits/chosen": -2.234231948852539,
324
+ "eval_logits/rejected": -2.2294156551361084,
325
+ "eval_logps/chosen": -33.776145935058594,
326
+ "eval_logps/rejected": -37.29295349121094,
327
+ "eval_loss": 0.49589911103248596,
328
+ "eval_rewards/accuracies": 0.5423588156700134,
329
+ "eval_rewards/chosen": 0.12920260429382324,
330
+ "eval_rewards/margins": 0.017370687797665596,
331
+ "eval_rewards/rejected": 0.1118319109082222,
332
+ "eval_runtime": 145.6148,
333
+ "eval_samples_per_second": 2.356,
334
  "eval_steps_per_second": 0.295,
335
  "step": 200
336
  },
337
  {
338
  "epoch": 0.55,
339
  "learning_rate": 2.5453962426402006e-06,
340
+ "logits/chosen": -2.0229763984680176,
341
+ "logits/rejected": -2.0335872173309326,
342
+ "logps/chosen": -31.537296295166016,
343
+ "logps/rejected": -33.72160339355469,
344
+ "loss": 0.466,
345
+ "rewards/accuracies": 0.637499988079071,
346
+ "rewards/chosen": 0.21700111031532288,
347
+ "rewards/margins": 0.14645527303218842,
348
+ "rewards/rejected": 0.07054580748081207,
349
  "step": 210
350
  },
351
  {
352
  "epoch": 0.57,
353
  "learning_rate": 2.3185646976551794e-06,
354
+ "logits/chosen": -1.9160172939300537,
355
+ "logits/rejected": -1.9307291507720947,
356
+ "logps/chosen": -29.558719635009766,
357
+ "logps/rejected": -31.404027938842773,
358
+ "loss": 0.4532,
359
+ "rewards/accuracies": 0.7250000238418579,
360
+ "rewards/chosen": 0.27640971541404724,
361
+ "rewards/margins": 0.19933710992336273,
362
+ "rewards/rejected": 0.07707259804010391,
363
  "step": 220
364
  },
365
  {
366
  "epoch": 0.6,
367
  "learning_rate": 2.0932279108998323e-06,
368
+ "logits/chosen": -1.9730017185211182,
369
+ "logits/rejected": -1.9770148992538452,
370
+ "logps/chosen": -32.84386444091797,
371
+ "logps/rejected": -31.42836570739746,
372
+ "loss": 0.4471,
373
+ "rewards/accuracies": 0.6875,
374
+ "rewards/chosen": 0.2868002951145172,
375
+ "rewards/margins": 0.22633206844329834,
376
+ "rewards/rejected": 0.06046823784708977,
377
  "step": 230
378
  },
379
  {
380
  "epoch": 0.62,
381
  "learning_rate": 1.8712423238279358e-06,
382
+ "logits/chosen": -1.9717906713485718,
383
+ "logits/rejected": -1.950059175491333,
384
+ "logps/chosen": -33.563629150390625,
385
+ "logps/rejected": -34.86870193481445,
386
+ "loss": 0.4421,
387
+ "rewards/accuracies": 0.7749999761581421,
388
+ "rewards/chosen": 0.2950619161128998,
389
+ "rewards/margins": 0.25401392579078674,
390
+ "rewards/rejected": 0.041047997772693634,
391
  "step": 240
392
  },
393
  {
394
  "epoch": 0.65,
395
  "learning_rate": 1.6544367689701824e-06,
396
+ "logits/chosen": -2.0124945640563965,
397
+ "logits/rejected": -2.0092015266418457,
398
+ "logps/chosen": -32.43610382080078,
399
+ "logps/rejected": -35.959327697753906,
400
+ "loss": 0.4732,
401
+ "rewards/accuracies": 0.5874999761581421,
402
+ "rewards/chosen": 0.24448053538799286,
403
+ "rewards/margins": 0.11168196052312851,
404
+ "rewards/rejected": 0.13279855251312256,
405
  "step": 250
406
  },
407
  {
408
  "epoch": 0.68,
409
  "learning_rate": 1.4445974030621963e-06,
410
+ "logits/chosen": -1.8797550201416016,
411
+ "logits/rejected": -1.8773235082626343,
412
+ "logps/chosen": -33.68698501586914,
413
+ "logps/rejected": -35.27508544921875,
414
+ "loss": 0.4682,
415
+ "rewards/accuracies": 0.6875,
416
+ "rewards/chosen": 0.25070956349372864,
417
+ "rewards/margins": 0.13594172894954681,
418
+ "rewards/rejected": 0.11476783454418182,
419
  "step": 260
420
  },
421
  {
422
  "epoch": 0.7,
423
  "learning_rate": 1.243452991757889e-06,
424
+ "logits/chosen": -1.8655035495758057,
425
+ "logits/rejected": -1.8629907369613647,
426
+ "logps/chosen": -33.91318893432617,
427
+ "logps/rejected": -31.576608657836914,
428
+ "loss": 0.4665,
429
+ "rewards/accuracies": 0.637499988079071,
430
+ "rewards/chosen": 0.24220648407936096,
431
+ "rewards/margins": 0.1448442041873932,
432
+ "rewards/rejected": 0.09736229479312897,
433
  "step": 270
434
  },
435
  {
436
  "epoch": 0.73,
437
  "learning_rate": 1.0526606671603523e-06,
438
+ "logits/chosen": -1.9690582752227783,
439
+ "logits/rejected": -1.9586395025253296,
440
+ "logps/chosen": -34.75696563720703,
441
+ "logps/rejected": -31.642765045166016,
442
+ "loss": 0.4519,
443
+ "rewards/accuracies": 0.6499999761581421,
444
+ "rewards/chosen": 0.2998362183570862,
445
+ "rewards/margins": 0.20072226226329803,
446
+ "rewards/rejected": 0.09911395609378815,
447
  "step": 280
448
  },
449
  {
450
  "epoch": 0.75,
451
  "learning_rate": 8.737922755071455e-07,
452
+ "logits/chosen": -2.0643129348754883,
453
+ "logits/rejected": -2.049489736557007,
454
+ "logps/chosen": -30.391122817993164,
455
+ "logps/rejected": -32.35709762573242,
456
+ "loss": 0.4722,
457
+ "rewards/accuracies": 0.574999988079071,
458
+ "rewards/chosen": 0.2665007710456848,
459
+ "rewards/margins": 0.11913253366947174,
460
+ "rewards/rejected": 0.14736825227737427,
461
  "step": 290
462
  },
463
  {
464
  "epoch": 0.78,
465
  "learning_rate": 7.08321427484816e-07,
466
+ "logits/chosen": -1.9364421367645264,
467
+ "logits/rejected": -1.933985710144043,
468
+ "logps/chosen": -32.0788688659668,
469
+ "logps/rejected": -30.654926300048828,
470
+ "loss": 0.4202,
471
+ "rewards/accuracies": 0.6875,
472
+ "rewards/chosen": 0.42412155866622925,
473
+ "rewards/margins": 0.34317898750305176,
474
+ "rewards/rejected": 0.08094261586666107,
475
  "step": 300
476
  },
477
  {
478
  "epoch": 0.78,
479
+ "eval_logits/chosen": -2.233308792114258,
480
+ "eval_logits/rejected": -2.2285048961639404,
481
+ "eval_logps/chosen": -33.75082015991211,
482
+ "eval_logps/rejected": -37.27885055541992,
483
+ "eval_loss": 0.4944371283054352,
484
+ "eval_rewards/accuracies": 0.5307309031486511,
485
+ "eval_rewards/chosen": 0.14186599850654602,
486
+ "eval_rewards/margins": 0.022981125861406326,
487
+ "eval_rewards/rejected": 0.1188848614692688,
488
+ "eval_runtime": 145.5333,
489
+ "eval_samples_per_second": 2.357,
490
  "eval_steps_per_second": 0.295,
491
  "step": 300
492
  },
493
  {
494
  "epoch": 0.81,
495
  "learning_rate": 5.576113578589035e-07,
496
+ "logits/chosen": -1.9192613363265991,
497
+ "logits/rejected": -1.9161159992218018,
498
+ "logps/chosen": -31.030216217041016,
499
+ "logps/rejected": -33.573768615722656,
500
+ "loss": 0.4538,
501
+ "rewards/accuracies": 0.75,
502
+ "rewards/chosen": 0.2865663468837738,
503
+ "rewards/margins": 0.20237243175506592,
504
+ "rewards/rejected": 0.0841938853263855,
505
  "step": 310
506
  },
507
  {
508
  "epoch": 0.83,
509
  "learning_rate": 4.229036944380913e-07,
510
+ "logits/chosen": -1.9703991413116455,
511
+ "logits/rejected": -1.9582570791244507,
512
+ "logps/chosen": -34.00434112548828,
513
+ "logps/rejected": -33.427330017089844,
514
+ "loss": 0.4439,
515
  "rewards/accuracies": 0.7124999761581421,
516
+ "rewards/chosen": 0.2958681583404541,
517
+ "rewards/margins": 0.24011309444904327,
518
+ "rewards/rejected": 0.05575507879257202,
519
  "step": 320
520
  },
521
  {
522
  "epoch": 0.86,
523
  "learning_rate": 3.053082288996112e-07,
524
+ "logits/chosen": -2.0051374435424805,
525
+ "logits/rejected": -2.003786563873291,
526
+ "logps/chosen": -32.876285552978516,
527
+ "logps/rejected": -32.25849914550781,
528
+ "loss": 0.4543,
529
+ "rewards/accuracies": 0.7124999761581421,
530
+ "rewards/chosen": 0.30475252866744995,
531
+ "rewards/margins": 0.1942104697227478,
532
+ "rewards/rejected": 0.11054208129644394,
533
  "step": 330
534
  },
535
  {
536
  "epoch": 0.88,
537
  "learning_rate": 2.0579377374915805e-07,
538
+ "logits/chosen": -2.092106580734253,
539
+ "logits/rejected": -2.0764684677124023,
540
+ "logps/chosen": -33.44976043701172,
541
+ "logps/rejected": -32.806304931640625,
542
+ "loss": 0.4513,
543
+ "rewards/accuracies": 0.675000011920929,
544
+ "rewards/chosen": 0.3648151457309723,
545
+ "rewards/margins": 0.2089495211839676,
546
+ "rewards/rejected": 0.1558656245470047,
547
  "step": 340
548
  },
549
  {
550
  "epoch": 0.91,
551
  "learning_rate": 1.2518018074041684e-07,
552
+ "logits/chosen": -1.9645261764526367,
553
+ "logits/rejected": -1.963702917098999,
554
+ "logps/chosen": -32.543540954589844,
555
+ "logps/rejected": -32.22251510620117,
556
+ "loss": 0.4427,
557
+ "rewards/accuracies": 0.675000011920929,
558
+ "rewards/chosen": 0.36927860975265503,
559
+ "rewards/margins": 0.24796243011951447,
560
+ "rewards/rejected": 0.12131617963314056,
561
  "step": 350
562
  },
563
  {
564
  "epoch": 0.94,
565
  "learning_rate": 6.41315865106129e-08,
566
+ "logits/chosen": -1.921724557876587,
567
+ "logits/rejected": -1.9320251941680908,
568
+ "logps/chosen": -31.573944091796875,
569
+ "logps/rejected": -35.00251007080078,
570
+ "loss": 0.4596,
571
+ "rewards/accuracies": 0.6499999761581421,
572
+ "rewards/chosen": 0.3153747618198395,
573
+ "rewards/margins": 0.17225053906440735,
574
+ "rewards/rejected": 0.14312422275543213,
575
  "step": 360
576
  },
577
  {
578
  "epoch": 0.96,
579
  "learning_rate": 2.3150941078050325e-08,
580
+ "logits/chosen": -2.059705972671509,
581
+ "logits/rejected": -2.0532097816467285,
582
+ "logps/chosen": -33.03301239013672,
583
+ "logps/rejected": -28.99460220336914,
584
+ "loss": 0.4533,
585
+ "rewards/accuracies": 0.7250000238418579,
586
+ "rewards/chosen": 0.3094797134399414,
587
+ "rewards/margins": 0.19356802105903625,
588
+ "rewards/rejected": 0.11591170728206635,
589
  "step": 370
590
  },
591
  {
592
  "epoch": 0.99,
593
  "learning_rate": 2.575864278703266e-09,
594
+ "logits/chosen": -1.919736623764038,
595
+ "logits/rejected": -1.9219011068344116,
596
+ "logps/chosen": -33.58899688720703,
597
+ "logps/rejected": -30.707019805908203,
598
+ "loss": 0.4436,
599
+ "rewards/accuracies": 0.737500011920929,
600
+ "rewards/chosen": 0.3250266909599304,
601
+ "rewards/margins": 0.24222226440906525,
602
+ "rewards/rejected": 0.08280440419912338,
603
  "step": 380
604
  },
605
  {
606
  "epoch": 1.0,
607
  "step": 385,
608
  "total_flos": 0.0,
609
+ "train_loss": 0.4687722819192069,
610
+ "train_runtime": 3250.758,
611
  "train_samples_per_second": 0.947,
612
  "train_steps_per_second": 0.118
613
  }