wxzhang commited on
Commit
7be11e4
·
verified ·
1 Parent(s): 7b87852

Model save

Browse files
README.md ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - trl
4
+ - dpo
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: dpo-selective-mixdata
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # dpo-selective-mixdata
15
+
16
+ This model was trained from scratch on the None dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: 0.5805
19
+ - Rewards/chosen: -4.7292
20
+ - Rewards/rejected: -5.2763
21
+ - Rewards/accuracies: 0.6934
22
+ - Rewards/margins: 0.5471
23
+ - Logps/rejected: -654.5243
24
+ - Logps/chosen: -590.7578
25
+ - Logits/rejected: 6.2956
26
+ - Logits/chosen: 6.4467
27
+
28
+ ## Model description
29
+
30
+ More information needed
31
+
32
+ ## Intended uses & limitations
33
+
34
+ More information needed
35
+
36
+ ## Training and evaluation data
37
+
38
+ More information needed
39
+
40
+ ## Training procedure
41
+
42
+ ### Training hyperparameters
43
+
44
+ The following hyperparameters were used during training:
45
+ - learning_rate: 5e-07
46
+ - train_batch_size: 4
47
+ - eval_batch_size: 8
48
+ - seed: 42
49
+ - distributed_type: multi-GPU
50
+ - num_devices: 4
51
+ - gradient_accumulation_steps: 4
52
+ - total_train_batch_size: 64
53
+ - total_eval_batch_size: 32
54
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
55
+ - lr_scheduler_type: cosine
56
+ - lr_scheduler_warmup_ratio: 0.1
57
+ - num_epochs: 1
58
+
59
+ ### Training results
60
+
61
+ | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
62
+ |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
63
+ | 0.5481 | 0.27 | 500 | 0.6089 | -2.6822 | -3.1236 | 0.6705 | 0.4414 | -439.2521 | -386.0565 | 3.9671 | 4.1604 |
64
+ | 0.5519 | 0.53 | 1000 | 0.5867 | -4.2523 | -4.7597 | 0.6894 | 0.5074 | -602.8671 | -543.0739 | 5.1974 | 5.3486 |
65
+ | 0.5597 | 0.8 | 1500 | 0.5821 | -4.7906 | -5.3218 | 0.6959 | 0.5311 | -659.0733 | -596.9037 | 6.4644 | 6.6294 |
66
+
67
+
68
+ ### Framework versions
69
+
70
+ - Transformers 4.36.2
71
+ - Pytorch 2.1.2
72
+ - Datasets 2.14.6
73
+ - Tokenizers 0.15.0
all_results.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_logits/chosen": 6.446723937988281,
4
+ "eval_logits/rejected": 6.295619964599609,
5
+ "eval_logps/chosen": -590.7577514648438,
6
+ "eval_logps/rejected": -654.5242919921875,
7
+ "eval_loss": 0.5805472135543823,
8
+ "eval_rewards/accuracies": 0.6934306621551514,
9
+ "eval_rewards/chosen": -4.729165554046631,
10
+ "eval_rewards/margins": 0.5471083521842957,
11
+ "eval_rewards/rejected": -5.276274681091309,
12
+ "eval_runtime": 2831.4775,
13
+ "eval_samples": 35044,
14
+ "eval_samples_per_second": 12.377,
15
+ "eval_steps_per_second": 0.387,
16
+ "train_loss": 0.5658232939218006,
17
+ "train_runtime": 45309.3847,
18
+ "train_samples": 120613,
19
+ "train_samples_per_second": 2.662,
20
+ "train_steps_per_second": 0.042
21
+ }
eval_results.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_logits/chosen": 6.446723937988281,
4
+ "eval_logits/rejected": 6.295619964599609,
5
+ "eval_logps/chosen": -590.7577514648438,
6
+ "eval_logps/rejected": -654.5242919921875,
7
+ "eval_loss": 0.5805472135543823,
8
+ "eval_rewards/accuracies": 0.6934306621551514,
9
+ "eval_rewards/chosen": -4.729165554046631,
10
+ "eval_rewards/margins": 0.5471083521842957,
11
+ "eval_rewards/rejected": -5.276274681091309,
12
+ "eval_runtime": 2831.4775,
13
+ "eval_samples": 35044,
14
+ "eval_samples_per_second": 12.377,
15
+ "eval_steps_per_second": 0.387
16
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.36.2"
6
+ }
model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd6f506ba8dcb4cfb8c0882cc3a8f8d66208b1900d110162d03eb9401cf1d6b2
3
+ size 4943162336
model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e3374e5638c392e79254199c91b4801e381cfe30e34a5fedc04a5f3f05353b8
3
+ size 4999819336
model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a01997f2bfbeb801204ce2953dcf7cf0cdb53cdbda344b89740d879ccb8f87b
3
+ size 4540516344
model.safetensors.index.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 14483464192
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00003-of-00003.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00003.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
15
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
16
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
17
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00003.safetensors",
18
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
19
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
20
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
21
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
22
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
23
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
24
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
25
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
26
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00003.safetensors",
27
+ "model.layers.10.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
28
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
29
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
30
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
31
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
32
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
33
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
34
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
35
+ "model.layers.11.input_layernorm.weight": "model-00002-of-00003.safetensors",
36
+ "model.layers.11.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
37
+ "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
38
+ "model.layers.11.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
39
+ "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
40
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
41
+ "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
42
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
43
+ "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
44
+ "model.layers.12.input_layernorm.weight": "model-00002-of-00003.safetensors",
45
+ "model.layers.12.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
46
+ "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
47
+ "model.layers.12.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
48
+ "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
49
+ "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
50
+ "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
51
+ "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
52
+ "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
53
+ "model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
54
+ "model.layers.13.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
55
+ "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
56
+ "model.layers.13.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
57
+ "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
58
+ "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
59
+ "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
60
+ "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
61
+ "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
62
+ "model.layers.14.input_layernorm.weight": "model-00002-of-00003.safetensors",
63
+ "model.layers.14.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
64
+ "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
65
+ "model.layers.14.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
66
+ "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
67
+ "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
68
+ "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
69
+ "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
70
+ "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
71
+ "model.layers.15.input_layernorm.weight": "model-00002-of-00003.safetensors",
72
+ "model.layers.15.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
73
+ "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
74
+ "model.layers.15.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
75
+ "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
76
+ "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
77
+ "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
78
+ "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
79
+ "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
80
+ "model.layers.16.input_layernorm.weight": "model-00002-of-00003.safetensors",
81
+ "model.layers.16.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
82
+ "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
83
+ "model.layers.16.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
84
+ "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
85
+ "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
86
+ "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
87
+ "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
88
+ "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
89
+ "model.layers.17.input_layernorm.weight": "model-00002-of-00003.safetensors",
90
+ "model.layers.17.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
91
+ "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
92
+ "model.layers.17.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
93
+ "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
94
+ "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
95
+ "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
96
+ "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
97
+ "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
98
+ "model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
99
+ "model.layers.18.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
100
+ "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
101
+ "model.layers.18.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
102
+ "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
103
+ "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
104
+ "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
105
+ "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
106
+ "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
107
+ "model.layers.19.input_layernorm.weight": "model-00002-of-00003.safetensors",
108
+ "model.layers.19.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
109
+ "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
110
+ "model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
111
+ "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
112
+ "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
113
+ "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
114
+ "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
115
+ "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
116
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00003.safetensors",
117
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
118
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
119
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
120
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
121
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
122
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
123
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
124
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
125
+ "model.layers.20.input_layernorm.weight": "model-00002-of-00003.safetensors",
126
+ "model.layers.20.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
127
+ "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
128
+ "model.layers.20.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
129
+ "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
130
+ "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
131
+ "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
132
+ "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
133
+ "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
134
+ "model.layers.21.input_layernorm.weight": "model-00002-of-00003.safetensors",
135
+ "model.layers.21.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
136
+ "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
137
+ "model.layers.21.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
138
+ "model.layers.21.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
139
+ "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
140
+ "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
141
+ "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
142
+ "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
143
+ "model.layers.22.input_layernorm.weight": "model-00003-of-00003.safetensors",
144
+ "model.layers.22.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
145
+ "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
146
+ "model.layers.22.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
147
+ "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
148
+ "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
149
+ "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
150
+ "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
151
+ "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
152
+ "model.layers.23.input_layernorm.weight": "model-00003-of-00003.safetensors",
153
+ "model.layers.23.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
154
+ "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
155
+ "model.layers.23.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
156
+ "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
157
+ "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
158
+ "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
159
+ "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
160
+ "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
161
+ "model.layers.24.input_layernorm.weight": "model-00003-of-00003.safetensors",
162
+ "model.layers.24.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
163
+ "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
164
+ "model.layers.24.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
165
+ "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
166
+ "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
167
+ "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
168
+ "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
169
+ "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
170
+ "model.layers.25.input_layernorm.weight": "model-00003-of-00003.safetensors",
171
+ "model.layers.25.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
172
+ "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
173
+ "model.layers.25.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
174
+ "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
175
+ "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
176
+ "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
177
+ "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
178
+ "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
179
+ "model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
180
+ "model.layers.26.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
181
+ "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
182
+ "model.layers.26.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
183
+ "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
184
+ "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
185
+ "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
186
+ "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
187
+ "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
188
+ "model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
189
+ "model.layers.27.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
190
+ "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
191
+ "model.layers.27.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
192
+ "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
193
+ "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
194
+ "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
195
+ "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
196
+ "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
197
+ "model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
198
+ "model.layers.28.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
199
+ "model.layers.28.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
200
+ "model.layers.28.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
201
+ "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
202
+ "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
203
+ "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
204
+ "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
205
+ "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
206
+ "model.layers.29.input_layernorm.weight": "model-00003-of-00003.safetensors",
207
+ "model.layers.29.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
208
+ "model.layers.29.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
209
+ "model.layers.29.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
210
+ "model.layers.29.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
211
+ "model.layers.29.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
212
+ "model.layers.29.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
213
+ "model.layers.29.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
214
+ "model.layers.29.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
215
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
216
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
217
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
218
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
219
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
220
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
221
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
222
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
223
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
224
+ "model.layers.30.input_layernorm.weight": "model-00003-of-00003.safetensors",
225
+ "model.layers.30.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
226
+ "model.layers.30.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
227
+ "model.layers.30.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
228
+ "model.layers.30.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
229
+ "model.layers.30.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
230
+ "model.layers.30.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
231
+ "model.layers.30.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
232
+ "model.layers.30.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
233
+ "model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
234
+ "model.layers.31.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
235
+ "model.layers.31.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
236
+ "model.layers.31.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
237
+ "model.layers.31.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
238
+ "model.layers.31.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
239
+ "model.layers.31.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
240
+ "model.layers.31.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
241
+ "model.layers.31.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
242
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00003.safetensors",
243
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
244
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
245
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
246
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
247
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
248
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
249
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
250
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
251
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00003.safetensors",
252
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
253
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
254
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
255
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
256
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
257
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
258
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
259
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
260
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00003.safetensors",
261
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
262
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
263
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
264
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
265
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
266
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
267
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
268
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
269
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00003.safetensors",
270
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
271
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
272
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
273
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
274
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
275
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
276
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
277
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
278
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00003.safetensors",
279
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
280
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
281
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
282
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
283
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
284
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
285
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
286
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
287
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00003.safetensors",
288
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
289
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
290
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
291
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
292
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
293
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
294
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
295
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
296
+ "model.norm.weight": "model-00003-of-00003.safetensors"
297
+ }
298
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "train_loss": 0.5658232939218006,
4
+ "train_runtime": 45309.3847,
5
+ "train_samples": 120613,
6
+ "train_samples_per_second": 2.662,
7
+ "train_steps_per_second": 0.042
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,2724 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.9996020692399522,
5
+ "eval_steps": 500,
6
+ "global_step": 1884,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.0,
13
+ "learning_rate": 2.645502645502645e-09,
14
+ "logits/chosen": -1.7827545404434204,
15
+ "logits/rejected": -1.8475866317749023,
16
+ "logps/chosen": -188.38653564453125,
17
+ "logps/rejected": -220.2354736328125,
18
+ "loss": 0.6931,
19
+ "rewards/accuracies": 0.0,
20
+ "rewards/chosen": 0.0,
21
+ "rewards/margins": 0.0,
22
+ "rewards/rejected": 0.0,
23
+ "step": 1
24
+ },
25
+ {
26
+ "epoch": 0.01,
27
+ "learning_rate": 2.6455026455026453e-08,
28
+ "logits/chosen": -1.9714936017990112,
29
+ "logits/rejected": -1.876778483390808,
30
+ "logps/chosen": -288.3788146972656,
31
+ "logps/rejected": -199.33457946777344,
32
+ "loss": 0.6931,
33
+ "rewards/accuracies": 0.3819444477558136,
34
+ "rewards/chosen": -0.00010286699398420751,
35
+ "rewards/margins": 0.00021866995666641742,
36
+ "rewards/rejected": -0.00032153705251403153,
37
+ "step": 10
38
+ },
39
+ {
40
+ "epoch": 0.01,
41
+ "learning_rate": 5.2910052910052905e-08,
42
+ "logits/chosen": -1.9301083087921143,
43
+ "logits/rejected": -1.9124500751495361,
44
+ "logps/chosen": -180.6164093017578,
45
+ "logps/rejected": -186.2301483154297,
46
+ "loss": 0.6931,
47
+ "rewards/accuracies": 0.48750001192092896,
48
+ "rewards/chosen": -4.3160351197002456e-05,
49
+ "rewards/margins": -7.290135545190424e-05,
50
+ "rewards/rejected": 2.974100425490178e-05,
51
+ "step": 20
52
+ },
53
+ {
54
+ "epoch": 0.02,
55
+ "learning_rate": 7.936507936507936e-08,
56
+ "logits/chosen": -1.9063631296157837,
57
+ "logits/rejected": -1.937107801437378,
58
+ "logps/chosen": -190.45803833007812,
59
+ "logps/rejected": -201.83729553222656,
60
+ "loss": 0.6933,
61
+ "rewards/accuracies": 0.518750011920929,
62
+ "rewards/chosen": -0.0003868662752211094,
63
+ "rewards/margins": -0.00044071293086744845,
64
+ "rewards/rejected": 5.3846673836233094e-05,
65
+ "step": 30
66
+ },
67
+ {
68
+ "epoch": 0.02,
69
+ "learning_rate": 1.0582010582010581e-07,
70
+ "logits/chosen": -1.8966213464736938,
71
+ "logits/rejected": -1.8653703927993774,
72
+ "logps/chosen": -212.13247680664062,
73
+ "logps/rejected": -186.06394958496094,
74
+ "loss": 0.693,
75
+ "rewards/accuracies": 0.4749999940395355,
76
+ "rewards/chosen": 0.00024294569448102266,
77
+ "rewards/margins": 0.00039846505387686193,
78
+ "rewards/rejected": -0.00015551943215541542,
79
+ "step": 40
80
+ },
81
+ {
82
+ "epoch": 0.03,
83
+ "learning_rate": 1.3227513227513225e-07,
84
+ "logits/chosen": -1.9822158813476562,
85
+ "logits/rejected": -1.8613579273223877,
86
+ "logps/chosen": -217.9025421142578,
87
+ "logps/rejected": -174.90078735351562,
88
+ "loss": 0.6923,
89
+ "rewards/accuracies": 0.59375,
90
+ "rewards/chosen": 0.003075198968872428,
91
+ "rewards/margins": 0.0013211520854383707,
92
+ "rewards/rejected": 0.001754046999849379,
93
+ "step": 50
94
+ },
95
+ {
96
+ "epoch": 0.03,
97
+ "learning_rate": 1.5873015873015872e-07,
98
+ "logits/chosen": -1.9449065923690796,
99
+ "logits/rejected": -1.9192440509796143,
100
+ "logps/chosen": -176.40371704101562,
101
+ "logps/rejected": -172.91717529296875,
102
+ "loss": 0.6921,
103
+ "rewards/accuracies": 0.5874999761581421,
104
+ "rewards/chosen": 0.003195743542164564,
105
+ "rewards/margins": 0.0035919342190027237,
106
+ "rewards/rejected": -0.0003961907350458205,
107
+ "step": 60
108
+ },
109
+ {
110
+ "epoch": 0.04,
111
+ "learning_rate": 1.8518518518518516e-07,
112
+ "logits/chosen": -1.9199409484863281,
113
+ "logits/rejected": -1.8474018573760986,
114
+ "logps/chosen": -174.99351501464844,
115
+ "logps/rejected": -197.31735229492188,
116
+ "loss": 0.6911,
117
+ "rewards/accuracies": 0.5687500238418579,
118
+ "rewards/chosen": 0.0023753070272505283,
119
+ "rewards/margins": 0.006421695463359356,
120
+ "rewards/rejected": -0.004046388436108828,
121
+ "step": 70
122
+ },
123
+ {
124
+ "epoch": 0.04,
125
+ "learning_rate": 2.1164021164021162e-07,
126
+ "logits/chosen": -1.8320897817611694,
127
+ "logits/rejected": -1.844348669052124,
128
+ "logps/chosen": -195.1573028564453,
129
+ "logps/rejected": -195.3834228515625,
130
+ "loss": 0.6894,
131
+ "rewards/accuracies": 0.518750011920929,
132
+ "rewards/chosen": -0.001768267946317792,
133
+ "rewards/margins": 0.007629372179508209,
134
+ "rewards/rejected": -0.009397639892995358,
135
+ "step": 80
136
+ },
137
+ {
138
+ "epoch": 0.05,
139
+ "learning_rate": 2.3809523809523806e-07,
140
+ "logits/chosen": -1.909330129623413,
141
+ "logits/rejected": -1.9080867767333984,
142
+ "logps/chosen": -196.3611297607422,
143
+ "logps/rejected": -187.08120727539062,
144
+ "loss": 0.6879,
145
+ "rewards/accuracies": 0.6187499761581421,
146
+ "rewards/chosen": -0.0239607784897089,
147
+ "rewards/margins": 0.006291805300861597,
148
+ "rewards/rejected": -0.030252579599618912,
149
+ "step": 90
150
+ },
151
+ {
152
+ "epoch": 0.05,
153
+ "learning_rate": 2.645502645502645e-07,
154
+ "logits/chosen": -1.8753902912139893,
155
+ "logits/rejected": -1.8588756322860718,
156
+ "logps/chosen": -197.9770050048828,
157
+ "logps/rejected": -210.04736328125,
158
+ "loss": 0.6842,
159
+ "rewards/accuracies": 0.5625,
160
+ "rewards/chosen": -0.05865932255983353,
161
+ "rewards/margins": 0.00952382292598486,
162
+ "rewards/rejected": -0.06818314641714096,
163
+ "step": 100
164
+ },
165
+ {
166
+ "epoch": 0.06,
167
+ "learning_rate": 2.9100529100529097e-07,
168
+ "logits/chosen": -1.942865014076233,
169
+ "logits/rejected": -1.8896948099136353,
170
+ "logps/chosen": -222.5757598876953,
171
+ "logps/rejected": -227.91476440429688,
172
+ "loss": 0.6796,
173
+ "rewards/accuracies": 0.6000000238418579,
174
+ "rewards/chosen": -0.07471577823162079,
175
+ "rewards/margins": 0.0264597050845623,
176
+ "rewards/rejected": -0.1011754721403122,
177
+ "step": 110
178
+ },
179
+ {
180
+ "epoch": 0.06,
181
+ "learning_rate": 3.1746031746031743e-07,
182
+ "logits/chosen": -1.8088843822479248,
183
+ "logits/rejected": -1.7518575191497803,
184
+ "logps/chosen": -222.64022827148438,
185
+ "logps/rejected": -209.8594512939453,
186
+ "loss": 0.6776,
187
+ "rewards/accuracies": 0.6875,
188
+ "rewards/chosen": -0.11961501836776733,
189
+ "rewards/margins": 0.06059970706701279,
190
+ "rewards/rejected": -0.18021473288536072,
191
+ "step": 120
192
+ },
193
+ {
194
+ "epoch": 0.07,
195
+ "learning_rate": 3.439153439153439e-07,
196
+ "logits/chosen": -1.7773793935775757,
197
+ "logits/rejected": -1.7670618295669556,
198
+ "logps/chosen": -222.9905548095703,
199
+ "logps/rejected": -228.2213134765625,
200
+ "loss": 0.669,
201
+ "rewards/accuracies": 0.59375,
202
+ "rewards/chosen": -0.18275949358940125,
203
+ "rewards/margins": 0.04632816091179848,
204
+ "rewards/rejected": -0.22908766567707062,
205
+ "step": 130
206
+ },
207
+ {
208
+ "epoch": 0.07,
209
+ "learning_rate": 3.703703703703703e-07,
210
+ "logits/chosen": -1.777341604232788,
211
+ "logits/rejected": -1.6979541778564453,
212
+ "logps/chosen": -227.1598358154297,
213
+ "logps/rejected": -202.8821258544922,
214
+ "loss": 0.6663,
215
+ "rewards/accuracies": 0.59375,
216
+ "rewards/chosen": -0.19891491532325745,
217
+ "rewards/margins": 0.05222553759813309,
218
+ "rewards/rejected": -0.25114041566848755,
219
+ "step": 140
220
+ },
221
+ {
222
+ "epoch": 0.08,
223
+ "learning_rate": 3.968253968253968e-07,
224
+ "logits/chosen": -1.639807939529419,
225
+ "logits/rejected": -1.5844175815582275,
226
+ "logps/chosen": -248.316650390625,
227
+ "logps/rejected": -248.49240112304688,
228
+ "loss": 0.6549,
229
+ "rewards/accuracies": 0.581250011920929,
230
+ "rewards/chosen": -0.45894140005111694,
231
+ "rewards/margins": 0.053535498678684235,
232
+ "rewards/rejected": -0.512476921081543,
233
+ "step": 150
234
+ },
235
+ {
236
+ "epoch": 0.08,
237
+ "learning_rate": 4.2328042328042324e-07,
238
+ "logits/chosen": -1.4632054567337036,
239
+ "logits/rejected": -1.4093120098114014,
240
+ "logps/chosen": -265.42877197265625,
241
+ "logps/rejected": -252.86727905273438,
242
+ "loss": 0.6609,
243
+ "rewards/accuracies": 0.6000000238418579,
244
+ "rewards/chosen": -0.6529312133789062,
245
+ "rewards/margins": 0.09157294034957886,
246
+ "rewards/rejected": -0.7445041537284851,
247
+ "step": 160
248
+ },
249
+ {
250
+ "epoch": 0.09,
251
+ "learning_rate": 4.497354497354497e-07,
252
+ "logits/chosen": -1.4973636865615845,
253
+ "logits/rejected": -1.3732246160507202,
254
+ "logps/chosen": -299.1934509277344,
255
+ "logps/rejected": -286.1786804199219,
256
+ "loss": 0.6473,
257
+ "rewards/accuracies": 0.59375,
258
+ "rewards/chosen": -0.6386870741844177,
259
+ "rewards/margins": 0.15349723398685455,
260
+ "rewards/rejected": -0.7921843528747559,
261
+ "step": 170
262
+ },
263
+ {
264
+ "epoch": 0.1,
265
+ "learning_rate": 4.761904761904761e-07,
266
+ "logits/chosen": -1.419722318649292,
267
+ "logits/rejected": -1.3405072689056396,
268
+ "logps/chosen": -262.6790466308594,
269
+ "logps/rejected": -247.8793487548828,
270
+ "loss": 0.6447,
271
+ "rewards/accuracies": 0.6000000238418579,
272
+ "rewards/chosen": -0.6575330495834351,
273
+ "rewards/margins": 0.12454842031002045,
274
+ "rewards/rejected": -0.7820814847946167,
275
+ "step": 180
276
+ },
277
+ {
278
+ "epoch": 0.1,
279
+ "learning_rate": 4.999995705919032e-07,
280
+ "logits/chosen": -1.4580373764038086,
281
+ "logits/rejected": -1.3449883460998535,
282
+ "logps/chosen": -265.0930480957031,
283
+ "logps/rejected": -266.20880126953125,
284
+ "loss": 0.649,
285
+ "rewards/accuracies": 0.6625000238418579,
286
+ "rewards/chosen": -0.6115543246269226,
287
+ "rewards/margins": 0.11900641769170761,
288
+ "rewards/rejected": -0.7305607795715332,
289
+ "step": 190
290
+ },
291
+ {
292
+ "epoch": 0.11,
293
+ "learning_rate": 4.999480434051858e-07,
294
+ "logits/chosen": -1.7100050449371338,
295
+ "logits/rejected": -1.6133902072906494,
296
+ "logps/chosen": -312.9882507324219,
297
+ "logps/rejected": -298.0437927246094,
298
+ "loss": 0.6499,
299
+ "rewards/accuracies": 0.5375000238418579,
300
+ "rewards/chosen": -0.9183656573295593,
301
+ "rewards/margins": 0.07652232050895691,
302
+ "rewards/rejected": -0.9948879480361938,
303
+ "step": 200
304
+ },
305
+ {
306
+ "epoch": 0.11,
307
+ "learning_rate": 4.998106548810311e-07,
308
+ "logits/chosen": -1.6054027080535889,
309
+ "logits/rejected": -1.5358479022979736,
310
+ "logps/chosen": -283.46551513671875,
311
+ "logps/rejected": -306.21820068359375,
312
+ "loss": 0.6371,
313
+ "rewards/accuracies": 0.637499988079071,
314
+ "rewards/chosen": -0.8442398905754089,
315
+ "rewards/margins": 0.2072862684726715,
316
+ "rewards/rejected": -1.0515261888504028,
317
+ "step": 210
318
+ },
319
+ {
320
+ "epoch": 0.12,
321
+ "learning_rate": 4.995874522146975e-07,
322
+ "logits/chosen": -1.5969936847686768,
323
+ "logits/rejected": -1.5442743301391602,
324
+ "logps/chosen": -294.7565002441406,
325
+ "logps/rejected": -295.92779541015625,
326
+ "loss": 0.6223,
327
+ "rewards/accuracies": 0.5874999761581421,
328
+ "rewards/chosen": -0.9133985638618469,
329
+ "rewards/margins": 0.20787055790424347,
330
+ "rewards/rejected": -1.1212691068649292,
331
+ "step": 220
332
+ },
333
+ {
334
+ "epoch": 0.12,
335
+ "learning_rate": 4.992785120800375e-07,
336
+ "logits/chosen": -1.0925428867340088,
337
+ "logits/rejected": -1.0447893142700195,
338
+ "logps/chosen": -327.7795104980469,
339
+ "logps/rejected": -339.69500732421875,
340
+ "loss": 0.6349,
341
+ "rewards/accuracies": 0.6499999761581421,
342
+ "rewards/chosen": -1.108730435371399,
343
+ "rewards/margins": 0.24315257370471954,
344
+ "rewards/rejected": -1.3518829345703125,
345
+ "step": 230
346
+ },
347
+ {
348
+ "epoch": 0.13,
349
+ "learning_rate": 4.988839406031596e-07,
350
+ "logits/chosen": -0.13352422416210175,
351
+ "logits/rejected": 0.016487866640090942,
352
+ "logps/chosen": -324.3548278808594,
353
+ "logps/rejected": -332.9447326660156,
354
+ "loss": 0.63,
355
+ "rewards/accuracies": 0.6187499761581421,
356
+ "rewards/chosen": -1.3696115016937256,
357
+ "rewards/margins": 0.30091235041618347,
358
+ "rewards/rejected": -1.6705236434936523,
359
+ "step": 240
360
+ },
361
+ {
362
+ "epoch": 0.13,
363
+ "learning_rate": 4.98403873325972e-07,
364
+ "logits/chosen": -0.25656959414482117,
365
+ "logits/rejected": -0.22962741553783417,
366
+ "logps/chosen": -276.41180419921875,
367
+ "logps/rejected": -297.1820983886719,
368
+ "loss": 0.6098,
369
+ "rewards/accuracies": 0.6937500238418579,
370
+ "rewards/chosen": -0.8686075210571289,
371
+ "rewards/margins": 0.2754421532154083,
372
+ "rewards/rejected": -1.1440496444702148,
373
+ "step": 250
374
+ },
375
+ {
376
+ "epoch": 0.14,
377
+ "learning_rate": 4.978384751596212e-07,
378
+ "logits/chosen": 0.7076885104179382,
379
+ "logits/rejected": 0.7922934293746948,
380
+ "logps/chosen": -315.2431640625,
381
+ "logps/rejected": -349.3119812011719,
382
+ "loss": 0.6038,
383
+ "rewards/accuracies": 0.6812499761581421,
384
+ "rewards/chosen": -1.1041444540023804,
385
+ "rewards/margins": 0.35109204053878784,
386
+ "rewards/rejected": -1.4552366733551025,
387
+ "step": 260
388
+ },
389
+ {
390
+ "epoch": 0.14,
391
+ "learning_rate": 4.971879403278432e-07,
392
+ "logits/chosen": 1.446623682975769,
393
+ "logits/rejected": 1.5377308130264282,
394
+ "logps/chosen": -370.0694580078125,
395
+ "logps/rejected": -390.1954650878906,
396
+ "loss": 0.604,
397
+ "rewards/accuracies": 0.643750011920929,
398
+ "rewards/chosen": -1.6548852920532227,
399
+ "rewards/margins": 0.3384679853916168,
400
+ "rewards/rejected": -1.9933536052703857,
401
+ "step": 270
402
+ },
403
+ {
404
+ "epoch": 0.15,
405
+ "learning_rate": 4.964524923002436e-07,
406
+ "logits/chosen": 0.9782055020332336,
407
+ "logits/rejected": 1.0214025974273682,
408
+ "logps/chosen": -374.83990478515625,
409
+ "logps/rejected": -407.66644287109375,
410
+ "loss": 0.5943,
411
+ "rewards/accuracies": 0.668749988079071,
412
+ "rewards/chosen": -1.666853666305542,
413
+ "rewards/margins": 0.4476628303527832,
414
+ "rewards/rejected": -2.1145167350769043,
415
+ "step": 280
416
+ },
417
+ {
418
+ "epoch": 0.15,
419
+ "learning_rate": 4.956323837155325e-07,
420
+ "logits/chosen": 1.1159193515777588,
421
+ "logits/rejected": 1.1185142993927002,
422
+ "logps/chosen": -357.0001220703125,
423
+ "logps/rejected": -394.30279541015625,
424
+ "loss": 0.6107,
425
+ "rewards/accuracies": 0.6625000238418579,
426
+ "rewards/chosen": -1.6487280130386353,
427
+ "rewards/margins": 0.3608878254890442,
428
+ "rewards/rejected": -2.009615898132324,
429
+ "step": 290
430
+ },
431
+ {
432
+ "epoch": 0.16,
433
+ "learning_rate": 4.947278962947386e-07,
434
+ "logits/chosen": 1.6991510391235352,
435
+ "logits/rejected": 1.9209808111190796,
436
+ "logps/chosen": -396.7066955566406,
437
+ "logps/rejected": -431.6412658691406,
438
+ "loss": 0.6001,
439
+ "rewards/accuracies": 0.6875,
440
+ "rewards/chosen": -2.0254971981048584,
441
+ "rewards/margins": 0.42958149313926697,
442
+ "rewards/rejected": -2.455078601837158,
443
+ "step": 300
444
+ },
445
+ {
446
+ "epoch": 0.16,
447
+ "learning_rate": 4.937393407444337e-07,
448
+ "logits/chosen": 1.100517988204956,
449
+ "logits/rejected": 1.2365609407424927,
450
+ "logps/chosen": -356.17681884765625,
451
+ "logps/rejected": -412.0787048339844,
452
+ "loss": 0.5925,
453
+ "rewards/accuracies": 0.65625,
454
+ "rewards/chosen": -1.7612950801849365,
455
+ "rewards/margins": 0.4073117673397064,
456
+ "rewards/rejected": -2.168606758117676,
457
+ "step": 310
458
+ },
459
+ {
460
+ "epoch": 0.17,
461
+ "learning_rate": 4.926670566499992e-07,
462
+ "logits/chosen": 1.4866538047790527,
463
+ "logits/rejected": 1.5181796550750732,
464
+ "logps/chosen": -370.19793701171875,
465
+ "logps/rejected": -399.6379089355469,
466
+ "loss": 0.5865,
467
+ "rewards/accuracies": 0.71875,
468
+ "rewards/chosen": -1.8391116857528687,
469
+ "rewards/margins": 0.3383946418762207,
470
+ "rewards/rejected": -2.1775062084198,
471
+ "step": 320
472
+ },
473
+ {
474
+ "epoch": 0.18,
475
+ "learning_rate": 4.915114123589732e-07,
476
+ "logits/chosen": 3.2927498817443848,
477
+ "logits/rejected": 3.4409637451171875,
478
+ "logps/chosen": -474.81134033203125,
479
+ "logps/rejected": -486.77264404296875,
480
+ "loss": 0.5826,
481
+ "rewards/accuracies": 0.699999988079071,
482
+ "rewards/chosen": -2.5577211380004883,
483
+ "rewards/margins": 0.5131586790084839,
484
+ "rewards/rejected": -3.0708799362182617,
485
+ "step": 330
486
+ },
487
+ {
488
+ "epoch": 0.18,
489
+ "learning_rate": 4.90272804854517e-07,
490
+ "logits/chosen": 4.935120582580566,
491
+ "logits/rejected": 5.241981029510498,
492
+ "logps/chosen": -558.8336181640625,
493
+ "logps/rejected": -587.92578125,
494
+ "loss": 0.6196,
495
+ "rewards/accuracies": 0.675000011920929,
496
+ "rewards/chosen": -3.244816303253174,
497
+ "rewards/margins": 0.4380508065223694,
498
+ "rewards/rejected": -3.6828665733337402,
499
+ "step": 340
500
+ },
501
+ {
502
+ "epoch": 0.19,
503
+ "learning_rate": 4.889516596190448e-07,
504
+ "logits/chosen": 3.1185314655303955,
505
+ "logits/rejected": 3.6742959022521973,
506
+ "logps/chosen": -524.4172973632812,
507
+ "logps/rejected": -528.1964111328125,
508
+ "loss": 0.602,
509
+ "rewards/accuracies": 0.699999988079071,
510
+ "rewards/chosen": -2.8538341522216797,
511
+ "rewards/margins": 0.4846366345882416,
512
+ "rewards/rejected": -3.338470935821533,
513
+ "step": 350
514
+ },
515
+ {
516
+ "epoch": 0.19,
517
+ "learning_rate": 4.875484304880629e-07,
518
+ "logits/chosen": 1.2977662086486816,
519
+ "logits/rejected": 1.6007190942764282,
520
+ "logps/chosen": -468.362060546875,
521
+ "logps/rejected": -483.639892578125,
522
+ "loss": 0.6246,
523
+ "rewards/accuracies": 0.71875,
524
+ "rewards/chosen": -2.143881320953369,
525
+ "rewards/margins": 0.36468884348869324,
526
+ "rewards/rejected": -2.5085701942443848,
527
+ "step": 360
528
+ },
529
+ {
530
+ "epoch": 0.2,
531
+ "learning_rate": 4.860635994942702e-07,
532
+ "logits/chosen": 1.5786962509155273,
533
+ "logits/rejected": 1.9745759963989258,
534
+ "logps/chosen": -395.7223205566406,
535
+ "logps/rejected": -411.2752990722656,
536
+ "loss": 0.5782,
537
+ "rewards/accuracies": 0.6937500238418579,
538
+ "rewards/chosen": -1.866349458694458,
539
+ "rewards/margins": 0.4314461350440979,
540
+ "rewards/rejected": -2.297795534133911,
541
+ "step": 370
542
+ },
543
+ {
544
+ "epoch": 0.2,
545
+ "learning_rate": 4.844976767019714e-07,
546
+ "logits/chosen": 2.636617660522461,
547
+ "logits/rejected": 2.620858669281006,
548
+ "logps/chosen": -369.3752136230469,
549
+ "logps/rejected": -416.45574951171875,
550
+ "loss": 0.576,
551
+ "rewards/accuracies": 0.6875,
552
+ "rewards/chosen": -2.1540982723236084,
553
+ "rewards/margins": 0.4857994019985199,
554
+ "rewards/rejected": -2.6398978233337402,
555
+ "step": 380
556
+ },
557
+ {
558
+ "epoch": 0.21,
559
+ "learning_rate": 4.828512000318616e-07,
560
+ "logits/chosen": 2.6557507514953613,
561
+ "logits/rejected": 3.367990016937256,
562
+ "logps/chosen": -490.51708984375,
563
+ "logps/rejected": -534.2076416015625,
564
+ "loss": 0.5691,
565
+ "rewards/accuracies": 0.75,
566
+ "rewards/chosen": -2.498900890350342,
567
+ "rewards/margins": 0.7616626620292664,
568
+ "rewards/rejected": -3.260563373565674,
569
+ "step": 390
570
+ },
571
+ {
572
+ "epoch": 0.21,
573
+ "learning_rate": 4.811247350762418e-07,
574
+ "logits/chosen": 2.237056016921997,
575
+ "logits/rejected": 2.7285072803497314,
576
+ "logps/chosen": -381.19000244140625,
577
+ "logps/rejected": -417.8363342285156,
578
+ "loss": 0.5659,
579
+ "rewards/accuracies": 0.6875,
580
+ "rewards/chosen": -1.9070937633514404,
581
+ "rewards/margins": 0.5606156587600708,
582
+ "rewards/rejected": -2.46770977973938,
583
+ "step": 400
584
+ },
585
+ {
586
+ "epoch": 0.22,
587
+ "learning_rate": 4.79318874904728e-07,
588
+ "logits/chosen": 4.107834339141846,
589
+ "logits/rejected": 4.020180702209473,
590
+ "logps/chosen": -422.3435974121094,
591
+ "logps/rejected": -466.9422912597656,
592
+ "loss": 0.5826,
593
+ "rewards/accuracies": 0.71875,
594
+ "rewards/chosen": -2.08524751663208,
595
+ "rewards/margins": 0.5145748853683472,
596
+ "rewards/rejected": -2.599822521209717,
597
+ "step": 410
598
+ },
599
+ {
600
+ "epoch": 0.22,
601
+ "learning_rate": 4.774342398605221e-07,
602
+ "logits/chosen": 4.6263508796691895,
603
+ "logits/rejected": 5.094324588775635,
604
+ "logps/chosen": -447.66790771484375,
605
+ "logps/rejected": -472.98541259765625,
606
+ "loss": 0.5864,
607
+ "rewards/accuracies": 0.737500011920929,
608
+ "rewards/chosen": -2.5026144981384277,
609
+ "rewards/margins": 0.5022978782653809,
610
+ "rewards/rejected": -3.0049126148223877,
611
+ "step": 420
612
+ },
613
+ {
614
+ "epoch": 0.23,
615
+ "learning_rate": 4.754714773473134e-07,
616
+ "logits/chosen": 3.137293577194214,
617
+ "logits/rejected": 3.4239907264709473,
618
+ "logps/chosen": -383.85272216796875,
619
+ "logps/rejected": -439.3240661621094,
620
+ "loss": 0.5614,
621
+ "rewards/accuracies": 0.6875,
622
+ "rewards/chosen": -1.8029075860977173,
623
+ "rewards/margins": 0.6045055389404297,
624
+ "rewards/rejected": -2.4074130058288574,
625
+ "step": 430
626
+ },
627
+ {
628
+ "epoch": 0.23,
629
+ "learning_rate": 4.734312616068851e-07,
630
+ "logits/chosen": 3.090550184249878,
631
+ "logits/rejected": 3.005370855331421,
632
+ "logps/chosen": -393.83636474609375,
633
+ "logps/rejected": -456.08917236328125,
634
+ "loss": 0.5636,
635
+ "rewards/accuracies": 0.699999988079071,
636
+ "rewards/chosen": -2.2565479278564453,
637
+ "rewards/margins": 0.5205822587013245,
638
+ "rewards/rejected": -2.777130365371704,
639
+ "step": 440
640
+ },
641
+ {
642
+ "epoch": 0.24,
643
+ "learning_rate": 4.713142934875005e-07,
644
+ "logits/chosen": 3.673936128616333,
645
+ "logits/rejected": 4.1635026931762695,
646
+ "logps/chosen": -492.8273010253906,
647
+ "logps/rejected": -526.05810546875,
648
+ "loss": 0.5837,
649
+ "rewards/accuracies": 0.699999988079071,
650
+ "rewards/chosen": -2.6836562156677246,
651
+ "rewards/margins": 0.5476377010345459,
652
+ "rewards/rejected": -3.2312939167022705,
653
+ "step": 450
654
+ },
655
+ {
656
+ "epoch": 0.24,
657
+ "learning_rate": 4.6912130020314996e-07,
658
+ "logits/chosen": 3.0116238594055176,
659
+ "logits/rejected": 3.2378668785095215,
660
+ "logps/chosen": -389.37811279296875,
661
+ "logps/rejected": -445.2620544433594,
662
+ "loss": 0.5657,
663
+ "rewards/accuracies": 0.6625000238418579,
664
+ "rewards/chosen": -2.161726951599121,
665
+ "rewards/margins": 0.5274554491043091,
666
+ "rewards/rejected": -2.6891825199127197,
667
+ "step": 460
668
+ },
669
+ {
670
+ "epoch": 0.25,
671
+ "learning_rate": 4.668530350837408e-07,
672
+ "logits/chosen": 2.4322941303253174,
673
+ "logits/rejected": 2.488206148147583,
674
+ "logps/chosen": -389.4209899902344,
675
+ "logps/rejected": -435.225341796875,
676
+ "loss": 0.5843,
677
+ "rewards/accuracies": 0.637499988079071,
678
+ "rewards/chosen": -1.8818559646606445,
679
+ "rewards/margins": 0.40773385763168335,
680
+ "rewards/rejected": -2.2895896434783936,
681
+ "step": 470
682
+ },
683
+ {
684
+ "epoch": 0.25,
685
+ "learning_rate": 4.64510277316316e-07,
686
+ "logits/chosen": 2.517045736312866,
687
+ "logits/rejected": 2.917738676071167,
688
+ "logps/chosen": -429.2723693847656,
689
+ "logps/rejected": -474.98681640625,
690
+ "loss": 0.568,
691
+ "rewards/accuracies": 0.699999988079071,
692
+ "rewards/chosen": -2.0465638637542725,
693
+ "rewards/margins": 0.6840267777442932,
694
+ "rewards/rejected": -2.730590581893921,
695
+ "step": 480
696
+ },
697
+ {
698
+ "epoch": 0.26,
699
+ "learning_rate": 4.6209383167739015e-07,
700
+ "logits/chosen": 2.8990161418914795,
701
+ "logits/rejected": 3.3626670837402344,
702
+ "logps/chosen": -449.22235107421875,
703
+ "logps/rejected": -482.134033203125,
704
+ "loss": 0.5814,
705
+ "rewards/accuracies": 0.6625000238418579,
706
+ "rewards/chosen": -2.392901659011841,
707
+ "rewards/margins": 0.5466801524162292,
708
+ "rewards/rejected": -2.939582109451294,
709
+ "step": 490
710
+ },
711
+ {
712
+ "epoch": 0.27,
713
+ "learning_rate": 4.5960452825649526e-07,
714
+ "logits/chosen": 2.796527147293091,
715
+ "logits/rejected": 2.8022494316101074,
716
+ "logps/chosen": -416.53692626953125,
717
+ "logps/rejected": -469.519287109375,
718
+ "loss": 0.5481,
719
+ "rewards/accuracies": 0.7124999761581421,
720
+ "rewards/chosen": -2.1346237659454346,
721
+ "rewards/margins": 0.5421117544174194,
722
+ "rewards/rejected": -2.6767354011535645,
723
+ "step": 500
724
+ },
725
+ {
726
+ "epoch": 0.27,
727
+ "eval_logits/chosen": 4.160411834716797,
728
+ "eval_logits/rejected": 3.967071294784546,
729
+ "eval_logps/chosen": -386.0565185546875,
730
+ "eval_logps/rejected": -439.2520751953125,
731
+ "eval_loss": 0.6089237928390503,
732
+ "eval_rewards/accuracies": 0.6705063581466675,
733
+ "eval_rewards/chosen": -2.6821529865264893,
734
+ "eval_rewards/margins": 0.44139912724494934,
735
+ "eval_rewards/rejected": -3.123551845550537,
736
+ "eval_runtime": 2833.1592,
737
+ "eval_samples_per_second": 12.369,
738
+ "eval_steps_per_second": 0.387,
739
+ "step": 500
740
+ },
741
+ {
742
+ "epoch": 0.27,
743
+ "learning_rate": 4.570432221710314e-07,
744
+ "logits/chosen": 2.99139666557312,
745
+ "logits/rejected": 2.9831995964050293,
746
+ "logps/chosen": -414.53643798828125,
747
+ "logps/rejected": -447.3922424316406,
748
+ "loss": 0.5896,
749
+ "rewards/accuracies": 0.78125,
750
+ "rewards/chosen": -1.9070537090301514,
751
+ "rewards/margins": 0.6989824175834656,
752
+ "rewards/rejected": -2.6060359477996826,
753
+ "step": 510
754
+ },
755
+ {
756
+ "epoch": 0.28,
757
+ "learning_rate": 4.5441079327251927e-07,
758
+ "logits/chosen": 2.8209099769592285,
759
+ "logits/rejected": 3.297544002532959,
760
+ "logps/chosen": -415.64990234375,
761
+ "logps/rejected": -483.9609375,
762
+ "loss": 0.5711,
763
+ "rewards/accuracies": 0.6812499761581421,
764
+ "rewards/chosen": -2.1026933193206787,
765
+ "rewards/margins": 0.5386781096458435,
766
+ "rewards/rejected": -2.641371726989746,
767
+ "step": 520
768
+ },
769
+ {
770
+ "epoch": 0.28,
771
+ "learning_rate": 4.5170814584435644e-07,
772
+ "logits/chosen": 3.1211986541748047,
773
+ "logits/rejected": 3.634293794631958,
774
+ "logps/chosen": -445.901611328125,
775
+ "logps/rejected": -477.486083984375,
776
+ "loss": 0.5723,
777
+ "rewards/accuracies": 0.7437499761581421,
778
+ "rewards/chosen": -2.098557949066162,
779
+ "rewards/margins": 0.7294325828552246,
780
+ "rewards/rejected": -2.8279905319213867,
781
+ "step": 530
782
+ },
783
+ {
784
+ "epoch": 0.29,
785
+ "learning_rate": 4.4893620829118124e-07,
786
+ "logits/chosen": 4.876758575439453,
787
+ "logits/rejected": 4.832733154296875,
788
+ "logps/chosen": -390.5097351074219,
789
+ "logps/rejected": -458.13433837890625,
790
+ "loss": 0.5553,
791
+ "rewards/accuracies": 0.737500011920929,
792
+ "rewards/chosen": -2.3329362869262695,
793
+ "rewards/margins": 0.6324361562728882,
794
+ "rewards/rejected": -2.965372323989868,
795
+ "step": 540
796
+ },
797
+ {
798
+ "epoch": 0.29,
799
+ "learning_rate": 4.460959328199497e-07,
800
+ "logits/chosen": 5.745593070983887,
801
+ "logits/rejected": 6.544689178466797,
802
+ "logps/chosen": -479.26312255859375,
803
+ "logps/rejected": -547.4561767578125,
804
+ "loss": 0.5592,
805
+ "rewards/accuracies": 0.6937500238418579,
806
+ "rewards/chosen": -2.804473876953125,
807
+ "rewards/margins": 0.6489624977111816,
808
+ "rewards/rejected": -3.4534363746643066,
809
+ "step": 550
810
+ },
811
+ {
812
+ "epoch": 0.3,
813
+ "learning_rate": 4.4318829511283707e-07,
814
+ "logits/chosen": 5.760529518127441,
815
+ "logits/rejected": 5.908038139343262,
816
+ "logps/chosen": -500.1220703125,
817
+ "logps/rejected": -579.2546997070312,
818
+ "loss": 0.5787,
819
+ "rewards/accuracies": 0.731249988079071,
820
+ "rewards/chosen": -3.121411085128784,
821
+ "rewards/margins": 0.5803627967834473,
822
+ "rewards/rejected": -3.7017738819122314,
823
+ "step": 560
824
+ },
825
+ {
826
+ "epoch": 0.3,
827
+ "learning_rate": 4.40214293992074e-07,
828
+ "logits/chosen": 4.968967914581299,
829
+ "logits/rejected": 5.733763694763184,
830
+ "logps/chosen": -449.43182373046875,
831
+ "logps/rejected": -481.9095153808594,
832
+ "loss": 0.5705,
833
+ "rewards/accuracies": 0.7250000238418579,
834
+ "rewards/chosen": -2.2817494869232178,
835
+ "rewards/margins": 0.5884373188018799,
836
+ "rewards/rejected": -2.8701870441436768,
837
+ "step": 570
838
+ },
839
+ {
840
+ "epoch": 0.31,
841
+ "learning_rate": 4.3717495107683516e-07,
842
+ "logits/chosen": 6.087460041046143,
843
+ "logits/rejected": 6.749306678771973,
844
+ "logps/chosen": -554.03173828125,
845
+ "logps/rejected": -600.6712646484375,
846
+ "loss": 0.5706,
847
+ "rewards/accuracies": 0.65625,
848
+ "rewards/chosen": -3.579016923904419,
849
+ "rewards/margins": 0.5537681579589844,
850
+ "rewards/rejected": -4.132784843444824,
851
+ "step": 580
852
+ },
853
+ {
854
+ "epoch": 0.31,
855
+ "learning_rate": 4.340713104322953e-07,
856
+ "logits/chosen": 5.846838474273682,
857
+ "logits/rejected": 6.166060447692871,
858
+ "logps/chosen": -547.8532104492188,
859
+ "logps/rejected": -635.3438110351562,
860
+ "loss": 0.5318,
861
+ "rewards/accuracies": 0.71875,
862
+ "rewards/chosen": -3.4989750385284424,
863
+ "rewards/margins": 0.7716847658157349,
864
+ "rewards/rejected": -4.270659446716309,
865
+ "step": 590
866
+ },
867
+ {
868
+ "epoch": 0.32,
869
+ "learning_rate": 4.3090443821097566e-07,
870
+ "logits/chosen": 4.9042229652404785,
871
+ "logits/rejected": 5.197253704071045,
872
+ "logps/chosen": -483.651123046875,
873
+ "logps/rejected": -557.0797119140625,
874
+ "loss": 0.5436,
875
+ "rewards/accuracies": 0.768750011920929,
876
+ "rewards/chosen": -2.6425604820251465,
877
+ "rewards/margins": 0.6980206370353699,
878
+ "rewards/rejected": -3.340581178665161,
879
+ "step": 600
880
+ },
881
+ {
882
+ "epoch": 0.32,
883
+ "learning_rate": 4.276754222865029e-07,
884
+ "logits/chosen": 4.094830513000488,
885
+ "logits/rejected": 4.941472053527832,
886
+ "logps/chosen": -461.6971130371094,
887
+ "logps/rejected": -490.43377685546875,
888
+ "loss": 0.5627,
889
+ "rewards/accuracies": 0.71875,
890
+ "rewards/chosen": -2.5023880004882812,
891
+ "rewards/margins": 0.6450821757316589,
892
+ "rewards/rejected": -3.147469997406006,
893
+ "step": 610
894
+ },
895
+ {
896
+ "epoch": 0.33,
897
+ "learning_rate": 4.2438537187990565e-07,
898
+ "logits/chosen": 4.136041641235352,
899
+ "logits/rejected": 4.636130332946777,
900
+ "logps/chosen": -462.34136962890625,
901
+ "logps/rejected": -513.6154174804688,
902
+ "loss": 0.5457,
903
+ "rewards/accuracies": 0.6937500238418579,
904
+ "rewards/chosen": -2.5274147987365723,
905
+ "rewards/margins": 0.8207530975341797,
906
+ "rewards/rejected": -3.348167896270752,
907
+ "step": 620
908
+ },
909
+ {
910
+ "epoch": 0.33,
911
+ "learning_rate": 4.210354171785795e-07,
912
+ "logits/chosen": 4.823207855224609,
913
+ "logits/rejected": 5.325142860412598,
914
+ "logps/chosen": -488.6162109375,
915
+ "logps/rejected": -548.1345825195312,
916
+ "loss": 0.5523,
917
+ "rewards/accuracies": 0.7562500238418579,
918
+ "rewards/chosen": -2.699159622192383,
919
+ "rewards/margins": 0.712811291217804,
920
+ "rewards/rejected": -3.4119713306427,
921
+ "step": 630
922
+ },
923
+ {
924
+ "epoch": 0.34,
925
+ "learning_rate": 4.1762670894804775e-07,
926
+ "logits/chosen": 4.793757915496826,
927
+ "logits/rejected": 5.079339027404785,
928
+ "logps/chosen": -517.7249755859375,
929
+ "logps/rejected": -555.1099243164062,
930
+ "loss": 0.5609,
931
+ "rewards/accuracies": 0.6937500238418579,
932
+ "rewards/chosen": -3.0854828357696533,
933
+ "rewards/margins": 0.5394307374954224,
934
+ "rewards/rejected": -3.624913454055786,
935
+ "step": 640
936
+ },
937
+ {
938
+ "epoch": 0.34,
939
+ "learning_rate": 4.1416041813665493e-07,
940
+ "logits/chosen": 4.005798816680908,
941
+ "logits/rejected": 4.380357265472412,
942
+ "logps/chosen": -465.93701171875,
943
+ "logps/rejected": -518.0279541015625,
944
+ "loss": 0.5483,
945
+ "rewards/accuracies": 0.6499999761581421,
946
+ "rewards/chosen": -2.6373207569122314,
947
+ "rewards/margins": 0.5303686857223511,
948
+ "rewards/rejected": -3.167689323425293,
949
+ "step": 650
950
+ },
951
+ {
952
+ "epoch": 0.35,
953
+ "learning_rate": 4.1063773547332584e-07,
954
+ "logits/chosen": 4.460536956787109,
955
+ "logits/rejected": 4.873723030090332,
956
+ "logps/chosen": -462.9532775878906,
957
+ "logps/rejected": -519.1263427734375,
958
+ "loss": 0.5651,
959
+ "rewards/accuracies": 0.6625000238418579,
960
+ "rewards/chosen": -2.6806063652038574,
961
+ "rewards/margins": 0.6171140670776367,
962
+ "rewards/rejected": -3.2977206707000732,
963
+ "step": 660
964
+ },
965
+ {
966
+ "epoch": 0.36,
967
+ "learning_rate": 4.0705987105853077e-07,
968
+ "logits/chosen": 4.531912803649902,
969
+ "logits/rejected": 4.875602722167969,
970
+ "logps/chosen": -430.66015625,
971
+ "logps/rejected": -490.43505859375,
972
+ "loss": 0.5664,
973
+ "rewards/accuracies": 0.699999988079071,
974
+ "rewards/chosen": -2.494072437286377,
975
+ "rewards/margins": 0.5989924669265747,
976
+ "rewards/rejected": -3.0930652618408203,
977
+ "step": 670
978
+ },
979
+ {
980
+ "epoch": 0.36,
981
+ "learning_rate": 4.034280539485952e-07,
982
+ "logits/chosen": 3.7882587909698486,
983
+ "logits/rejected": 4.701764106750488,
984
+ "logps/chosen": -472.9560546875,
985
+ "logps/rejected": -545.2316284179688,
986
+ "loss": 0.5388,
987
+ "rewards/accuracies": 0.737500011920929,
988
+ "rewards/chosen": -2.2864925861358643,
989
+ "rewards/margins": 0.8441749811172485,
990
+ "rewards/rejected": -3.130667209625244,
991
+ "step": 680
992
+ },
993
+ {
994
+ "epoch": 0.37,
995
+ "learning_rate": 3.997435317334988e-07,
996
+ "logits/chosen": 5.6066789627075195,
997
+ "logits/rejected": 6.343667507171631,
998
+ "logps/chosen": -546.0533447265625,
999
+ "logps/rejected": -598.7782592773438,
1000
+ "loss": 0.5491,
1001
+ "rewards/accuracies": 0.6812499761581421,
1002
+ "rewards/chosen": -3.1853551864624023,
1003
+ "rewards/margins": 0.7388944625854492,
1004
+ "rewards/rejected": -3.9242501258850098,
1005
+ "step": 690
1006
+ },
1007
+ {
1008
+ "epoch": 0.37,
1009
+ "learning_rate": 3.960075701083074e-07,
1010
+ "logits/chosen": 6.01334285736084,
1011
+ "logits/rejected": 6.114634037017822,
1012
+ "logps/chosen": -553.4805908203125,
1013
+ "logps/rejected": -612.23095703125,
1014
+ "loss": 0.5694,
1015
+ "rewards/accuracies": 0.668749988079071,
1016
+ "rewards/chosen": -3.704514980316162,
1017
+ "rewards/margins": 0.5731467008590698,
1018
+ "rewards/rejected": -4.27766227722168,
1019
+ "step": 700
1020
+ },
1021
+ {
1022
+ "epoch": 0.38,
1023
+ "learning_rate": 3.92221452438385e-07,
1024
+ "logits/chosen": 5.397694110870361,
1025
+ "logits/rejected": 5.5348615646362305,
1026
+ "logps/chosen": -530.5993041992188,
1027
+ "logps/rejected": -563.1759033203125,
1028
+ "loss": 0.5445,
1029
+ "rewards/accuracies": 0.6937500238418579,
1030
+ "rewards/chosen": -3.2607131004333496,
1031
+ "rewards/margins": 0.5572507977485657,
1032
+ "rewards/rejected": -3.817964553833008,
1033
+ "step": 710
1034
+ },
1035
+ {
1036
+ "epoch": 0.38,
1037
+ "learning_rate": 3.8838647931853684e-07,
1038
+ "logits/chosen": 5.427793025970459,
1039
+ "logits/rejected": 5.946757793426514,
1040
+ "logps/chosen": -589.3401489257812,
1041
+ "logps/rejected": -668.6383666992188,
1042
+ "loss": 0.5744,
1043
+ "rewards/accuracies": 0.6937500238418579,
1044
+ "rewards/chosen": -3.8846659660339355,
1045
+ "rewards/margins": 0.7481729388237,
1046
+ "rewards/rejected": -4.632839679718018,
1047
+ "step": 720
1048
+ },
1049
+ {
1050
+ "epoch": 0.39,
1051
+ "learning_rate": 3.845039681262332e-07,
1052
+ "logits/chosen": 5.047539710998535,
1053
+ "logits/rejected": 5.657217025756836,
1054
+ "logps/chosen": -554.4708862304688,
1055
+ "logps/rejected": -616.8558349609375,
1056
+ "loss": 0.5279,
1057
+ "rewards/accuracies": 0.731249988079071,
1058
+ "rewards/chosen": -3.465214967727661,
1059
+ "rewards/margins": 0.6963861584663391,
1060
+ "rewards/rejected": -4.1616010665893555,
1061
+ "step": 730
1062
+ },
1063
+ {
1064
+ "epoch": 0.39,
1065
+ "learning_rate": 3.805752525690681e-07,
1066
+ "logits/chosen": 5.46529483795166,
1067
+ "logits/rejected": 5.889483451843262,
1068
+ "logps/chosen": -564.59228515625,
1069
+ "logps/rejected": -660.7513427734375,
1070
+ "loss": 0.5431,
1071
+ "rewards/accuracies": 0.768750011920929,
1072
+ "rewards/chosen": -3.814227342605591,
1073
+ "rewards/margins": 0.7818762063980103,
1074
+ "rewards/rejected": -4.596103191375732,
1075
+ "step": 740
1076
+ },
1077
+ {
1078
+ "epoch": 0.4,
1079
+ "learning_rate": 3.7660168222660824e-07,
1080
+ "logits/chosen": 5.152674198150635,
1081
+ "logits/rejected": 5.510188102722168,
1082
+ "logps/chosen": -587.24560546875,
1083
+ "logps/rejected": -626.5244140625,
1084
+ "loss": 0.5788,
1085
+ "rewards/accuracies": 0.699999988079071,
1086
+ "rewards/chosen": -3.4678547382354736,
1087
+ "rewards/margins": 0.5621030926704407,
1088
+ "rewards/rejected": -4.0299577713012695,
1089
+ "step": 750
1090
+ },
1091
+ {
1092
+ "epoch": 0.4,
1093
+ "learning_rate": 3.725846220867901e-07,
1094
+ "logits/chosen": 4.9715166091918945,
1095
+ "logits/rejected": 5.927099227905273,
1096
+ "logps/chosen": -540.8013916015625,
1097
+ "logps/rejected": -573.8363037109375,
1098
+ "loss": 0.5665,
1099
+ "rewards/accuracies": 0.6625000238418579,
1100
+ "rewards/chosen": -3.4430599212646484,
1101
+ "rewards/margins": 0.5011590123176575,
1102
+ "rewards/rejected": -3.944218873977661,
1103
+ "step": 760
1104
+ },
1105
+ {
1106
+ "epoch": 0.41,
1107
+ "learning_rate": 3.6852545207702393e-07,
1108
+ "logits/chosen": 5.292831897735596,
1109
+ "logits/rejected": 6.226684093475342,
1110
+ "logps/chosen": -585.9905395507812,
1111
+ "logps/rejected": -621.8675537109375,
1112
+ "loss": 0.5763,
1113
+ "rewards/accuracies": 0.75,
1114
+ "rewards/chosen": -3.4966049194335938,
1115
+ "rewards/margins": 0.7289837598800659,
1116
+ "rewards/rejected": -4.225588798522949,
1117
+ "step": 770
1118
+ },
1119
+ {
1120
+ "epoch": 0.41,
1121
+ "learning_rate": 3.6442556659016475e-07,
1122
+ "logits/chosen": 5.347090244293213,
1123
+ "logits/rejected": 6.540925025939941,
1124
+ "logps/chosen": -568.36572265625,
1125
+ "logps/rejected": -611.6026611328125,
1126
+ "loss": 0.5235,
1127
+ "rewards/accuracies": 0.7562500238418579,
1128
+ "rewards/chosen": -3.558441162109375,
1129
+ "rewards/margins": 0.7847374081611633,
1130
+ "rewards/rejected": -4.343178749084473,
1131
+ "step": 780
1132
+ },
1133
+ {
1134
+ "epoch": 0.42,
1135
+ "learning_rate": 3.602863740055161e-07,
1136
+ "logits/chosen": 5.637946128845215,
1137
+ "logits/rejected": 6.208116054534912,
1138
+ "logps/chosen": -585.2955322265625,
1139
+ "logps/rejected": -676.1448974609375,
1140
+ "loss": 0.5476,
1141
+ "rewards/accuracies": 0.7749999761581421,
1142
+ "rewards/chosen": -3.8783912658691406,
1143
+ "rewards/margins": 0.8531737327575684,
1144
+ "rewards/rejected": -4.731564998626709,
1145
+ "step": 790
1146
+ },
1147
+ {
1148
+ "epoch": 0.42,
1149
+ "learning_rate": 3.5610929620502747e-07,
1150
+ "logits/chosen": 4.267381191253662,
1151
+ "logits/rejected": 5.2271528244018555,
1152
+ "logps/chosen": -578.3743286132812,
1153
+ "logps/rejected": -665.9542846679688,
1154
+ "loss": 0.5381,
1155
+ "rewards/accuracies": 0.6875,
1156
+ "rewards/chosen": -3.779019594192505,
1157
+ "rewards/margins": 0.8331762552261353,
1158
+ "rewards/rejected": -4.61219596862793,
1159
+ "step": 800
1160
+ },
1161
+ {
1162
+ "epoch": 0.43,
1163
+ "learning_rate": 3.5189576808485404e-07,
1164
+ "logits/chosen": 4.498036861419678,
1165
+ "logits/rejected": 5.546736717224121,
1166
+ "logps/chosen": -622.6543579101562,
1167
+ "logps/rejected": -662.39990234375,
1168
+ "loss": 0.5631,
1169
+ "rewards/accuracies": 0.7562500238418579,
1170
+ "rewards/chosen": -3.9034676551818848,
1171
+ "rewards/margins": 0.7405894994735718,
1172
+ "rewards/rejected": -4.644057273864746,
1173
+ "step": 810
1174
+ },
1175
+ {
1176
+ "epoch": 0.44,
1177
+ "learning_rate": 3.476472370624464e-07,
1178
+ "logits/chosen": 5.130145072937012,
1179
+ "logits/rejected": 5.417891502380371,
1180
+ "logps/chosen": -601.1119384765625,
1181
+ "logps/rejected": -665.1697387695312,
1182
+ "loss": 0.5375,
1183
+ "rewards/accuracies": 0.71875,
1184
+ "rewards/chosen": -4.171017646789551,
1185
+ "rewards/margins": 0.6281455159187317,
1186
+ "rewards/rejected": -4.799162864685059,
1187
+ "step": 820
1188
+ },
1189
+ {
1190
+ "epoch": 0.44,
1191
+ "learning_rate": 3.43365162579338e-07,
1192
+ "logits/chosen": 5.479987144470215,
1193
+ "logits/rejected": 5.379258155822754,
1194
+ "logps/chosen": -553.6651611328125,
1195
+ "logps/rejected": -656.02880859375,
1196
+ "loss": 0.5448,
1197
+ "rewards/accuracies": 0.71875,
1198
+ "rewards/chosen": -3.905271053314209,
1199
+ "rewards/margins": 0.8390300869941711,
1200
+ "rewards/rejected": -4.744300842285156,
1201
+ "step": 830
1202
+ },
1203
+ {
1204
+ "epoch": 0.45,
1205
+ "learning_rate": 3.390510155998023e-07,
1206
+ "logits/chosen": 5.235553741455078,
1207
+ "logits/rejected": 5.778375148773193,
1208
+ "logps/chosen": -578.1744384765625,
1209
+ "logps/rejected": -612.1620483398438,
1210
+ "loss": 0.5477,
1211
+ "rewards/accuracies": 0.7124999761581421,
1212
+ "rewards/chosen": -3.762741804122925,
1213
+ "rewards/margins": 0.507239043712616,
1214
+ "rewards/rejected": -4.269980430603027,
1215
+ "step": 840
1216
+ },
1217
+ {
1218
+ "epoch": 0.45,
1219
+ "learning_rate": 3.347062781055526e-07,
1220
+ "logits/chosen": 5.177485942840576,
1221
+ "logits/rejected": 5.3884077072143555,
1222
+ "logps/chosen": -517.0951538085938,
1223
+ "logps/rejected": -611.8797607421875,
1224
+ "loss": 0.5674,
1225
+ "rewards/accuracies": 0.675000011920929,
1226
+ "rewards/chosen": -3.408946990966797,
1227
+ "rewards/margins": 0.6858997344970703,
1228
+ "rewards/rejected": -4.094846725463867,
1229
+ "step": 850
1230
+ },
1231
+ {
1232
+ "epoch": 0.46,
1233
+ "learning_rate": 3.303324425866559e-07,
1234
+ "logits/chosen": 4.779083728790283,
1235
+ "logits/rejected": 4.99451208114624,
1236
+ "logps/chosen": -534.8001708984375,
1237
+ "logps/rejected": -588.9842529296875,
1238
+ "loss": 0.5692,
1239
+ "rewards/accuracies": 0.731249988079071,
1240
+ "rewards/chosen": -3.043729543685913,
1241
+ "rewards/margins": 0.7375171184539795,
1242
+ "rewards/rejected": -3.7812466621398926,
1243
+ "step": 860
1244
+ },
1245
+ {
1246
+ "epoch": 0.46,
1247
+ "learning_rate": 3.2593101152883795e-07,
1248
+ "logits/chosen": 4.893560409545898,
1249
+ "logits/rejected": 4.912612438201904,
1250
+ "logps/chosen": -477.8765563964844,
1251
+ "logps/rejected": -577.2595825195312,
1252
+ "loss": 0.5559,
1253
+ "rewards/accuracies": 0.7875000238418579,
1254
+ "rewards/chosen": -2.9390878677368164,
1255
+ "rewards/margins": 0.735144853591919,
1256
+ "rewards/rejected": -3.6742329597473145,
1257
+ "step": 870
1258
+ },
1259
+ {
1260
+ "epoch": 0.47,
1261
+ "learning_rate": 3.21503496897354e-07,
1262
+ "logits/chosen": 4.999721527099609,
1263
+ "logits/rejected": 5.3239426612854,
1264
+ "logps/chosen": -539.9871215820312,
1265
+ "logps/rejected": -568.0904541015625,
1266
+ "loss": 0.5742,
1267
+ "rewards/accuracies": 0.65625,
1268
+ "rewards/chosen": -3.2623775005340576,
1269
+ "rewards/margins": 0.4343135356903076,
1270
+ "rewards/rejected": -3.6966910362243652,
1271
+ "step": 880
1272
+ },
1273
+ {
1274
+ "epoch": 0.47,
1275
+ "learning_rate": 3.170514196176037e-07,
1276
+ "logits/chosen": 4.89313268661499,
1277
+ "logits/rejected": 4.976138114929199,
1278
+ "logps/chosen": -531.0595703125,
1279
+ "logps/rejected": -615.2716064453125,
1280
+ "loss": 0.5382,
1281
+ "rewards/accuracies": 0.71875,
1282
+ "rewards/chosen": -3.470829486846924,
1283
+ "rewards/margins": 0.6645902395248413,
1284
+ "rewards/rejected": -4.1354193687438965,
1285
+ "step": 890
1286
+ },
1287
+ {
1288
+ "epoch": 0.48,
1289
+ "learning_rate": 3.125763090526674e-07,
1290
+ "logits/chosen": 4.8970746994018555,
1291
+ "logits/rejected": 5.335020542144775,
1292
+ "logps/chosen": -570.1527709960938,
1293
+ "logps/rejected": -649.9376831054688,
1294
+ "loss": 0.5099,
1295
+ "rewards/accuracies": 0.7562500238418579,
1296
+ "rewards/chosen": -3.5448403358459473,
1297
+ "rewards/margins": 0.861894965171814,
1298
+ "rewards/rejected": -4.406734943389893,
1299
+ "step": 900
1300
+ },
1301
+ {
1302
+ "epoch": 0.48,
1303
+ "learning_rate": 3.080797024779447e-07,
1304
+ "logits/chosen": 4.570401191711426,
1305
+ "logits/rejected": 5.245053291320801,
1306
+ "logps/chosen": -538.8987426757812,
1307
+ "logps/rejected": -598.7467651367188,
1308
+ "loss": 0.529,
1309
+ "rewards/accuracies": 0.768750011920929,
1310
+ "rewards/chosen": -3.5106894969940186,
1311
+ "rewards/margins": 0.7811245918273926,
1312
+ "rewards/rejected": -4.29181432723999,
1313
+ "step": 910
1314
+ },
1315
+ {
1316
+ "epoch": 0.49,
1317
+ "learning_rate": 3.035631445530743e-07,
1318
+ "logits/chosen": 5.412999629974365,
1319
+ "logits/rejected": 5.755091190338135,
1320
+ "logps/chosen": -584.6710205078125,
1321
+ "logps/rejected": -669.8858642578125,
1322
+ "loss": 0.5345,
1323
+ "rewards/accuracies": 0.699999988079071,
1324
+ "rewards/chosen": -3.7473666667938232,
1325
+ "rewards/margins": 0.8245238065719604,
1326
+ "rewards/rejected": -4.571890830993652,
1327
+ "step": 920
1328
+ },
1329
+ {
1330
+ "epoch": 0.49,
1331
+ "learning_rate": 2.9902818679131775e-07,
1332
+ "logits/chosen": 5.077193260192871,
1333
+ "logits/rejected": 5.424740791320801,
1334
+ "logps/chosen": -592.6712646484375,
1335
+ "logps/rejected": -663.4064331054688,
1336
+ "loss": 0.5308,
1337
+ "rewards/accuracies": 0.7437499761581421,
1338
+ "rewards/chosen": -4.113033294677734,
1339
+ "rewards/margins": 0.7320185899734497,
1340
+ "rewards/rejected": -4.845052242279053,
1341
+ "step": 930
1342
+ },
1343
+ {
1344
+ "epoch": 0.5,
1345
+ "learning_rate": 2.944763870265886e-07,
1346
+ "logits/chosen": 4.157199859619141,
1347
+ "logits/rejected": 5.055262565612793,
1348
+ "logps/chosen": -593.670166015625,
1349
+ "logps/rejected": -670.92578125,
1350
+ "loss": 0.5311,
1351
+ "rewards/accuracies": 0.75,
1352
+ "rewards/chosen": -3.861306667327881,
1353
+ "rewards/margins": 0.8220443725585938,
1354
+ "rewards/rejected": -4.683350563049316,
1355
+ "step": 940
1356
+ },
1357
+ {
1358
+ "epoch": 0.5,
1359
+ "learning_rate": 2.899093088783105e-07,
1360
+ "logits/chosen": 4.214892864227295,
1361
+ "logits/rejected": 4.872607707977295,
1362
+ "logps/chosen": -627.0443115234375,
1363
+ "logps/rejected": -696.4767456054688,
1364
+ "loss": 0.5336,
1365
+ "rewards/accuracies": 0.7124999761581421,
1366
+ "rewards/chosen": -3.9733481407165527,
1367
+ "rewards/margins": 0.7376152276992798,
1368
+ "rewards/rejected": -4.710963726043701,
1369
+ "step": 950
1370
+ },
1371
+ {
1372
+ "epoch": 0.51,
1373
+ "learning_rate": 2.8532852121428733e-07,
1374
+ "logits/chosen": 4.1724419593811035,
1375
+ "logits/rejected": 4.470523834228516,
1376
+ "logps/chosen": -547.5701904296875,
1377
+ "logps/rejected": -630.979736328125,
1378
+ "loss": 0.5351,
1379
+ "rewards/accuracies": 0.768750011920929,
1380
+ "rewards/chosen": -3.5739071369171143,
1381
+ "rewards/margins": 0.9093371629714966,
1382
+ "rewards/rejected": -4.4832444190979,
1383
+ "step": 960
1384
+ },
1385
+ {
1386
+ "epoch": 0.51,
1387
+ "learning_rate": 2.807355976117716e-07,
1388
+ "logits/chosen": 3.997878313064575,
1389
+ "logits/rejected": 4.5441575050354,
1390
+ "logps/chosen": -573.764892578125,
1391
+ "logps/rejected": -626.8602294921875,
1392
+ "loss": 0.5637,
1393
+ "rewards/accuracies": 0.706250011920929,
1394
+ "rewards/chosen": -3.4168217182159424,
1395
+ "rewards/margins": 0.7197359800338745,
1396
+ "rewards/rejected": -4.1365580558776855,
1397
+ "step": 970
1398
+ },
1399
+ {
1400
+ "epoch": 0.52,
1401
+ "learning_rate": 2.761321158169134e-07,
1402
+ "logits/chosen": 3.821613311767578,
1403
+ "logits/rejected": 4.272428512573242,
1404
+ "logps/chosen": -539.5350341796875,
1405
+ "logps/rejected": -624.6610717773438,
1406
+ "loss": 0.5437,
1407
+ "rewards/accuracies": 0.7749999761581421,
1408
+ "rewards/chosen": -3.425632953643799,
1409
+ "rewards/margins": 0.7200672030448914,
1410
+ "rewards/rejected": -4.145700454711914,
1411
+ "step": 980
1412
+ },
1413
+ {
1414
+ "epoch": 0.53,
1415
+ "learning_rate": 2.715196572027789e-07,
1416
+ "logits/chosen": 3.830516815185547,
1417
+ "logits/rejected": 3.83209228515625,
1418
+ "logps/chosen": -553.0765380859375,
1419
+ "logps/rejected": -614.8712768554688,
1420
+ "loss": 0.5591,
1421
+ "rewards/accuracies": 0.7250000238418579,
1422
+ "rewards/chosen": -3.4271316528320312,
1423
+ "rewards/margins": 0.6863579750061035,
1424
+ "rewards/rejected": -4.113489627838135,
1425
+ "step": 990
1426
+ },
1427
+ {
1428
+ "epoch": 0.53,
1429
+ "learning_rate": 2.6689980622612204e-07,
1430
+ "logits/chosen": 4.027700424194336,
1431
+ "logits/rejected": 4.4086809158325195,
1432
+ "logps/chosen": -504.0587463378906,
1433
+ "logps/rejected": -589.3375854492188,
1434
+ "loss": 0.5519,
1435
+ "rewards/accuracies": 0.706250011920929,
1436
+ "rewards/chosen": -3.1788105964660645,
1437
+ "rewards/margins": 0.6995212435722351,
1438
+ "rewards/rejected": -3.8783316612243652,
1439
+ "step": 1000
1440
+ },
1441
+ {
1442
+ "epoch": 0.53,
1443
+ "eval_logits/chosen": 5.348609447479248,
1444
+ "eval_logits/rejected": 5.197441101074219,
1445
+ "eval_logps/chosen": -543.0738525390625,
1446
+ "eval_logps/rejected": -602.8671264648438,
1447
+ "eval_loss": 0.5866538882255554,
1448
+ "eval_rewards/accuracies": 0.6894388794898987,
1449
+ "eval_rewards/chosen": -4.252326488494873,
1450
+ "eval_rewards/margins": 0.5073760151863098,
1451
+ "eval_rewards/rejected": -4.759702205657959,
1452
+ "eval_runtime": 2829.1466,
1453
+ "eval_samples_per_second": 12.387,
1454
+ "eval_steps_per_second": 0.387,
1455
+ "step": 1000
1456
+ },
1457
+ {
1458
+ "epoch": 0.54,
1459
+ "learning_rate": 2.622741498830969e-07,
1460
+ "logits/chosen": 4.356446266174316,
1461
+ "logits/rejected": 4.390655517578125,
1462
+ "logps/chosen": -531.9536743164062,
1463
+ "logps/rejected": -595.4317626953125,
1464
+ "loss": 0.5685,
1465
+ "rewards/accuracies": 0.71875,
1466
+ "rewards/chosen": -3.2892327308654785,
1467
+ "rewards/margins": 0.6568211913108826,
1468
+ "rewards/rejected": -3.9460537433624268,
1469
+ "step": 1010
1470
+ },
1471
+ {
1472
+ "epoch": 0.54,
1473
+ "learning_rate": 2.5764427716409815e-07,
1474
+ "logits/chosen": 3.622380495071411,
1475
+ "logits/rejected": 4.40954065322876,
1476
+ "logps/chosen": -515.4381103515625,
1477
+ "logps/rejected": -565.3131103515625,
1478
+ "loss": 0.5428,
1479
+ "rewards/accuracies": 0.6937500238418579,
1480
+ "rewards/chosen": -3.0056488513946533,
1481
+ "rewards/margins": 0.683438777923584,
1482
+ "rewards/rejected": -3.6890876293182373,
1483
+ "step": 1020
1484
+ },
1485
+ {
1486
+ "epoch": 0.55,
1487
+ "learning_rate": 2.5301177850791616e-07,
1488
+ "logits/chosen": 3.9004855155944824,
1489
+ "logits/rejected": 4.826648712158203,
1490
+ "logps/chosen": -568.6829833984375,
1491
+ "logps/rejected": -630.1458740234375,
1492
+ "loss": 0.5394,
1493
+ "rewards/accuracies": 0.7124999761581421,
1494
+ "rewards/chosen": -3.4316725730895996,
1495
+ "rewards/margins": 0.825964093208313,
1496
+ "rewards/rejected": -4.257637023925781,
1497
+ "step": 1030
1498
+ },
1499
+ {
1500
+ "epoch": 0.55,
1501
+ "learning_rate": 2.4837824525539477e-07,
1502
+ "logits/chosen": 3.6745972633361816,
1503
+ "logits/rejected": 4.580860614776611,
1504
+ "logps/chosen": -539.4935302734375,
1505
+ "logps/rejected": -591.1002197265625,
1506
+ "loss": 0.5549,
1507
+ "rewards/accuracies": 0.731249988079071,
1508
+ "rewards/chosen": -3.340437650680542,
1509
+ "rewards/margins": 0.6588428020477295,
1510
+ "rewards/rejected": -3.9992804527282715,
1511
+ "step": 1040
1512
+ },
1513
+ {
1514
+ "epoch": 0.56,
1515
+ "learning_rate": 2.4374526910277886e-07,
1516
+ "logits/chosen": 4.059652805328369,
1517
+ "logits/rejected": 4.569832801818848,
1518
+ "logps/chosen": -524.096435546875,
1519
+ "logps/rejected": -599.4136962890625,
1520
+ "loss": 0.566,
1521
+ "rewards/accuracies": 0.7124999761581421,
1522
+ "rewards/chosen": -3.2275619506835938,
1523
+ "rewards/margins": 0.767623782157898,
1524
+ "rewards/rejected": -3.9951858520507812,
1525
+ "step": 1050
1526
+ },
1527
+ {
1528
+ "epoch": 0.56,
1529
+ "learning_rate": 2.391144415549403e-07,
1530
+ "logits/chosen": 4.6053290367126465,
1531
+ "logits/rejected": 4.822667121887207,
1532
+ "logps/chosen": -529.7521362304688,
1533
+ "logps/rejected": -581.6422729492188,
1534
+ "loss": 0.5563,
1535
+ "rewards/accuracies": 0.6499999761581421,
1536
+ "rewards/chosen": -3.478646755218506,
1537
+ "rewards/margins": 0.4808022081851959,
1538
+ "rewards/rejected": -3.959449052810669,
1539
+ "step": 1060
1540
+ },
1541
+ {
1542
+ "epoch": 0.57,
1543
+ "learning_rate": 2.3448735337866919e-07,
1544
+ "logits/chosen": 4.680068016052246,
1545
+ "logits/rejected": 5.09098482131958,
1546
+ "logps/chosen": -486.0177307128906,
1547
+ "logps/rejected": -539.49658203125,
1548
+ "loss": 0.5976,
1549
+ "rewards/accuracies": 0.6875,
1550
+ "rewards/chosen": -3.101942539215088,
1551
+ "rewards/margins": 0.5219632983207703,
1552
+ "rewards/rejected": -3.623905897140503,
1553
+ "step": 1070
1554
+ },
1555
+ {
1556
+ "epoch": 0.57,
1557
+ "learning_rate": 2.2986559405621886e-07,
1558
+ "logits/chosen": 4.115077495574951,
1559
+ "logits/rejected": 4.646478652954102,
1560
+ "logps/chosen": -508.98480224609375,
1561
+ "logps/rejected": -547.0084838867188,
1562
+ "loss": 0.5705,
1563
+ "rewards/accuracies": 0.699999988079071,
1564
+ "rewards/chosen": -2.912672281265259,
1565
+ "rewards/margins": 0.5327907204627991,
1566
+ "rewards/rejected": -3.445463180541992,
1567
+ "step": 1080
1568
+ },
1569
+ {
1570
+ "epoch": 0.58,
1571
+ "learning_rate": 2.2525075123929213e-07,
1572
+ "logits/chosen": 4.412012100219727,
1573
+ "logits/rejected": 4.801966190338135,
1574
+ "logps/chosen": -494.9298400878906,
1575
+ "logps/rejected": -553.3126220703125,
1576
+ "loss": 0.53,
1577
+ "rewards/accuracies": 0.737500011920929,
1578
+ "rewards/chosen": -3.0377612113952637,
1579
+ "rewards/margins": 0.6897190809249878,
1580
+ "rewards/rejected": -3.72748064994812,
1581
+ "step": 1090
1582
+ },
1583
+ {
1584
+ "epoch": 0.58,
1585
+ "learning_rate": 2.206444102036565e-07,
1586
+ "logits/chosen": 4.899998664855957,
1587
+ "logits/rejected": 5.011576175689697,
1588
+ "logps/chosen": -533.6759643554688,
1589
+ "logps/rejected": -604.164306640625,
1590
+ "loss": 0.5622,
1591
+ "rewards/accuracies": 0.71875,
1592
+ "rewards/chosen": -3.346374034881592,
1593
+ "rewards/margins": 0.5992105603218079,
1594
+ "rewards/rejected": -3.945584774017334,
1595
+ "step": 1100
1596
+ },
1597
+ {
1598
+ "epoch": 0.59,
1599
+ "learning_rate": 2.160481533045751e-07,
1600
+ "logits/chosen": 4.868978977203369,
1601
+ "logits/rejected": 5.403256416320801,
1602
+ "logps/chosen": -569.0045776367188,
1603
+ "logps/rejected": -625.4342041015625,
1604
+ "loss": 0.547,
1605
+ "rewards/accuracies": 0.699999988079071,
1606
+ "rewards/chosen": -3.556847333908081,
1607
+ "rewards/margins": 0.7111159563064575,
1608
+ "rewards/rejected": -4.267963409423828,
1609
+ "step": 1110
1610
+ },
1611
+ {
1612
+ "epoch": 0.59,
1613
+ "learning_rate": 2.1146355943324148e-07,
1614
+ "logits/chosen": 5.191876411437988,
1615
+ "logits/rejected": 5.507510185241699,
1616
+ "logps/chosen": -557.5202026367188,
1617
+ "logps/rejected": -637.4208374023438,
1618
+ "loss": 0.5456,
1619
+ "rewards/accuracies": 0.71875,
1620
+ "rewards/chosen": -3.72900652885437,
1621
+ "rewards/margins": 0.7132843732833862,
1622
+ "rewards/rejected": -4.442290782928467,
1623
+ "step": 1120
1624
+ },
1625
+ {
1626
+ "epoch": 0.6,
1627
+ "learning_rate": 2.0689220347440374e-07,
1628
+ "logits/chosen": 4.4830732345581055,
1629
+ "logits/rejected": 5.013645172119141,
1630
+ "logps/chosen": -587.7266235351562,
1631
+ "logps/rejected": -625.2584228515625,
1632
+ "loss": 0.5546,
1633
+ "rewards/accuracies": 0.7124999761581421,
1634
+ "rewards/chosen": -3.556133985519409,
1635
+ "rewards/margins": 0.6282913684844971,
1636
+ "rewards/rejected": -4.184426307678223,
1637
+ "step": 1130
1638
+ },
1639
+ {
1640
+ "epoch": 0.6,
1641
+ "learning_rate": 2.0233565576536564e-07,
1642
+ "logits/chosen": 4.331673622131348,
1643
+ "logits/rejected": 4.588160037994385,
1644
+ "logps/chosen": -579.3844604492188,
1645
+ "logps/rejected": -643.15380859375,
1646
+ "loss": 0.5457,
1647
+ "rewards/accuracies": 0.6812499761581421,
1648
+ "rewards/chosen": -3.580660581588745,
1649
+ "rewards/margins": 0.58534175157547,
1650
+ "rewards/rejected": -4.16600227355957,
1651
+ "step": 1140
1652
+ },
1653
+ {
1654
+ "epoch": 0.61,
1655
+ "learning_rate": 1.97795481556549e-07,
1656
+ "logits/chosen": 3.8659698963165283,
1657
+ "logits/rejected": 4.494038105010986,
1658
+ "logps/chosen": -526.2404174804688,
1659
+ "logps/rejected": -560.0923461914062,
1660
+ "loss": 0.5521,
1661
+ "rewards/accuracies": 0.7124999761581421,
1662
+ "rewards/chosen": -3.2603347301483154,
1663
+ "rewards/margins": 0.6364775896072388,
1664
+ "rewards/rejected": -3.8968119621276855,
1665
+ "step": 1150
1666
+ },
1667
+ {
1668
+ "epoch": 0.62,
1669
+ "learning_rate": 1.9327324047380422e-07,
1670
+ "logits/chosen": 3.7696309089660645,
1671
+ "logits/rejected": 4.594429016113281,
1672
+ "logps/chosen": -504.9520568847656,
1673
+ "logps/rejected": -586.5645751953125,
1674
+ "loss": 0.5119,
1675
+ "rewards/accuracies": 0.7437499761581421,
1676
+ "rewards/chosen": -3.0566916465759277,
1677
+ "rewards/margins": 0.8342288732528687,
1678
+ "rewards/rejected": -3.8909201622009277,
1679
+ "step": 1160
1680
+ },
1681
+ {
1682
+ "epoch": 0.62,
1683
+ "learning_rate": 1.887704859826528e-07,
1684
+ "logits/chosen": 4.6678924560546875,
1685
+ "logits/rejected": 5.033539772033691,
1686
+ "logps/chosen": -579.7088012695312,
1687
+ "logps/rejected": -635.5621948242188,
1688
+ "loss": 0.5359,
1689
+ "rewards/accuracies": 0.65625,
1690
+ "rewards/chosen": -3.5427486896514893,
1691
+ "rewards/margins": 0.6539695262908936,
1692
+ "rewards/rejected": -4.196718692779541,
1693
+ "step": 1170
1694
+ },
1695
+ {
1696
+ "epoch": 0.63,
1697
+ "learning_rate": 1.8428876485464572e-07,
1698
+ "logits/chosen": 4.120070934295654,
1699
+ "logits/rejected": 4.802687644958496,
1700
+ "logps/chosen": -546.7890014648438,
1701
+ "logps/rejected": -595.0975341796875,
1702
+ "loss": 0.5792,
1703
+ "rewards/accuracies": 0.6812499761581421,
1704
+ "rewards/chosen": -3.6790192127227783,
1705
+ "rewards/margins": 0.6043803095817566,
1706
+ "rewards/rejected": -4.28339958190918,
1707
+ "step": 1180
1708
+ },
1709
+ {
1710
+ "epoch": 0.63,
1711
+ "learning_rate": 1.798296166360216e-07,
1712
+ "logits/chosen": 4.212487697601318,
1713
+ "logits/rejected": 5.026057243347168,
1714
+ "logps/chosen": -570.1341552734375,
1715
+ "logps/rejected": -631.1887817382812,
1716
+ "loss": 0.554,
1717
+ "rewards/accuracies": 0.7562500238418579,
1718
+ "rewards/chosen": -3.4938530921936035,
1719
+ "rewards/margins": 0.8111255764961243,
1720
+ "rewards/rejected": -4.304978370666504,
1721
+ "step": 1190
1722
+ },
1723
+ {
1724
+ "epoch": 0.64,
1725
+ "learning_rate": 1.7539457311884675e-07,
1726
+ "logits/chosen": 4.184502601623535,
1727
+ "logits/rejected": 4.767962455749512,
1728
+ "logps/chosen": -532.1156005859375,
1729
+ "logps/rejected": -596.5296630859375,
1730
+ "loss": 0.5289,
1731
+ "rewards/accuracies": 0.75,
1732
+ "rewards/chosen": -3.360400438308716,
1733
+ "rewards/margins": 0.7973008155822754,
1734
+ "rewards/rejected": -4.157701015472412,
1735
+ "step": 1200
1736
+ },
1737
+ {
1738
+ "epoch": 0.64,
1739
+ "learning_rate": 1.7098515781481883e-07,
1740
+ "logits/chosen": 5.043318748474121,
1741
+ "logits/rejected": 5.477365970611572,
1742
+ "logps/chosen": -554.041015625,
1743
+ "logps/rejected": -602.6400146484375,
1744
+ "loss": 0.5381,
1745
+ "rewards/accuracies": 0.675000011920929,
1746
+ "rewards/chosen": -3.55816388130188,
1747
+ "rewards/margins": 0.6002733111381531,
1748
+ "rewards/rejected": -4.1584367752075195,
1749
+ "step": 1210
1750
+ },
1751
+ {
1752
+ "epoch": 0.65,
1753
+ "learning_rate": 1.6660288543191568e-07,
1754
+ "logits/chosen": 4.342132568359375,
1755
+ "logits/rejected": 5.391766548156738,
1756
+ "logps/chosen": -545.3358764648438,
1757
+ "logps/rejected": -593.21044921875,
1758
+ "loss": 0.5273,
1759
+ "rewards/accuracies": 0.71875,
1760
+ "rewards/chosen": -3.2604167461395264,
1761
+ "rewards/margins": 0.7034675478935242,
1762
+ "rewards/rejected": -3.9638848304748535,
1763
+ "step": 1220
1764
+ },
1765
+ {
1766
+ "epoch": 0.65,
1767
+ "learning_rate": 1.6224926135406693e-07,
1768
+ "logits/chosen": 4.523472785949707,
1769
+ "logits/rejected": 4.9561238288879395,
1770
+ "logps/chosen": -541.8357543945312,
1771
+ "logps/rejected": -617.3576049804688,
1772
+ "loss": 0.5279,
1773
+ "rewards/accuracies": 0.7124999761581421,
1774
+ "rewards/chosen": -3.246079206466675,
1775
+ "rewards/margins": 0.7863640785217285,
1776
+ "rewards/rejected": -4.032443046569824,
1777
+ "step": 1230
1778
+ },
1779
+ {
1780
+ "epoch": 0.66,
1781
+ "learning_rate": 1.579257811240298e-07,
1782
+ "logits/chosen": 4.174516201019287,
1783
+ "logits/rejected": 4.948155403137207,
1784
+ "logps/chosen": -566.966796875,
1785
+ "logps/rejected": -629.2576293945312,
1786
+ "loss": 0.5383,
1787
+ "rewards/accuracies": 0.6937500238418579,
1788
+ "rewards/chosen": -3.5496108531951904,
1789
+ "rewards/margins": 0.6297346949577332,
1790
+ "rewards/rejected": -4.179345607757568,
1791
+ "step": 1240
1792
+ },
1793
+ {
1794
+ "epoch": 0.66,
1795
+ "learning_rate": 1.5363392992964523e-07,
1796
+ "logits/chosen": 4.812657833099365,
1797
+ "logits/rejected": 5.1759843826293945,
1798
+ "logps/chosen": -545.2940673828125,
1799
+ "logps/rejected": -610.6871948242188,
1800
+ "loss": 0.5599,
1801
+ "rewards/accuracies": 0.6812499761581421,
1802
+ "rewards/chosen": -3.605778217315674,
1803
+ "rewards/margins": 0.5686632394790649,
1804
+ "rewards/rejected": -4.174441337585449,
1805
+ "step": 1250
1806
+ },
1807
+ {
1808
+ "epoch": 0.67,
1809
+ "learning_rate": 1.4937518209365108e-07,
1810
+ "logits/chosen": 4.399840354919434,
1811
+ "logits/rejected": 5.024414539337158,
1812
+ "logps/chosen": -565.787353515625,
1813
+ "logps/rejected": -621.1328125,
1814
+ "loss": 0.5422,
1815
+ "rewards/accuracies": 0.71875,
1816
+ "rewards/chosen": -3.319702625274658,
1817
+ "rewards/margins": 0.7489235997200012,
1818
+ "rewards/rejected": -4.0686259269714355,
1819
+ "step": 1260
1820
+ },
1821
+ {
1822
+ "epoch": 0.67,
1823
+ "learning_rate": 1.4515100056722708e-07,
1824
+ "logits/chosen": 4.322859287261963,
1825
+ "logits/rejected": 4.647487163543701,
1826
+ "logps/chosen": -473.44915771484375,
1827
+ "logps/rejected": -556.26806640625,
1828
+ "loss": 0.5335,
1829
+ "rewards/accuracies": 0.762499988079071,
1830
+ "rewards/chosen": -2.9221551418304443,
1831
+ "rewards/margins": 0.8288917541503906,
1832
+ "rewards/rejected": -3.7510476112365723,
1833
+ "step": 1270
1834
+ },
1835
+ {
1836
+ "epoch": 0.68,
1837
+ "learning_rate": 1.4096283642744716e-07,
1838
+ "logits/chosen": 4.445618152618408,
1839
+ "logits/rejected": 5.310964107513428,
1840
+ "logps/chosen": -529.9594116210938,
1841
+ "logps/rejected": -583.6397094726562,
1842
+ "loss": 0.5489,
1843
+ "rewards/accuracies": 0.6875,
1844
+ "rewards/chosen": -3.098684310913086,
1845
+ "rewards/margins": 0.7321470975875854,
1846
+ "rewards/rejected": -3.8308310508728027,
1847
+ "step": 1280
1848
+ },
1849
+ {
1850
+ "epoch": 0.68,
1851
+ "learning_rate": 1.3681212837880977e-07,
1852
+ "logits/chosen": 4.398858070373535,
1853
+ "logits/rejected": 4.8625688552856445,
1854
+ "logps/chosen": -509.6946716308594,
1855
+ "logps/rejected": -594.6348876953125,
1856
+ "loss": 0.5397,
1857
+ "rewards/accuracies": 0.71875,
1858
+ "rewards/chosen": -2.9974119663238525,
1859
+ "rewards/margins": 0.8787434697151184,
1860
+ "rewards/rejected": -3.876155376434326,
1861
+ "step": 1290
1862
+ },
1863
+ {
1864
+ "epoch": 0.69,
1865
+ "learning_rate": 1.3270030225901908e-07,
1866
+ "logits/chosen": 4.207759857177734,
1867
+ "logits/rejected": 5.357243537902832,
1868
+ "logps/chosen": -540.0505981445312,
1869
+ "logps/rejected": -591.9456787109375,
1870
+ "loss": 0.5405,
1871
+ "rewards/accuracies": 0.7749999761581421,
1872
+ "rewards/chosen": -2.9670040607452393,
1873
+ "rewards/margins": 0.9191484451293945,
1874
+ "rewards/rejected": -3.8861522674560547,
1875
+ "step": 1300
1876
+ },
1877
+ {
1878
+ "epoch": 0.7,
1879
+ "learning_rate": 1.2862877054918572e-07,
1880
+ "logits/chosen": 3.810641050338745,
1881
+ "logits/rejected": 4.167244911193848,
1882
+ "logps/chosen": -455.54962158203125,
1883
+ "logps/rejected": -525.954833984375,
1884
+ "loss": 0.558,
1885
+ "rewards/accuracies": 0.699999988079071,
1886
+ "rewards/chosen": -2.536816358566284,
1887
+ "rewards/margins": 0.7640112638473511,
1888
+ "rewards/rejected": -3.300827741622925,
1889
+ "step": 1310
1890
+ },
1891
+ {
1892
+ "epoch": 0.7,
1893
+ "learning_rate": 1.2459893188861613e-07,
1894
+ "logits/chosen": 3.5092411041259766,
1895
+ "logits/rejected": 4.095297813415527,
1896
+ "logps/chosen": -398.01702880859375,
1897
+ "logps/rejected": -486.40106201171875,
1898
+ "loss": 0.5194,
1899
+ "rewards/accuracies": 0.7250000238418579,
1900
+ "rewards/chosen": -2.352292060852051,
1901
+ "rewards/margins": 0.8707995414733887,
1902
+ "rewards/rejected": -3.2230911254882812,
1903
+ "step": 1320
1904
+ },
1905
+ {
1906
+ "epoch": 0.71,
1907
+ "learning_rate": 1.206121705943558e-07,
1908
+ "logits/chosen": 3.8648457527160645,
1909
+ "logits/rejected": 4.79974365234375,
1910
+ "logps/chosen": -450.3089904785156,
1911
+ "logps/rejected": -489.1319274902344,
1912
+ "loss": 0.5501,
1913
+ "rewards/accuracies": 0.762499988079071,
1914
+ "rewards/chosen": -2.3793253898620605,
1915
+ "rewards/margins": 0.6881033778190613,
1916
+ "rewards/rejected": -3.0674290657043457,
1917
+ "step": 1330
1918
+ },
1919
+ {
1920
+ "epoch": 0.71,
1921
+ "learning_rate": 1.1666985618565422e-07,
1922
+ "logits/chosen": 4.9529547691345215,
1923
+ "logits/rejected": 5.19863224029541,
1924
+ "logps/chosen": -469.70855712890625,
1925
+ "logps/rejected": -529.8267822265625,
1926
+ "loss": 0.5422,
1927
+ "rewards/accuracies": 0.6875,
1928
+ "rewards/chosen": -2.916710615158081,
1929
+ "rewards/margins": 0.6288083791732788,
1930
+ "rewards/rejected": -3.545518398284912,
1931
+ "step": 1340
1932
+ },
1933
+ {
1934
+ "epoch": 0.72,
1935
+ "learning_rate": 1.1277334291351145e-07,
1936
+ "logits/chosen": 5.23069953918457,
1937
+ "logits/rejected": 5.482142925262451,
1938
+ "logps/chosen": -469.21307373046875,
1939
+ "logps/rejected": -589.8045654296875,
1940
+ "loss": 0.5183,
1941
+ "rewards/accuracies": 0.75,
1942
+ "rewards/chosen": -2.9651408195495605,
1943
+ "rewards/margins": 0.8228071331977844,
1944
+ "rewards/rejected": -3.7879481315612793,
1945
+ "step": 1350
1946
+ },
1947
+ {
1948
+ "epoch": 0.72,
1949
+ "learning_rate": 1.089239692954701e-07,
1950
+ "logits/chosen": 4.438393592834473,
1951
+ "logits/rejected": 5.2009711265563965,
1952
+ "logps/chosen": -496.31463623046875,
1953
+ "logps/rejected": -561.5128173828125,
1954
+ "loss": 0.5497,
1955
+ "rewards/accuracies": 0.7124999761581421,
1956
+ "rewards/chosen": -2.9592137336730957,
1957
+ "rewards/margins": 0.6417844891548157,
1958
+ "rewards/rejected": -3.6009984016418457,
1959
+ "step": 1360
1960
+ },
1961
+ {
1962
+ "epoch": 0.73,
1963
+ "learning_rate": 1.051230576558127e-07,
1964
+ "logits/chosen": 5.329151630401611,
1965
+ "logits/rejected": 5.1115946769714355,
1966
+ "logps/chosen": -510.32379150390625,
1967
+ "logps/rejected": -633.5717163085938,
1968
+ "loss": 0.5281,
1969
+ "rewards/accuracies": 0.78125,
1970
+ "rewards/chosen": -3.305920124053955,
1971
+ "rewards/margins": 0.6949362754821777,
1972
+ "rewards/rejected": -4.000856399536133,
1973
+ "step": 1370
1974
+ },
1975
+ {
1976
+ "epoch": 0.73,
1977
+ "learning_rate": 1.0137191367132078e-07,
1978
+ "logits/chosen": 4.635811805725098,
1979
+ "logits/rejected": 5.594052314758301,
1980
+ "logps/chosen": -538.316650390625,
1981
+ "logps/rejected": -593.9783325195312,
1982
+ "loss": 0.5179,
1983
+ "rewards/accuracies": 0.7124999761581421,
1984
+ "rewards/chosen": -3.257719039916992,
1985
+ "rewards/margins": 0.8566193580627441,
1986
+ "rewards/rejected": -4.1143388748168945,
1987
+ "step": 1380
1988
+ },
1989
+ {
1990
+ "epoch": 0.74,
1991
+ "learning_rate": 9.76718259227532e-08,
1992
+ "logits/chosen": 5.040810585021973,
1993
+ "logits/rejected": 5.872241973876953,
1994
+ "logps/chosen": -559.94140625,
1995
+ "logps/rejected": -600.4876098632812,
1996
+ "loss": 0.5346,
1997
+ "rewards/accuracies": 0.7437499761581421,
1998
+ "rewards/chosen": -3.3652939796447754,
1999
+ "rewards/margins": 0.7677735090255737,
2000
+ "rewards/rejected": -4.1330671310424805,
2001
+ "step": 1390
2002
+ },
2003
+ {
2004
+ "epoch": 0.74,
2005
+ "learning_rate": 9.402406545219676e-08,
2006
+ "logits/chosen": 5.209423542022705,
2007
+ "logits/rejected": 5.7783403396606445,
2008
+ "logps/chosen": -570.754150390625,
2009
+ "logps/rejected": -631.2447509765625,
2010
+ "loss": 0.5343,
2011
+ "rewards/accuracies": 0.737500011920929,
2012
+ "rewards/chosen": -3.6044132709503174,
2013
+ "rewards/margins": 0.8434345126152039,
2014
+ "rewards/rejected": -4.447848320007324,
2015
+ "step": 1400
2016
+ },
2017
+ {
2018
+ "epoch": 0.75,
2019
+ "learning_rate": 9.042988532644249e-08,
2020
+ "logits/chosen": 5.212072372436523,
2021
+ "logits/rejected": 5.8972063064575195,
2022
+ "logps/chosen": -583.494873046875,
2023
+ "logps/rejected": -651.96630859375,
2024
+ "loss": 0.5139,
2025
+ "rewards/accuracies": 0.78125,
2026
+ "rewards/chosen": -3.3433117866516113,
2027
+ "rewards/margins": 0.9961609840393066,
2028
+ "rewards/rejected": -4.339472770690918,
2029
+ "step": 1410
2030
+ },
2031
+ {
2032
+ "epoch": 0.75,
2033
+ "learning_rate": 8.689052020653592e-08,
2034
+ "logits/chosen": 4.78833532333374,
2035
+ "logits/rejected": 5.89553165435791,
2036
+ "logps/chosen": -561.2659912109375,
2037
+ "logps/rejected": -631.704833984375,
2038
+ "loss": 0.5514,
2039
+ "rewards/accuracies": 0.7124999761581421,
2040
+ "rewards/chosen": -3.425673007965088,
2041
+ "rewards/margins": 0.8838043212890625,
2042
+ "rewards/rejected": -4.309477806091309,
2043
+ "step": 1420
2044
+ },
2045
+ {
2046
+ "epoch": 0.76,
2047
+ "learning_rate": 8.340718592365037e-08,
2048
+ "logits/chosen": 5.993227958679199,
2049
+ "logits/rejected": 6.045858860015869,
2050
+ "logps/chosen": -575.0015869140625,
2051
+ "logps/rejected": -656.5886840820312,
2052
+ "loss": 0.5134,
2053
+ "rewards/accuracies": 0.6812499761581421,
2054
+ "rewards/chosen": -3.8804218769073486,
2055
+ "rewards/margins": 0.6545805931091309,
2056
+ "rewards/rejected": -4.535001754760742,
2057
+ "step": 1430
2058
+ },
2059
+ {
2060
+ "epoch": 0.76,
2061
+ "learning_rate": 7.998107906142839e-08,
2062
+ "logits/chosen": 5.676799774169922,
2063
+ "logits/rejected": 6.391809940338135,
2064
+ "logps/chosen": -568.3199462890625,
2065
+ "logps/rejected": -637.5294189453125,
2066
+ "loss": 0.5254,
2067
+ "rewards/accuracies": 0.71875,
2068
+ "rewards/chosen": -3.7702269554138184,
2069
+ "rewards/margins": 0.8325515985488892,
2070
+ "rewards/rejected": -4.602778434753418,
2071
+ "step": 1440
2072
+ },
2073
+ {
2074
+ "epoch": 0.77,
2075
+ "learning_rate": 7.661337654493575e-08,
2076
+ "logits/chosen": 5.199474334716797,
2077
+ "logits/rejected": 5.9724626541137695,
2078
+ "logps/chosen": -604.0126953125,
2079
+ "logps/rejected": -679.3408203125,
2080
+ "loss": 0.5511,
2081
+ "rewards/accuracies": 0.7562500238418579,
2082
+ "rewards/chosen": -3.842377185821533,
2083
+ "rewards/margins": 0.8208445310592651,
2084
+ "rewards/rejected": -4.663221836090088,
2085
+ "step": 1450
2086
+ },
2087
+ {
2088
+ "epoch": 0.77,
2089
+ "learning_rate": 7.330523523636751e-08,
2090
+ "logits/chosen": 4.888028144836426,
2091
+ "logits/rejected": 5.282634258270264,
2092
+ "logps/chosen": -574.2249755859375,
2093
+ "logps/rejected": -649.12158203125,
2094
+ "loss": 0.5353,
2095
+ "rewards/accuracies": 0.699999988079071,
2096
+ "rewards/chosen": -3.574962615966797,
2097
+ "rewards/margins": 0.7398016452789307,
2098
+ "rewards/rejected": -4.314764499664307,
2099
+ "step": 1460
2100
+ },
2101
+ {
2102
+ "epoch": 0.78,
2103
+ "learning_rate": 7.005779153764682e-08,
2104
+ "logits/chosen": 5.4382123947143555,
2105
+ "logits/rejected": 5.483049392700195,
2106
+ "logps/chosen": -550.1482543945312,
2107
+ "logps/rejected": -629.5477294921875,
2108
+ "loss": 0.5274,
2109
+ "rewards/accuracies": 0.75,
2110
+ "rewards/chosen": -3.7003333568573,
2111
+ "rewards/margins": 0.6894980669021606,
2112
+ "rewards/rejected": -4.389832019805908,
2113
+ "step": 1470
2114
+ },
2115
+ {
2116
+ "epoch": 0.79,
2117
+ "learning_rate": 6.687216100005138e-08,
2118
+ "logits/chosen": 5.554679870605469,
2119
+ "logits/rejected": 6.150310516357422,
2120
+ "logps/chosen": -604.5882568359375,
2121
+ "logps/rejected": -694.1824340820312,
2122
+ "loss": 0.5241,
2123
+ "rewards/accuracies": 0.6812499761581421,
2124
+ "rewards/chosen": -3.837186336517334,
2125
+ "rewards/margins": 0.8338931202888489,
2126
+ "rewards/rejected": -4.671080589294434,
2127
+ "step": 1480
2128
+ },
2129
+ {
2130
+ "epoch": 0.79,
2131
+ "learning_rate": 6.374943794100349e-08,
2132
+ "logits/chosen": 4.995267868041992,
2133
+ "logits/rejected": 6.089870452880859,
2134
+ "logps/chosen": -540.8165283203125,
2135
+ "logps/rejected": -621.503662109375,
2136
+ "loss": 0.5181,
2137
+ "rewards/accuracies": 0.7562500238418579,
2138
+ "rewards/chosen": -3.447101593017578,
2139
+ "rewards/margins": 0.998428463935852,
2140
+ "rewards/rejected": -4.445529937744141,
2141
+ "step": 1490
2142
+ },
2143
+ {
2144
+ "epoch": 0.8,
2145
+ "learning_rate": 6.069069506815325e-08,
2146
+ "logits/chosen": 5.665683269500732,
2147
+ "logits/rejected": 5.919187545776367,
2148
+ "logps/chosen": -557.50341796875,
2149
+ "logps/rejected": -635.5902099609375,
2150
+ "loss": 0.5597,
2151
+ "rewards/accuracies": 0.706250011920929,
2152
+ "rewards/chosen": -3.7898292541503906,
2153
+ "rewards/margins": 0.7044454216957092,
2154
+ "rewards/rejected": -4.494275093078613,
2155
+ "step": 1500
2156
+ },
2157
+ {
2158
+ "epoch": 0.8,
2159
+ "eval_logits/chosen": 6.629395008087158,
2160
+ "eval_logits/rejected": 6.464417934417725,
2161
+ "eval_logps/chosen": -596.9036865234375,
2162
+ "eval_logps/rejected": -659.0733032226562,
2163
+ "eval_loss": 0.5820685625076294,
2164
+ "eval_rewards/accuracies": 0.6959397792816162,
2165
+ "eval_rewards/chosen": -4.79062557220459,
2166
+ "eval_rewards/margins": 0.5311391949653625,
2167
+ "eval_rewards/rejected": -5.321763515472412,
2168
+ "eval_runtime": 2832.2152,
2169
+ "eval_samples_per_second": 12.373,
2170
+ "eval_steps_per_second": 0.387,
2171
+ "step": 1500
2172
+ },
2173
+ {
2174
+ "epoch": 0.8,
2175
+ "learning_rate": 5.7696983110885746e-08,
2176
+ "logits/chosen": 6.144343376159668,
2177
+ "logits/rejected": 6.1946892738342285,
2178
+ "logps/chosen": -568.4108276367188,
2179
+ "logps/rejected": -656.3994750976562,
2180
+ "loss": 0.5713,
2181
+ "rewards/accuracies": 0.7124999761581421,
2182
+ "rewards/chosen": -3.8781726360321045,
2183
+ "rewards/margins": 0.7777887582778931,
2184
+ "rewards/rejected": -4.655961036682129,
2185
+ "step": 1510
2186
+ },
2187
+ {
2188
+ "epoch": 0.81,
2189
+ "learning_rate": 5.47693304593777e-08,
2190
+ "logits/chosen": 5.610565185546875,
2191
+ "logits/rejected": 6.231629371643066,
2192
+ "logps/chosen": -578.0348510742188,
2193
+ "logps/rejected": -634.0302124023438,
2194
+ "loss": 0.5254,
2195
+ "rewards/accuracies": 0.7562500238418579,
2196
+ "rewards/chosen": -3.640714168548584,
2197
+ "rewards/margins": 0.8447046279907227,
2198
+ "rewards/rejected": -4.485418796539307,
2199
+ "step": 1520
2200
+ },
2201
+ {
2202
+ "epoch": 0.81,
2203
+ "learning_rate": 5.190874281132851e-08,
2204
+ "logits/chosen": 5.386014938354492,
2205
+ "logits/rejected": 5.870030403137207,
2206
+ "logps/chosen": -544.43212890625,
2207
+ "logps/rejected": -605.0162963867188,
2208
+ "loss": 0.5313,
2209
+ "rewards/accuracies": 0.71875,
2210
+ "rewards/chosen": -3.533620834350586,
2211
+ "rewards/margins": 0.7047785520553589,
2212
+ "rewards/rejected": -4.238399505615234,
2213
+ "step": 1530
2214
+ },
2215
+ {
2216
+ "epoch": 0.82,
2217
+ "learning_rate": 4.9116202826486045e-08,
2218
+ "logits/chosen": 5.818629741668701,
2219
+ "logits/rejected": 6.154656410217285,
2220
+ "logps/chosen": -583.7796630859375,
2221
+ "logps/rejected": -635.4794921875,
2222
+ "loss": 0.5337,
2223
+ "rewards/accuracies": 0.6812499761581421,
2224
+ "rewards/chosen": -3.777665615081787,
2225
+ "rewards/margins": 0.6795842051506042,
2226
+ "rewards/rejected": -4.457250118255615,
2227
+ "step": 1540
2228
+ },
2229
+ {
2230
+ "epoch": 0.82,
2231
+ "learning_rate": 4.639266978908676e-08,
2232
+ "logits/chosen": 5.711030006408691,
2233
+ "logits/rejected": 6.358799934387207,
2234
+ "logps/chosen": -611.2662353515625,
2235
+ "logps/rejected": -676.8931884765625,
2236
+ "loss": 0.5187,
2237
+ "rewards/accuracies": 0.706250011920929,
2238
+ "rewards/chosen": -3.886317491531372,
2239
+ "rewards/margins": 0.7436650991439819,
2240
+ "rewards/rejected": -4.629981994628906,
2241
+ "step": 1550
2242
+ },
2243
+ {
2244
+ "epoch": 0.83,
2245
+ "learning_rate": 4.373907927832513e-08,
2246
+ "logits/chosen": 5.366969108581543,
2247
+ "logits/rejected": 5.813288688659668,
2248
+ "logps/chosen": -553.2315673828125,
2249
+ "logps/rejected": -654.0559692382812,
2250
+ "loss": 0.5274,
2251
+ "rewards/accuracies": 0.7250000238418579,
2252
+ "rewards/chosen": -3.5649421215057373,
2253
+ "rewards/margins": 0.8516329526901245,
2254
+ "rewards/rejected": -4.416574954986572,
2255
+ "step": 1560
2256
+ },
2257
+ {
2258
+ "epoch": 0.83,
2259
+ "learning_rate": 4.115634284696698e-08,
2260
+ "logits/chosen": 5.473841667175293,
2261
+ "logits/rejected": 6.142959117889404,
2262
+ "logps/chosen": -595.4747314453125,
2263
+ "logps/rejected": -675.5442504882812,
2264
+ "loss": 0.4995,
2265
+ "rewards/accuracies": 0.7749999761581421,
2266
+ "rewards/chosen": -3.8835277557373047,
2267
+ "rewards/margins": 0.8822458386421204,
2268
+ "rewards/rejected": -4.765774250030518,
2269
+ "step": 1570
2270
+ },
2271
+ {
2272
+ "epoch": 0.84,
2273
+ "learning_rate": 3.864534770821559e-08,
2274
+ "logits/chosen": 5.575949668884277,
2275
+ "logits/rejected": 6.3084797859191895,
2276
+ "logps/chosen": -557.6798706054688,
2277
+ "logps/rejected": -628.6642456054688,
2278
+ "loss": 0.5549,
2279
+ "rewards/accuracies": 0.71875,
2280
+ "rewards/chosen": -3.6573452949523926,
2281
+ "rewards/margins": 0.8320104479789734,
2282
+ "rewards/rejected": -4.489355564117432,
2283
+ "step": 1580
2284
+ },
2285
+ {
2286
+ "epoch": 0.84,
2287
+ "learning_rate": 3.620695643093924e-08,
2288
+ "logits/chosen": 5.1766839027404785,
2289
+ "logits/rejected": 6.165072441101074,
2290
+ "logps/chosen": -578.5045166015625,
2291
+ "logps/rejected": -645.0850830078125,
2292
+ "loss": 0.5174,
2293
+ "rewards/accuracies": 0.7437499761581421,
2294
+ "rewards/chosen": -3.7886593341827393,
2295
+ "rewards/margins": 0.9527637362480164,
2296
+ "rewards/rejected": -4.7414231300354,
2297
+ "step": 1590
2298
+ },
2299
+ {
2300
+ "epoch": 0.85,
2301
+ "learning_rate": 3.384200664336412e-08,
2302
+ "logits/chosen": 5.577385425567627,
2303
+ "logits/rejected": 6.328660011291504,
2304
+ "logps/chosen": -591.4779052734375,
2305
+ "logps/rejected": -644.4967041015625,
2306
+ "loss": 0.5339,
2307
+ "rewards/accuracies": 0.7562500238418579,
2308
+ "rewards/chosen": -3.7442920207977295,
2309
+ "rewards/margins": 0.8128402829170227,
2310
+ "rewards/rejected": -4.557132244110107,
2311
+ "step": 1600
2312
+ },
2313
+ {
2314
+ "epoch": 0.85,
2315
+ "learning_rate": 3.155131074533529e-08,
2316
+ "logits/chosen": 4.875297546386719,
2317
+ "logits/rejected": 5.868630886077881,
2318
+ "logps/chosen": -590.7601318359375,
2319
+ "logps/rejected": -644.1740112304688,
2320
+ "loss": 0.5532,
2321
+ "rewards/accuracies": 0.7250000238418579,
2322
+ "rewards/chosen": -3.6774280071258545,
2323
+ "rewards/margins": 0.69183349609375,
2324
+ "rewards/rejected": -4.369261741638184,
2325
+ "step": 1610
2326
+ },
2327
+ {
2328
+ "epoch": 0.86,
2329
+ "learning_rate": 2.9335655629243645e-08,
2330
+ "logits/chosen": 5.1115217208862305,
2331
+ "logits/rejected": 5.944631099700928,
2332
+ "logps/chosen": -572.9088134765625,
2333
+ "logps/rejected": -635.9785766601562,
2334
+ "loss": 0.5635,
2335
+ "rewards/accuracies": 0.7124999761581421,
2336
+ "rewards/chosen": -3.67173433303833,
2337
+ "rewards/margins": 0.6962157487869263,
2338
+ "rewards/rejected": -4.367949962615967,
2339
+ "step": 1620
2340
+ },
2341
+ {
2342
+ "epoch": 0.86,
2343
+ "learning_rate": 2.7195802409715197e-08,
2344
+ "logits/chosen": 5.198755741119385,
2345
+ "logits/rejected": 5.86753511428833,
2346
+ "logps/chosen": -608.7496948242188,
2347
+ "logps/rejected": -661.3677978515625,
2348
+ "loss": 0.5482,
2349
+ "rewards/accuracies": 0.737500011920929,
2350
+ "rewards/chosen": -3.853783130645752,
2351
+ "rewards/margins": 0.7783033847808838,
2352
+ "rewards/rejected": -4.632086753845215,
2353
+ "step": 1630
2354
+ },
2355
+ {
2356
+ "epoch": 0.87,
2357
+ "learning_rate": 2.513248616215527e-08,
2358
+ "logits/chosen": 4.92915153503418,
2359
+ "logits/rejected": 5.753048896789551,
2360
+ "logps/chosen": -574.3338623046875,
2361
+ "logps/rejected": -647.4317626953125,
2362
+ "loss": 0.5148,
2363
+ "rewards/accuracies": 0.793749988079071,
2364
+ "rewards/chosen": -3.5955843925476074,
2365
+ "rewards/margins": 0.8324259519577026,
2366
+ "rewards/rejected": -4.428009986877441,
2367
+ "step": 1640
2368
+ },
2369
+ {
2370
+ "epoch": 0.88,
2371
+ "learning_rate": 2.31464156702382e-08,
2372
+ "logits/chosen": 4.705880165100098,
2373
+ "logits/rejected": 5.527700901031494,
2374
+ "logps/chosen": -576.8226928710938,
2375
+ "logps/rejected": -628.70849609375,
2376
+ "loss": 0.5499,
2377
+ "rewards/accuracies": 0.699999988079071,
2378
+ "rewards/chosen": -3.532841205596924,
2379
+ "rewards/margins": 0.7848026752471924,
2380
+ "rewards/rejected": -4.317643642425537,
2381
+ "step": 1650
2382
+ },
2383
+ {
2384
+ "epoch": 0.88,
2385
+ "learning_rate": 2.1238273182427933e-08,
2386
+ "logits/chosen": 5.681646347045898,
2387
+ "logits/rejected": 6.597903251647949,
2388
+ "logps/chosen": -569.5862426757812,
2389
+ "logps/rejected": -620.6526489257812,
2390
+ "loss": 0.5361,
2391
+ "rewards/accuracies": 0.699999988079071,
2392
+ "rewards/chosen": -3.6919121742248535,
2393
+ "rewards/margins": 0.7138382196426392,
2394
+ "rewards/rejected": -4.405750274658203,
2395
+ "step": 1660
2396
+ },
2397
+ {
2398
+ "epoch": 0.89,
2399
+ "learning_rate": 1.9408714177614306e-08,
2400
+ "logits/chosen": 5.2946391105651855,
2401
+ "logits/rejected": 6.044753074645996,
2402
+ "logps/chosen": -564.3060302734375,
2403
+ "logps/rejected": -626.62353515625,
2404
+ "loss": 0.5135,
2405
+ "rewards/accuracies": 0.762499988079071,
2406
+ "rewards/chosen": -3.511505126953125,
2407
+ "rewards/margins": 0.8757253885269165,
2408
+ "rewards/rejected": -4.38723087310791,
2409
+ "step": 1670
2410
+ },
2411
+ {
2412
+ "epoch": 0.89,
2413
+ "learning_rate": 1.7658367139945228e-08,
2414
+ "logits/chosen": 5.4068779945373535,
2415
+ "logits/rejected": 5.994559288024902,
2416
+ "logps/chosen": -599.1517333984375,
2417
+ "logps/rejected": -650.2825317382812,
2418
+ "loss": 0.5239,
2419
+ "rewards/accuracies": 0.6875,
2420
+ "rewards/chosen": -3.7978172302246094,
2421
+ "rewards/margins": 0.7726759910583496,
2422
+ "rewards/rejected": -4.570493221282959,
2423
+ "step": 1680
2424
+ },
2425
+ {
2426
+ "epoch": 0.9,
2427
+ "learning_rate": 1.5987833342931745e-08,
2428
+ "logits/chosen": 5.3490800857543945,
2429
+ "logits/rejected": 6.137597560882568,
2430
+ "logps/chosen": -583.5009155273438,
2431
+ "logps/rejected": -632.0625,
2432
+ "loss": 0.5467,
2433
+ "rewards/accuracies": 0.768750011920929,
2434
+ "rewards/chosen": -3.8200180530548096,
2435
+ "rewards/margins": 0.7188345193862915,
2436
+ "rewards/rejected": -4.538852691650391,
2437
+ "step": 1690
2438
+ },
2439
+ {
2440
+ "epoch": 0.9,
2441
+ "learning_rate": 1.439768664290053e-08,
2442
+ "logits/chosen": 5.457294940948486,
2443
+ "logits/rejected": 5.772470474243164,
2444
+ "logps/chosen": -603.9612426757812,
2445
+ "logps/rejected": -673.6254272460938,
2446
+ "loss": 0.5239,
2447
+ "rewards/accuracies": 0.71875,
2448
+ "rewards/chosen": -3.798058032989502,
2449
+ "rewards/margins": 0.787087619304657,
2450
+ "rewards/rejected": -4.585144996643066,
2451
+ "step": 1700
2452
+ },
2453
+ {
2454
+ "epoch": 0.91,
2455
+ "learning_rate": 1.2888473281864597e-08,
2456
+ "logits/chosen": 5.0094194412231445,
2457
+ "logits/rejected": 5.467037200927734,
2458
+ "logps/chosen": -566.6424560546875,
2459
+ "logps/rejected": -654.4783935546875,
2460
+ "loss": 0.5305,
2461
+ "rewards/accuracies": 0.71875,
2462
+ "rewards/chosen": -3.774543046951294,
2463
+ "rewards/margins": 0.7515050172805786,
2464
+ "rewards/rejected": -4.526047706604004,
2465
+ "step": 1710
2466
+ },
2467
+ {
2468
+ "epoch": 0.91,
2469
+ "learning_rate": 1.1460711699880082e-08,
2470
+ "logits/chosen": 4.981472015380859,
2471
+ "logits/rejected": 6.010613918304443,
2472
+ "logps/chosen": -587.7794189453125,
2473
+ "logps/rejected": -658.3820190429688,
2474
+ "loss": 0.5277,
2475
+ "rewards/accuracies": 0.7437499761581421,
2476
+ "rewards/chosen": -3.5816383361816406,
2477
+ "rewards/margins": 0.8929538726806641,
2478
+ "rewards/rejected": -4.474592208862305,
2479
+ "step": 1720
2480
+ },
2481
+ {
2482
+ "epoch": 0.92,
2483
+ "learning_rate": 1.0114892356953397e-08,
2484
+ "logits/chosen": 4.724175453186035,
2485
+ "logits/rejected": 5.377623558044434,
2486
+ "logps/chosen": -572.1964111328125,
2487
+ "logps/rejected": -616.9322509765625,
2488
+ "loss": 0.5707,
2489
+ "rewards/accuracies": 0.6937500238418579,
2490
+ "rewards/chosen": -3.6713428497314453,
2491
+ "rewards/margins": 0.6362916231155396,
2492
+ "rewards/rejected": -4.3076348304748535,
2493
+ "step": 1730
2494
+ },
2495
+ {
2496
+ "epoch": 0.92,
2497
+ "learning_rate": 8.851477564560061e-09,
2498
+ "logits/chosen": 5.281374454498291,
2499
+ "logits/rejected": 5.823667526245117,
2500
+ "logps/chosen": -574.9295654296875,
2501
+ "logps/rejected": -682.01611328125,
2502
+ "loss": 0.5272,
2503
+ "rewards/accuracies": 0.7875000238418579,
2504
+ "rewards/chosen": -3.782670497894287,
2505
+ "rewards/margins": 1.0396969318389893,
2506
+ "rewards/rejected": -4.8223676681518555,
2507
+ "step": 1740
2508
+ },
2509
+ {
2510
+ "epoch": 0.93,
2511
+ "learning_rate": 7.670901326832763e-09,
2512
+ "logits/chosen": 5.813240051269531,
2513
+ "logits/rejected": 6.113523006439209,
2514
+ "logps/chosen": -583.0259399414062,
2515
+ "logps/rejected": -707.0408935546875,
2516
+ "loss": 0.5068,
2517
+ "rewards/accuracies": 0.737500011920929,
2518
+ "rewards/chosen": -3.896685838699341,
2519
+ "rewards/margins": 0.9632736444473267,
2520
+ "rewards/rejected": -4.859958648681641,
2521
+ "step": 1750
2522
+ },
2523
+ {
2524
+ "epoch": 0.93,
2525
+ "learning_rate": 6.5735691914738936e-09,
2526
+ "logits/chosen": 4.764095783233643,
2527
+ "logits/rejected": 5.158614158630371,
2528
+ "logps/chosen": -563.710693359375,
2529
+ "logps/rejected": -646.1812744140625,
2530
+ "loss": 0.5635,
2531
+ "rewards/accuracies": 0.699999988079071,
2532
+ "rewards/chosen": -3.6481738090515137,
2533
+ "rewards/margins": 0.7430108785629272,
2534
+ "rewards/rejected": -4.3911848068237305,
2535
+ "step": 1760
2536
+ },
2537
+ {
2538
+ "epoch": 0.94,
2539
+ "learning_rate": 5.559858110443016e-09,
2540
+ "logits/chosen": 4.941697120666504,
2541
+ "logits/rejected": 5.857718467712402,
2542
+ "logps/chosen": -589.8416137695312,
2543
+ "logps/rejected": -651.1143798828125,
2544
+ "loss": 0.5218,
2545
+ "rewards/accuracies": 0.75,
2546
+ "rewards/chosen": -3.781978130340576,
2547
+ "rewards/margins": 0.8180473446846008,
2548
+ "rewards/rejected": -4.600025177001953,
2549
+ "step": 1770
2550
+ },
2551
+ {
2552
+ "epoch": 0.94,
2553
+ "learning_rate": 4.6301163104676685e-09,
2554
+ "logits/chosen": 5.284440517425537,
2555
+ "logits/rejected": 5.553745269775391,
2556
+ "logps/chosen": -551.4322509765625,
2557
+ "logps/rejected": -651.212646484375,
2558
+ "loss": 0.5407,
2559
+ "rewards/accuracies": 0.6875,
2560
+ "rewards/chosen": -3.6018593311309814,
2561
+ "rewards/margins": 0.7678863406181335,
2562
+ "rewards/rejected": -4.36974573135376,
2563
+ "step": 1780
2564
+ },
2565
+ {
2566
+ "epoch": 0.95,
2567
+ "learning_rate": 3.784663173421438e-09,
2568
+ "logits/chosen": 5.374969959259033,
2569
+ "logits/rejected": 5.7753376960754395,
2570
+ "logps/chosen": -622.2989501953125,
2571
+ "logps/rejected": -689.6795654296875,
2572
+ "loss": 0.5308,
2573
+ "rewards/accuracies": 0.7437499761581421,
2574
+ "rewards/chosen": -3.73246693611145,
2575
+ "rewards/margins": 0.8092583417892456,
2576
+ "rewards/rejected": -4.5417256355285645,
2577
+ "step": 1790
2578
+ },
2579
+ {
2580
+ "epoch": 0.96,
2581
+ "learning_rate": 3.023789126611137e-09,
2582
+ "logits/chosen": 5.826882839202881,
2583
+ "logits/rejected": 6.438210964202881,
2584
+ "logps/chosen": -580.4129638671875,
2585
+ "logps/rejected": -642.33740234375,
2586
+ "loss": 0.5188,
2587
+ "rewards/accuracies": 0.7250000238418579,
2588
+ "rewards/chosen": -3.765054702758789,
2589
+ "rewards/margins": 0.8300830721855164,
2590
+ "rewards/rejected": -4.595137596130371,
2591
+ "step": 1800
2592
+ },
2593
+ {
2594
+ "epoch": 0.96,
2595
+ "learning_rate": 2.3477555430100604e-09,
2596
+ "logits/chosen": 5.330014228820801,
2597
+ "logits/rejected": 6.266722679138184,
2598
+ "logps/chosen": -545.0280151367188,
2599
+ "logps/rejected": -623.4840087890625,
2600
+ "loss": 0.5255,
2601
+ "rewards/accuracies": 0.768750011920929,
2602
+ "rewards/chosen": -3.382579803466797,
2603
+ "rewards/margins": 1.0013315677642822,
2604
+ "rewards/rejected": -4.383911609649658,
2605
+ "step": 1810
2606
+ },
2607
+ {
2608
+ "epoch": 0.97,
2609
+ "learning_rate": 1.7567946514721322e-09,
2610
+ "logits/chosen": 5.459120273590088,
2611
+ "logits/rejected": 5.671516418457031,
2612
+ "logps/chosen": -565.1712036132812,
2613
+ "logps/rejected": -651.7200927734375,
2614
+ "loss": 0.5631,
2615
+ "rewards/accuracies": 0.699999988079071,
2616
+ "rewards/chosen": -3.7416863441467285,
2617
+ "rewards/margins": 0.7198008894920349,
2618
+ "rewards/rejected": -4.46148681640625,
2619
+ "step": 1820
2620
+ },
2621
+ {
2622
+ "epoch": 0.97,
2623
+ "learning_rate": 1.2511094569571668e-09,
2624
+ "logits/chosen": 4.77047061920166,
2625
+ "logits/rejected": 5.680732727050781,
2626
+ "logps/chosen": -540.1213989257812,
2627
+ "logps/rejected": -612.4068603515625,
2628
+ "loss": 0.5294,
2629
+ "rewards/accuracies": 0.7437499761581421,
2630
+ "rewards/chosen": -3.508056640625,
2631
+ "rewards/margins": 0.8371505737304688,
2632
+ "rewards/rejected": -4.345207214355469,
2633
+ "step": 1830
2634
+ },
2635
+ {
2636
+ "epoch": 0.98,
2637
+ "learning_rate": 8.308736707954289e-10,
2638
+ "logits/chosen": 5.112320899963379,
2639
+ "logits/rejected": 6.371096611022949,
2640
+ "logps/chosen": -591.5335693359375,
2641
+ "logps/rejected": -622.4219360351562,
2642
+ "loss": 0.5351,
2643
+ "rewards/accuracies": 0.7250000238418579,
2644
+ "rewards/chosen": -3.7684085369110107,
2645
+ "rewards/margins": 0.8042453527450562,
2646
+ "rewards/rejected": -4.572653770446777,
2647
+ "step": 1840
2648
+ },
2649
+ {
2650
+ "epoch": 0.98,
2651
+ "learning_rate": 4.962316510149222e-10,
2652
+ "logits/chosen": 4.924439430236816,
2653
+ "logits/rejected": 5.723077774047852,
2654
+ "logps/chosen": -528.1791381835938,
2655
+ "logps/rejected": -613.794189453125,
2656
+ "loss": 0.5155,
2657
+ "rewards/accuracies": 0.7124999761581421,
2658
+ "rewards/chosen": -3.5361733436584473,
2659
+ "rewards/margins": 0.8941107988357544,
2660
+ "rewards/rejected": -4.43028450012207,
2661
+ "step": 1850
2662
+ },
2663
+ {
2664
+ "epoch": 0.99,
2665
+ "learning_rate": 2.4729835275189016e-10,
2666
+ "logits/chosen": 5.2165093421936035,
2667
+ "logits/rejected": 5.809523105621338,
2668
+ "logps/chosen": -512.5826416015625,
2669
+ "logps/rejected": -562.5357055664062,
2670
+ "loss": 0.5279,
2671
+ "rewards/accuracies": 0.7562500238418579,
2672
+ "rewards/chosen": -3.3880043029785156,
2673
+ "rewards/margins": 0.6139224767684937,
2674
+ "rewards/rejected": -4.001926898956299,
2675
+ "step": 1860
2676
+ },
2677
+ {
2678
+ "epoch": 0.99,
2679
+ "learning_rate": 8.415928876176482e-11,
2680
+ "logits/chosen": 5.226788520812988,
2681
+ "logits/rejected": 5.463950157165527,
2682
+ "logps/chosen": -568.6761474609375,
2683
+ "logps/rejected": -642.3060302734375,
2684
+ "loss": 0.5796,
2685
+ "rewards/accuracies": 0.6812499761581421,
2686
+ "rewards/chosen": -3.871469020843506,
2687
+ "rewards/margins": 0.6007556319236755,
2688
+ "rewards/rejected": -4.472224235534668,
2689
+ "step": 1870
2690
+ },
2691
+ {
2692
+ "epoch": 1.0,
2693
+ "learning_rate": 6.870500044303673e-12,
2694
+ "logits/chosen": 5.113773345947266,
2695
+ "logits/rejected": 5.360236167907715,
2696
+ "logps/chosen": -532.0494995117188,
2697
+ "logps/rejected": -629.88818359375,
2698
+ "loss": 0.537,
2699
+ "rewards/accuracies": 0.7437499761581421,
2700
+ "rewards/chosen": -3.353315830230713,
2701
+ "rewards/margins": 0.727187991142273,
2702
+ "rewards/rejected": -4.080503940582275,
2703
+ "step": 1880
2704
+ },
2705
+ {
2706
+ "epoch": 1.0,
2707
+ "step": 1884,
2708
+ "total_flos": 0.0,
2709
+ "train_loss": 0.5658232939218006,
2710
+ "train_runtime": 45309.3847,
2711
+ "train_samples_per_second": 2.662,
2712
+ "train_steps_per_second": 0.042
2713
+ }
2714
+ ],
2715
+ "logging_steps": 10,
2716
+ "max_steps": 1884,
2717
+ "num_input_tokens_seen": 0,
2718
+ "num_train_epochs": 1,
2719
+ "save_steps": 500,
2720
+ "total_flos": 0.0,
2721
+ "train_batch_size": 4,
2722
+ "trial_name": null,
2723
+ "trial_params": null
2724
+ }