{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 252, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01984126984126984, "grad_norm": 18.047539550790006, "learning_rate": 5.7692307692307695e-08, "logits/chosen": -1.703125, "logits/rejected": -1.6953125, "logps/chosen": -0.376953125, "logps/rejected": -0.427734375, "loss": 2.4555, "rewards/accuracies": 1.0, "rewards/chosen": -3.765625, "rewards/margins": 0.50390625, "rewards/rejected": -4.28125, "step": 5 }, { "epoch": 0.03968253968253968, "grad_norm": 17.185434459136403, "learning_rate": 1.1538461538461539e-07, "logits/chosen": -1.2578125, "logits/rejected": -1.2578125, "logps/chosen": -0.32421875, "logps/rejected": -0.390625, "loss": 2.4857, "rewards/accuracies": 1.0, "rewards/chosen": -3.234375, "rewards/margins": 0.6640625, "rewards/rejected": -3.90625, "step": 10 }, { "epoch": 0.05952380952380952, "grad_norm": 17.039133125856026, "learning_rate": 1.7307692307692305e-07, "logits/chosen": -1.5078125, "logits/rejected": -1.5078125, "logps/chosen": -0.279296875, "logps/rejected": -0.34765625, "loss": 2.402, "rewards/accuracies": 1.0, "rewards/chosen": -2.78125, "rewards/margins": 0.6796875, "rewards/rejected": -3.46875, "step": 15 }, { "epoch": 0.07936507936507936, "grad_norm": 17.10226682999204, "learning_rate": 2.3076923076923078e-07, "logits/chosen": -1.4765625, "logits/rejected": -1.4765625, "logps/chosen": -0.310546875, "logps/rejected": -0.39453125, "loss": 2.4703, "rewards/accuracies": 1.0, "rewards/chosen": -3.09375, "rewards/margins": 0.84375, "rewards/rejected": -3.9375, "step": 20 }, { "epoch": 0.0992063492063492, "grad_norm": 14.47393175785549, "learning_rate": 2.8846153846153846e-07, "logits/chosen": -1.78125, "logits/rejected": -1.78125, "logps/chosen": -0.34375, "logps/rejected": -0.40234375, "loss": 2.3167, "rewards/accuracies": 1.0, "rewards/chosen": -3.421875, "rewards/margins": 0.5859375, "rewards/rejected": -4.0, "step": 25 }, { "epoch": 0.11904761904761904, "grad_norm": 16.9612236979206, "learning_rate": 2.997681792980754e-07, "logits/chosen": -1.3359375, "logits/rejected": -1.3359375, "logps/chosen": -0.3046875, "logps/rejected": -0.357421875, "loss": 2.3476, "rewards/accuracies": 1.0, "rewards/chosen": -3.03125, "rewards/margins": 0.54296875, "rewards/rejected": -3.578125, "step": 30 }, { "epoch": 0.1388888888888889, "grad_norm": 16.48721079930822, "learning_rate": 2.9882763562415516e-07, "logits/chosen": -1.5078125, "logits/rejected": -1.5078125, "logps/chosen": -0.408203125, "logps/rejected": -0.482421875, "loss": 2.391, "rewards/accuracies": 1.0, "rewards/chosen": -4.0625, "rewards/margins": 0.75390625, "rewards/rejected": -4.8125, "step": 35 }, { "epoch": 0.15873015873015872, "grad_norm": 21.37387348387286, "learning_rate": 2.9716841847369104e-07, "logits/chosen": -1.5, "logits/rejected": -1.5, "logps/chosen": -0.2890625, "logps/rejected": -0.365234375, "loss": 2.3684, "rewards/accuracies": 1.0, "rewards/chosen": -2.890625, "rewards/margins": 0.7578125, "rewards/rejected": -3.640625, "step": 40 }, { "epoch": 0.17857142857142858, "grad_norm": 17.871422332796683, "learning_rate": 2.94798540037113e-07, "logits/chosen": -1.546875, "logits/rejected": -1.546875, "logps/chosen": -0.287109375, "logps/rejected": -0.3515625, "loss": 2.3699, "rewards/accuracies": 1.0, "rewards/chosen": -2.859375, "rewards/margins": 0.640625, "rewards/rejected": -3.5, "step": 45 }, { "epoch": 0.1984126984126984, "grad_norm": 29.190246701669263, "learning_rate": 2.917294442157359e-07, "logits/chosen": -1.3828125, "logits/rejected": -1.3828125, "logps/chosen": -0.265625, "logps/rejected": -0.34375, "loss": 2.3773, "rewards/accuracies": 1.0, "rewards/chosen": -2.65625, "rewards/margins": 0.78125, "rewards/rejected": -3.4375, "step": 50 }, { "epoch": 0.21825396825396826, "grad_norm": 20.420712184187625, "learning_rate": 2.8797595136032675e-07, "logits/chosen": -1.4765625, "logits/rejected": -1.4765625, "logps/chosen": -0.31640625, "logps/rejected": -0.39453125, "loss": 2.3586, "rewards/accuracies": 1.0, "rewards/chosen": -3.171875, "rewards/margins": 0.7734375, "rewards/rejected": -3.953125, "step": 55 }, { "epoch": 0.23809523809523808, "grad_norm": 17.704268736560135, "learning_rate": 2.8355618670514256e-07, "logits/chosen": -1.2890625, "logits/rejected": -1.2890625, "logps/chosen": -0.27734375, "logps/rejected": -0.328125, "loss": 2.3299, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -2.765625, "rewards/margins": 0.515625, "rewards/rejected": -3.28125, "step": 60 }, { "epoch": 0.25793650793650796, "grad_norm": 15.278992243473061, "learning_rate": 2.784914928430218e-07, "logits/chosen": -1.40625, "logits/rejected": -1.40625, "logps/chosen": -0.294921875, "logps/rejected": -0.37109375, "loss": 2.3791, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -2.953125, "rewards/margins": 0.7578125, "rewards/rejected": -3.703125, "step": 65 }, { "epoch": 0.2777777777777778, "grad_norm": 21.138997719744832, "learning_rate": 2.728063266641801e-07, "logits/chosen": -1.6484375, "logits/rejected": -1.6484375, "logps/chosen": -0.26953125, "logps/rejected": -0.361328125, "loss": 2.3719, "rewards/accuracies": 1.0, "rewards/chosen": -2.703125, "rewards/margins": 0.9140625, "rewards/rejected": -3.625, "step": 70 }, { "epoch": 0.2976190476190476, "grad_norm": 26.097345440204215, "learning_rate": 2.665281412563814e-07, "logits/chosen": -1.875, "logits/rejected": -1.875, "logps/chosen": -0.26171875, "logps/rejected": -0.37109375, "loss": 2.318, "rewards/accuracies": 1.0, "rewards/chosen": -2.625, "rewards/margins": 1.0859375, "rewards/rejected": -3.703125, "step": 75 }, { "epoch": 0.31746031746031744, "grad_norm": 21.061149357230455, "learning_rate": 2.596872533367763e-07, "logits/chosen": -1.5, "logits/rejected": -1.5, "logps/chosen": -0.28125, "logps/rejected": -0.373046875, "loss": 2.2812, "rewards/accuracies": 1.0, "rewards/chosen": -2.828125, "rewards/margins": 0.89453125, "rewards/rejected": -3.71875, "step": 80 }, { "epoch": 0.3373015873015873, "grad_norm": 34.0182184293618, "learning_rate": 2.5231669685556633e-07, "logits/chosen": -1.1875, "logits/rejected": -1.1875, "logps/chosen": -0.27734375, "logps/rejected": -0.39453125, "loss": 2.166, "rewards/accuracies": 1.0, "rewards/chosen": -2.78125, "rewards/margins": 1.15625, "rewards/rejected": -3.9375, "step": 85 }, { "epoch": 0.35714285714285715, "grad_norm": 20.821123274301424, "learning_rate": 2.444520634784271e-07, "logits/chosen": -1.3359375, "logits/rejected": -1.3359375, "logps/chosen": -0.267578125, "logps/rejected": -0.361328125, "loss": 2.1671, "rewards/accuracies": 1.0, "rewards/chosen": -2.671875, "rewards/margins": 0.9453125, "rewards/rejected": -3.625, "step": 90 }, { "epoch": 0.376984126984127, "grad_norm": 22.90486064935448, "learning_rate": 2.361313307179837e-07, "logits/chosen": -1.7265625, "logits/rejected": -1.7265625, "logps/chosen": -0.279296875, "logps/rejected": -0.353515625, "loss": 2.0894, "rewards/accuracies": 1.0, "rewards/chosen": -2.78125, "rewards/margins": 0.74609375, "rewards/rejected": -3.53125, "step": 95 }, { "epoch": 0.3968253968253968, "grad_norm": 27.68913080972186, "learning_rate": 2.2739467854427513e-07, "logits/chosen": -1.65625, "logits/rejected": -1.65625, "logps/chosen": -0.353515625, "logps/rejected": -0.46875, "loss": 2.1033, "rewards/accuracies": 1.0, "rewards/chosen": -3.546875, "rewards/margins": 1.1484375, "rewards/rejected": -4.6875, "step": 100 }, { "epoch": 0.4166666666666667, "grad_norm": 29.979641972346496, "learning_rate": 2.1828429535977582e-07, "logits/chosen": -1.6796875, "logits/rejected": -1.6796875, "logps/chosen": -0.3203125, "logps/rejected": -0.4296875, "loss": 2.0674, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -3.1875, "rewards/margins": 1.09375, "rewards/rejected": -4.28125, "step": 105 }, { "epoch": 0.4365079365079365, "grad_norm": 44.14626475514875, "learning_rate": 2.0884417427590215e-07, "logits/chosen": -1.65625, "logits/rejected": -1.65625, "logps/chosen": -0.34375, "logps/rejected": -0.46875, "loss": 2.0088, "rewards/accuracies": 1.0, "rewards/chosen": -3.4375, "rewards/margins": 1.2578125, "rewards/rejected": -4.6875, "step": 110 }, { "epoch": 0.45634920634920634, "grad_norm": 33.80229631943301, "learning_rate": 1.9911990067476334e-07, "logits/chosen": -1.15625, "logits/rejected": -1.15625, "logps/chosen": -0.23828125, "logps/rejected": -0.34375, "loss": 1.9458, "rewards/accuracies": 1.0, "rewards/chosen": -2.390625, "rewards/margins": 1.046875, "rewards/rejected": -3.4375, "step": 115 }, { "epoch": 0.47619047619047616, "grad_norm": 39.86801120602329, "learning_rate": 1.8915843208199963e-07, "logits/chosen": -1.625, "logits/rejected": -1.625, "logps/chosen": -0.4296875, "logps/rejected": -0.546875, "loss": 1.9224, "rewards/accuracies": 1.0, "rewards/chosen": -4.3125, "rewards/margins": 1.171875, "rewards/rejected": -5.46875, "step": 120 }, { "epoch": 0.49603174603174605, "grad_norm": 33.949958184137344, "learning_rate": 1.7900787141367918e-07, "logits/chosen": -1.453125, "logits/rejected": -1.453125, "logps/chosen": -0.357421875, "logps/rejected": -0.4765625, "loss": 2.0021, "rewards/accuracies": 1.0, "rewards/chosen": -3.578125, "rewards/margins": 1.1875, "rewards/rejected": -4.78125, "step": 125 }, { "epoch": 0.5158730158730159, "grad_norm": 37.35189108229666, "learning_rate": 1.687172346922213e-07, "logits/chosen": -1.546875, "logits/rejected": -1.546875, "logps/chosen": -0.419921875, "logps/rejected": -0.546875, "loss": 1.7533, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -4.1875, "rewards/margins": 1.2578125, "rewards/rejected": -5.46875, "step": 130 }, { "epoch": 0.5357142857142857, "grad_norm": 33.938784524888526, "learning_rate": 1.5833621435302245e-07, "logits/chosen": -1.734375, "logits/rejected": -1.734375, "logps/chosen": -0.421875, "logps/rejected": -0.63671875, "loss": 1.716, "rewards/accuracies": 1.0, "rewards/chosen": -4.21875, "rewards/margins": 2.15625, "rewards/rejected": -6.34375, "step": 135 }, { "epoch": 0.5555555555555556, "grad_norm": 35.12217870275606, "learning_rate": 1.4791493928475273e-07, "logits/chosen": -1.5234375, "logits/rejected": -1.5234375, "logps/chosen": -0.30078125, "logps/rejected": -0.578125, "loss": 1.7479, "rewards/accuracies": 1.0, "rewards/chosen": -3.015625, "rewards/margins": 2.765625, "rewards/rejected": -5.78125, "step": 140 }, { "epoch": 0.5753968253968254, "grad_norm": 34.0474825303708, "learning_rate": 1.3750373276206429e-07, "logits/chosen": -1.59375, "logits/rejected": -1.59375, "logps/chosen": -0.404296875, "logps/rejected": -0.60546875, "loss": 1.5267, "rewards/accuracies": 1.0, "rewards/chosen": -4.03125, "rewards/margins": 2.015625, "rewards/rejected": -6.0625, "step": 145 }, { "epoch": 0.5952380952380952, "grad_norm": 40.81441971424475, "learning_rate": 1.2715286943962924e-07, "logits/chosen": -1.484375, "logits/rejected": -1.484375, "logps/chosen": -0.36328125, "logps/rejected": -0.5546875, "loss": 1.4843, "rewards/accuracies": 1.0, "rewards/chosen": -3.625, "rewards/margins": 1.9375, "rewards/rejected": -5.5625, "step": 150 }, { "epoch": 0.6150793650793651, "grad_norm": 34.79982376085805, "learning_rate": 1.1691233258095889e-07, "logits/chosen": -1.484375, "logits/rejected": -1.484375, "logps/chosen": -0.306640625, "logps/rejected": -0.51953125, "loss": 1.3835, "rewards/accuracies": 1.0, "rewards/chosen": -3.0625, "rewards/margins": 2.125, "rewards/rejected": -5.1875, "step": 155 }, { "epoch": 0.6349206349206349, "grad_norm": 42.42547672022662, "learning_rate": 1.0683157269432096e-07, "logits/chosen": -1.8046875, "logits/rejected": -1.8046875, "logps/chosen": -0.408203125, "logps/rejected": -0.5859375, "loss": 1.4352, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -4.0625, "rewards/margins": 1.796875, "rewards/rejected": -5.875, "step": 160 }, { "epoch": 0.6547619047619048, "grad_norm": 31.589553355930356, "learning_rate": 9.695926874127765e-08, "logits/chosen": -1.609375, "logits/rejected": -1.609375, "logps/chosen": -0.3125, "logps/rejected": -0.5546875, "loss": 1.2363, "rewards/accuracies": 1.0, "rewards/chosen": -3.109375, "rewards/margins": 2.4375, "rewards/rejected": -5.5625, "step": 165 }, { "epoch": 0.6746031746031746, "grad_norm": 49.83108895416612, "learning_rate": 8.734309307094381e-08, "logits/chosen": -1.515625, "logits/rejected": -1.515625, "logps/chosen": -0.255859375, "logps/rejected": -0.39453125, "loss": 1.7128, "rewards/accuracies": 1.0, "rewards/chosen": -2.5625, "rewards/margins": 1.3984375, "rewards/rejected": -3.953125, "step": 170 }, { "epoch": 0.6944444444444444, "grad_norm": 53.80975894133238, "learning_rate": 7.802948121507461e-08, "logits/chosen": -1.7265625, "logits/rejected": -1.7265625, "logps/chosen": -0.33203125, "logps/rejected": -0.52734375, "loss": 1.7579, "rewards/accuracies": 1.0, "rewards/chosen": -3.3125, "rewards/margins": 1.953125, "rewards/rejected": -5.28125, "step": 175 }, { "epoch": 0.7142857142857143, "grad_norm": 26.370363764214527, "learning_rate": 6.906340765561734e-08, "logits/chosen": -1.265625, "logits/rejected": -1.265625, "logps/chosen": -0.42578125, "logps/rejected": -0.69140625, "loss": 1.1591, "rewards/accuracies": 1.0, "rewards/chosen": -4.25, "rewards/margins": 2.65625, "rewards/rejected": -6.90625, "step": 180 }, { "epoch": 0.7341269841269841, "grad_norm": 36.18162633880284, "learning_rate": 6.048816864752422e-08, "logits/chosen": -1.609375, "logits/rejected": -1.609375, "logps/chosen": -0.421875, "logps/rejected": -0.7421875, "loss": 1.2978, "rewards/accuracies": 1.0, "rewards/chosen": -4.21875, "rewards/margins": 3.1875, "rewards/rejected": -7.40625, "step": 185 }, { "epoch": 0.753968253968254, "grad_norm": 42.731435648770926, "learning_rate": 5.2345173145552125e-08, "logits/chosen": -1.78125, "logits/rejected": -1.78125, "logps/chosen": -0.32421875, "logps/rejected": -0.55859375, "loss": 1.5185, "rewards/accuracies": 1.0, "rewards/chosen": -3.234375, "rewards/margins": 2.34375, "rewards/rejected": -5.5625, "step": 190 }, { "epoch": 0.7738095238095238, "grad_norm": 47.99973669165908, "learning_rate": 4.467374284464271e-08, "logits/chosen": -1.6171875, "logits/rejected": -1.6171875, "logps/chosen": -0.35546875, "logps/rejected": -0.62890625, "loss": 1.2557, "rewards/accuracies": 1.0, "rewards/chosen": -3.546875, "rewards/margins": 2.75, "rewards/rejected": -6.28125, "step": 195 }, { "epoch": 0.7936507936507936, "grad_norm": 31.799179942413893, "learning_rate": 3.751092229946681e-08, "logits/chosen": -1.7421875, "logits/rejected": -1.7421875, "logps/chosen": -0.37890625, "logps/rejected": -0.578125, "loss": 1.3694, "rewards/accuracies": 1.0, "rewards/chosen": -3.78125, "rewards/margins": 1.984375, "rewards/rejected": -5.78125, "step": 200 }, { "epoch": 0.8134920634920635, "grad_norm": 32.210040778412214, "learning_rate": 3.0891300040047544e-08, "logits/chosen": -1.390625, "logits/rejected": -1.390625, "logps/chosen": -0.423828125, "logps/rejected": -0.68359375, "loss": 1.3582, "rewards/accuracies": 1.0, "rewards/chosen": -4.25, "rewards/margins": 2.578125, "rewards/rejected": -6.8125, "step": 205 }, { "epoch": 0.8333333333333334, "grad_norm": 41.74711583345937, "learning_rate": 2.4846841547275915e-08, "logits/chosen": -1.1640625, "logits/rejected": -1.1640625, "logps/chosen": -0.259765625, "logps/rejected": -0.396484375, "loss": 1.3447, "rewards/accuracies": 1.0, "rewards/chosen": -2.59375, "rewards/margins": 1.3671875, "rewards/rejected": -3.96875, "step": 210 }, { "epoch": 0.8531746031746031, "grad_norm": 46.7470602788351, "learning_rate": 1.9406734894862847e-08, "logits/chosen": -1.1875, "logits/rejected": -1.1875, "logps/chosen": -0.412109375, "logps/rejected": -0.71484375, "loss": 1.2536, "rewards/accuracies": 1.0, "rewards/chosen": -4.125, "rewards/margins": 3.03125, "rewards/rejected": -7.15625, "step": 215 }, { "epoch": 0.873015873015873, "grad_norm": 60.766181922921575, "learning_rate": 1.459724980310767e-08, "logits/chosen": -1.4765625, "logits/rejected": -1.484375, "logps/chosen": -0.34765625, "logps/rejected": -0.625, "loss": 1.2171, "rewards/accuracies": 1.0, "rewards/chosen": -3.46875, "rewards/margins": 2.8125, "rewards/rejected": -6.28125, "step": 220 }, { "epoch": 0.8928571428571429, "grad_norm": 31.72970074806701, "learning_rate": 1.0441610785097471e-08, "logits/chosen": -1.3125, "logits/rejected": -1.3125, "logps/chosen": -0.34765625, "logps/rejected": -0.453125, "loss": 1.2928, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -3.46875, "rewards/margins": 1.0625, "rewards/rejected": -4.53125, "step": 225 }, { "epoch": 0.9126984126984127, "grad_norm": 35.90988337509379, "learning_rate": 6.959884997901705e-09, "logits/chosen": -1.203125, "logits/rejected": -1.203125, "logps/chosen": -0.2451171875, "logps/rejected": -0.484375, "loss": 1.2009, "rewards/accuracies": 1.0, "rewards/chosen": -2.453125, "rewards/margins": 2.390625, "rewards/rejected": -4.84375, "step": 230 }, { "epoch": 0.9325396825396826, "grad_norm": 39.759595151629654, "learning_rate": 4.168885340316719e-09, "logits/chosen": -1.4296875, "logits/rejected": -1.4296875, "logps/chosen": -0.4453125, "logps/rejected": -0.5390625, "loss": 1.4925, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -4.4375, "rewards/margins": 0.9296875, "rewards/rejected": -5.375, "step": 235 }, { "epoch": 0.9523809523809523, "grad_norm": 45.2835359838723, "learning_rate": 2.0820892650920686e-09, "logits/chosen": -1.40625, "logits/rejected": -1.40625, "logps/chosen": -0.4296875, "logps/rejected": -0.8671875, "loss": 1.0995, "rewards/accuracies": 1.0, "rewards/chosen": -4.3125, "rewards/margins": 4.375, "rewards/rejected": -8.6875, "step": 240 }, { "epoch": 0.9722222222222222, "grad_norm": 38.0802575489829, "learning_rate": 7.095736976853894e-10, "logits/chosen": -1.2578125, "logits/rejected": -1.2578125, "logps/chosen": -0.48046875, "logps/rejected": -0.890625, "loss": 1.3758, "rewards/accuracies": 1.0, "rewards/chosen": -4.78125, "rewards/margins": 4.125, "rewards/rejected": -8.875, "step": 245 }, { "epoch": 0.9920634920634921, "grad_norm": 54.22110017894445, "learning_rate": 5.796637581689245e-11, "logits/chosen": -1.484375, "logits/rejected": -1.484375, "logps/chosen": -0.44921875, "logps/rejected": -0.6015625, "loss": 1.466, "rewards/accuracies": 1.0, "rewards/chosen": -4.5, "rewards/margins": 1.515625, "rewards/rejected": -6.0, "step": 250 }, { "epoch": 1.0, "step": 252, "total_flos": 0.0, "train_loss": 1.8338732454511855, "train_runtime": 6099.2148, "train_samples_per_second": 0.33, "train_steps_per_second": 0.041 } ], "logging_steps": 5, "max_steps": 252, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }