SelfCite-8B / trainer_state.json
voidism's picture
first commit
e452aff
raw
history blame
23.6 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 252,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01984126984126984,
"grad_norm": 18.047539550790006,
"learning_rate": 5.7692307692307695e-08,
"logits/chosen": -1.703125,
"logits/rejected": -1.6953125,
"logps/chosen": -0.376953125,
"logps/rejected": -0.427734375,
"loss": 2.4555,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.765625,
"rewards/margins": 0.50390625,
"rewards/rejected": -4.28125,
"step": 5
},
{
"epoch": 0.03968253968253968,
"grad_norm": 17.185434459136403,
"learning_rate": 1.1538461538461539e-07,
"logits/chosen": -1.2578125,
"logits/rejected": -1.2578125,
"logps/chosen": -0.32421875,
"logps/rejected": -0.390625,
"loss": 2.4857,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.234375,
"rewards/margins": 0.6640625,
"rewards/rejected": -3.90625,
"step": 10
},
{
"epoch": 0.05952380952380952,
"grad_norm": 17.039133125856026,
"learning_rate": 1.7307692307692305e-07,
"logits/chosen": -1.5078125,
"logits/rejected": -1.5078125,
"logps/chosen": -0.279296875,
"logps/rejected": -0.34765625,
"loss": 2.402,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.78125,
"rewards/margins": 0.6796875,
"rewards/rejected": -3.46875,
"step": 15
},
{
"epoch": 0.07936507936507936,
"grad_norm": 17.10226682999204,
"learning_rate": 2.3076923076923078e-07,
"logits/chosen": -1.4765625,
"logits/rejected": -1.4765625,
"logps/chosen": -0.310546875,
"logps/rejected": -0.39453125,
"loss": 2.4703,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.09375,
"rewards/margins": 0.84375,
"rewards/rejected": -3.9375,
"step": 20
},
{
"epoch": 0.0992063492063492,
"grad_norm": 14.47393175785549,
"learning_rate": 2.8846153846153846e-07,
"logits/chosen": -1.78125,
"logits/rejected": -1.78125,
"logps/chosen": -0.34375,
"logps/rejected": -0.40234375,
"loss": 2.3167,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.421875,
"rewards/margins": 0.5859375,
"rewards/rejected": -4.0,
"step": 25
},
{
"epoch": 0.11904761904761904,
"grad_norm": 16.9612236979206,
"learning_rate": 2.997681792980754e-07,
"logits/chosen": -1.3359375,
"logits/rejected": -1.3359375,
"logps/chosen": -0.3046875,
"logps/rejected": -0.357421875,
"loss": 2.3476,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.03125,
"rewards/margins": 0.54296875,
"rewards/rejected": -3.578125,
"step": 30
},
{
"epoch": 0.1388888888888889,
"grad_norm": 16.48721079930822,
"learning_rate": 2.9882763562415516e-07,
"logits/chosen": -1.5078125,
"logits/rejected": -1.5078125,
"logps/chosen": -0.408203125,
"logps/rejected": -0.482421875,
"loss": 2.391,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.0625,
"rewards/margins": 0.75390625,
"rewards/rejected": -4.8125,
"step": 35
},
{
"epoch": 0.15873015873015872,
"grad_norm": 21.37387348387286,
"learning_rate": 2.9716841847369104e-07,
"logits/chosen": -1.5,
"logits/rejected": -1.5,
"logps/chosen": -0.2890625,
"logps/rejected": -0.365234375,
"loss": 2.3684,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.890625,
"rewards/margins": 0.7578125,
"rewards/rejected": -3.640625,
"step": 40
},
{
"epoch": 0.17857142857142858,
"grad_norm": 17.871422332796683,
"learning_rate": 2.94798540037113e-07,
"logits/chosen": -1.546875,
"logits/rejected": -1.546875,
"logps/chosen": -0.287109375,
"logps/rejected": -0.3515625,
"loss": 2.3699,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.859375,
"rewards/margins": 0.640625,
"rewards/rejected": -3.5,
"step": 45
},
{
"epoch": 0.1984126984126984,
"grad_norm": 29.190246701669263,
"learning_rate": 2.917294442157359e-07,
"logits/chosen": -1.3828125,
"logits/rejected": -1.3828125,
"logps/chosen": -0.265625,
"logps/rejected": -0.34375,
"loss": 2.3773,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.65625,
"rewards/margins": 0.78125,
"rewards/rejected": -3.4375,
"step": 50
},
{
"epoch": 0.21825396825396826,
"grad_norm": 20.420712184187625,
"learning_rate": 2.8797595136032675e-07,
"logits/chosen": -1.4765625,
"logits/rejected": -1.4765625,
"logps/chosen": -0.31640625,
"logps/rejected": -0.39453125,
"loss": 2.3586,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.171875,
"rewards/margins": 0.7734375,
"rewards/rejected": -3.953125,
"step": 55
},
{
"epoch": 0.23809523809523808,
"grad_norm": 17.704268736560135,
"learning_rate": 2.8355618670514256e-07,
"logits/chosen": -1.2890625,
"logits/rejected": -1.2890625,
"logps/chosen": -0.27734375,
"logps/rejected": -0.328125,
"loss": 2.3299,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": -2.765625,
"rewards/margins": 0.515625,
"rewards/rejected": -3.28125,
"step": 60
},
{
"epoch": 0.25793650793650796,
"grad_norm": 15.278992243473061,
"learning_rate": 2.784914928430218e-07,
"logits/chosen": -1.40625,
"logits/rejected": -1.40625,
"logps/chosen": -0.294921875,
"logps/rejected": -0.37109375,
"loss": 2.3791,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": -2.953125,
"rewards/margins": 0.7578125,
"rewards/rejected": -3.703125,
"step": 65
},
{
"epoch": 0.2777777777777778,
"grad_norm": 21.138997719744832,
"learning_rate": 2.728063266641801e-07,
"logits/chosen": -1.6484375,
"logits/rejected": -1.6484375,
"logps/chosen": -0.26953125,
"logps/rejected": -0.361328125,
"loss": 2.3719,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.703125,
"rewards/margins": 0.9140625,
"rewards/rejected": -3.625,
"step": 70
},
{
"epoch": 0.2976190476190476,
"grad_norm": 26.097345440204215,
"learning_rate": 2.665281412563814e-07,
"logits/chosen": -1.875,
"logits/rejected": -1.875,
"logps/chosen": -0.26171875,
"logps/rejected": -0.37109375,
"loss": 2.318,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.625,
"rewards/margins": 1.0859375,
"rewards/rejected": -3.703125,
"step": 75
},
{
"epoch": 0.31746031746031744,
"grad_norm": 21.061149357230455,
"learning_rate": 2.596872533367763e-07,
"logits/chosen": -1.5,
"logits/rejected": -1.5,
"logps/chosen": -0.28125,
"logps/rejected": -0.373046875,
"loss": 2.2812,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.828125,
"rewards/margins": 0.89453125,
"rewards/rejected": -3.71875,
"step": 80
},
{
"epoch": 0.3373015873015873,
"grad_norm": 34.0182184293618,
"learning_rate": 2.5231669685556633e-07,
"logits/chosen": -1.1875,
"logits/rejected": -1.1875,
"logps/chosen": -0.27734375,
"logps/rejected": -0.39453125,
"loss": 2.166,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.78125,
"rewards/margins": 1.15625,
"rewards/rejected": -3.9375,
"step": 85
},
{
"epoch": 0.35714285714285715,
"grad_norm": 20.821123274301424,
"learning_rate": 2.444520634784271e-07,
"logits/chosen": -1.3359375,
"logits/rejected": -1.3359375,
"logps/chosen": -0.267578125,
"logps/rejected": -0.361328125,
"loss": 2.1671,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.671875,
"rewards/margins": 0.9453125,
"rewards/rejected": -3.625,
"step": 90
},
{
"epoch": 0.376984126984127,
"grad_norm": 22.90486064935448,
"learning_rate": 2.361313307179837e-07,
"logits/chosen": -1.7265625,
"logits/rejected": -1.7265625,
"logps/chosen": -0.279296875,
"logps/rejected": -0.353515625,
"loss": 2.0894,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.78125,
"rewards/margins": 0.74609375,
"rewards/rejected": -3.53125,
"step": 95
},
{
"epoch": 0.3968253968253968,
"grad_norm": 27.68913080972186,
"learning_rate": 2.2739467854427513e-07,
"logits/chosen": -1.65625,
"logits/rejected": -1.65625,
"logps/chosen": -0.353515625,
"logps/rejected": -0.46875,
"loss": 2.1033,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.546875,
"rewards/margins": 1.1484375,
"rewards/rejected": -4.6875,
"step": 100
},
{
"epoch": 0.4166666666666667,
"grad_norm": 29.979641972346496,
"learning_rate": 2.1828429535977582e-07,
"logits/chosen": -1.6796875,
"logits/rejected": -1.6796875,
"logps/chosen": -0.3203125,
"logps/rejected": -0.4296875,
"loss": 2.0674,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": -3.1875,
"rewards/margins": 1.09375,
"rewards/rejected": -4.28125,
"step": 105
},
{
"epoch": 0.4365079365079365,
"grad_norm": 44.14626475514875,
"learning_rate": 2.0884417427590215e-07,
"logits/chosen": -1.65625,
"logits/rejected": -1.65625,
"logps/chosen": -0.34375,
"logps/rejected": -0.46875,
"loss": 2.0088,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.4375,
"rewards/margins": 1.2578125,
"rewards/rejected": -4.6875,
"step": 110
},
{
"epoch": 0.45634920634920634,
"grad_norm": 33.80229631943301,
"learning_rate": 1.9911990067476334e-07,
"logits/chosen": -1.15625,
"logits/rejected": -1.15625,
"logps/chosen": -0.23828125,
"logps/rejected": -0.34375,
"loss": 1.9458,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.390625,
"rewards/margins": 1.046875,
"rewards/rejected": -3.4375,
"step": 115
},
{
"epoch": 0.47619047619047616,
"grad_norm": 39.86801120602329,
"learning_rate": 1.8915843208199963e-07,
"logits/chosen": -1.625,
"logits/rejected": -1.625,
"logps/chosen": -0.4296875,
"logps/rejected": -0.546875,
"loss": 1.9224,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.3125,
"rewards/margins": 1.171875,
"rewards/rejected": -5.46875,
"step": 120
},
{
"epoch": 0.49603174603174605,
"grad_norm": 33.949958184137344,
"learning_rate": 1.7900787141367918e-07,
"logits/chosen": -1.453125,
"logits/rejected": -1.453125,
"logps/chosen": -0.357421875,
"logps/rejected": -0.4765625,
"loss": 2.0021,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.578125,
"rewards/margins": 1.1875,
"rewards/rejected": -4.78125,
"step": 125
},
{
"epoch": 0.5158730158730159,
"grad_norm": 37.35189108229666,
"learning_rate": 1.687172346922213e-07,
"logits/chosen": -1.546875,
"logits/rejected": -1.546875,
"logps/chosen": -0.419921875,
"logps/rejected": -0.546875,
"loss": 1.7533,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": -4.1875,
"rewards/margins": 1.2578125,
"rewards/rejected": -5.46875,
"step": 130
},
{
"epoch": 0.5357142857142857,
"grad_norm": 33.938784524888526,
"learning_rate": 1.5833621435302245e-07,
"logits/chosen": -1.734375,
"logits/rejected": -1.734375,
"logps/chosen": -0.421875,
"logps/rejected": -0.63671875,
"loss": 1.716,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.21875,
"rewards/margins": 2.15625,
"rewards/rejected": -6.34375,
"step": 135
},
{
"epoch": 0.5555555555555556,
"grad_norm": 35.12217870275606,
"learning_rate": 1.4791493928475273e-07,
"logits/chosen": -1.5234375,
"logits/rejected": -1.5234375,
"logps/chosen": -0.30078125,
"logps/rejected": -0.578125,
"loss": 1.7479,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.015625,
"rewards/margins": 2.765625,
"rewards/rejected": -5.78125,
"step": 140
},
{
"epoch": 0.5753968253968254,
"grad_norm": 34.0474825303708,
"learning_rate": 1.3750373276206429e-07,
"logits/chosen": -1.59375,
"logits/rejected": -1.59375,
"logps/chosen": -0.404296875,
"logps/rejected": -0.60546875,
"loss": 1.5267,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.03125,
"rewards/margins": 2.015625,
"rewards/rejected": -6.0625,
"step": 145
},
{
"epoch": 0.5952380952380952,
"grad_norm": 40.81441971424475,
"learning_rate": 1.2715286943962924e-07,
"logits/chosen": -1.484375,
"logits/rejected": -1.484375,
"logps/chosen": -0.36328125,
"logps/rejected": -0.5546875,
"loss": 1.4843,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.625,
"rewards/margins": 1.9375,
"rewards/rejected": -5.5625,
"step": 150
},
{
"epoch": 0.6150793650793651,
"grad_norm": 34.79982376085805,
"learning_rate": 1.1691233258095889e-07,
"logits/chosen": -1.484375,
"logits/rejected": -1.484375,
"logps/chosen": -0.306640625,
"logps/rejected": -0.51953125,
"loss": 1.3835,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.0625,
"rewards/margins": 2.125,
"rewards/rejected": -5.1875,
"step": 155
},
{
"epoch": 0.6349206349206349,
"grad_norm": 42.42547672022662,
"learning_rate": 1.0683157269432096e-07,
"logits/chosen": -1.8046875,
"logits/rejected": -1.8046875,
"logps/chosen": -0.408203125,
"logps/rejected": -0.5859375,
"loss": 1.4352,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": -4.0625,
"rewards/margins": 1.796875,
"rewards/rejected": -5.875,
"step": 160
},
{
"epoch": 0.6547619047619048,
"grad_norm": 31.589553355930356,
"learning_rate": 9.695926874127765e-08,
"logits/chosen": -1.609375,
"logits/rejected": -1.609375,
"logps/chosen": -0.3125,
"logps/rejected": -0.5546875,
"loss": 1.2363,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.109375,
"rewards/margins": 2.4375,
"rewards/rejected": -5.5625,
"step": 165
},
{
"epoch": 0.6746031746031746,
"grad_norm": 49.83108895416612,
"learning_rate": 8.734309307094381e-08,
"logits/chosen": -1.515625,
"logits/rejected": -1.515625,
"logps/chosen": -0.255859375,
"logps/rejected": -0.39453125,
"loss": 1.7128,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.5625,
"rewards/margins": 1.3984375,
"rewards/rejected": -3.953125,
"step": 170
},
{
"epoch": 0.6944444444444444,
"grad_norm": 53.80975894133238,
"learning_rate": 7.802948121507461e-08,
"logits/chosen": -1.7265625,
"logits/rejected": -1.7265625,
"logps/chosen": -0.33203125,
"logps/rejected": -0.52734375,
"loss": 1.7579,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.3125,
"rewards/margins": 1.953125,
"rewards/rejected": -5.28125,
"step": 175
},
{
"epoch": 0.7142857142857143,
"grad_norm": 26.370363764214527,
"learning_rate": 6.906340765561734e-08,
"logits/chosen": -1.265625,
"logits/rejected": -1.265625,
"logps/chosen": -0.42578125,
"logps/rejected": -0.69140625,
"loss": 1.1591,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.25,
"rewards/margins": 2.65625,
"rewards/rejected": -6.90625,
"step": 180
},
{
"epoch": 0.7341269841269841,
"grad_norm": 36.18162633880284,
"learning_rate": 6.048816864752422e-08,
"logits/chosen": -1.609375,
"logits/rejected": -1.609375,
"logps/chosen": -0.421875,
"logps/rejected": -0.7421875,
"loss": 1.2978,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.21875,
"rewards/margins": 3.1875,
"rewards/rejected": -7.40625,
"step": 185
},
{
"epoch": 0.753968253968254,
"grad_norm": 42.731435648770926,
"learning_rate": 5.2345173145552125e-08,
"logits/chosen": -1.78125,
"logits/rejected": -1.78125,
"logps/chosen": -0.32421875,
"logps/rejected": -0.55859375,
"loss": 1.5185,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.234375,
"rewards/margins": 2.34375,
"rewards/rejected": -5.5625,
"step": 190
},
{
"epoch": 0.7738095238095238,
"grad_norm": 47.99973669165908,
"learning_rate": 4.467374284464271e-08,
"logits/chosen": -1.6171875,
"logits/rejected": -1.6171875,
"logps/chosen": -0.35546875,
"logps/rejected": -0.62890625,
"loss": 1.2557,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.546875,
"rewards/margins": 2.75,
"rewards/rejected": -6.28125,
"step": 195
},
{
"epoch": 0.7936507936507936,
"grad_norm": 31.799179942413893,
"learning_rate": 3.751092229946681e-08,
"logits/chosen": -1.7421875,
"logits/rejected": -1.7421875,
"logps/chosen": -0.37890625,
"logps/rejected": -0.578125,
"loss": 1.3694,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.78125,
"rewards/margins": 1.984375,
"rewards/rejected": -5.78125,
"step": 200
},
{
"epoch": 0.8134920634920635,
"grad_norm": 32.210040778412214,
"learning_rate": 3.0891300040047544e-08,
"logits/chosen": -1.390625,
"logits/rejected": -1.390625,
"logps/chosen": -0.423828125,
"logps/rejected": -0.68359375,
"loss": 1.3582,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.25,
"rewards/margins": 2.578125,
"rewards/rejected": -6.8125,
"step": 205
},
{
"epoch": 0.8333333333333334,
"grad_norm": 41.74711583345937,
"learning_rate": 2.4846841547275915e-08,
"logits/chosen": -1.1640625,
"logits/rejected": -1.1640625,
"logps/chosen": -0.259765625,
"logps/rejected": -0.396484375,
"loss": 1.3447,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.59375,
"rewards/margins": 1.3671875,
"rewards/rejected": -3.96875,
"step": 210
},
{
"epoch": 0.8531746031746031,
"grad_norm": 46.7470602788351,
"learning_rate": 1.9406734894862847e-08,
"logits/chosen": -1.1875,
"logits/rejected": -1.1875,
"logps/chosen": -0.412109375,
"logps/rejected": -0.71484375,
"loss": 1.2536,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.125,
"rewards/margins": 3.03125,
"rewards/rejected": -7.15625,
"step": 215
},
{
"epoch": 0.873015873015873,
"grad_norm": 60.766181922921575,
"learning_rate": 1.459724980310767e-08,
"logits/chosen": -1.4765625,
"logits/rejected": -1.484375,
"logps/chosen": -0.34765625,
"logps/rejected": -0.625,
"loss": 1.2171,
"rewards/accuracies": 1.0,
"rewards/chosen": -3.46875,
"rewards/margins": 2.8125,
"rewards/rejected": -6.28125,
"step": 220
},
{
"epoch": 0.8928571428571429,
"grad_norm": 31.72970074806701,
"learning_rate": 1.0441610785097471e-08,
"logits/chosen": -1.3125,
"logits/rejected": -1.3125,
"logps/chosen": -0.34765625,
"logps/rejected": -0.453125,
"loss": 1.2928,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": -3.46875,
"rewards/margins": 1.0625,
"rewards/rejected": -4.53125,
"step": 225
},
{
"epoch": 0.9126984126984127,
"grad_norm": 35.90988337509379,
"learning_rate": 6.959884997901705e-09,
"logits/chosen": -1.203125,
"logits/rejected": -1.203125,
"logps/chosen": -0.2451171875,
"logps/rejected": -0.484375,
"loss": 1.2009,
"rewards/accuracies": 1.0,
"rewards/chosen": -2.453125,
"rewards/margins": 2.390625,
"rewards/rejected": -4.84375,
"step": 230
},
{
"epoch": 0.9325396825396826,
"grad_norm": 39.759595151629654,
"learning_rate": 4.168885340316719e-09,
"logits/chosen": -1.4296875,
"logits/rejected": -1.4296875,
"logps/chosen": -0.4453125,
"logps/rejected": -0.5390625,
"loss": 1.4925,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": -4.4375,
"rewards/margins": 0.9296875,
"rewards/rejected": -5.375,
"step": 235
},
{
"epoch": 0.9523809523809523,
"grad_norm": 45.2835359838723,
"learning_rate": 2.0820892650920686e-09,
"logits/chosen": -1.40625,
"logits/rejected": -1.40625,
"logps/chosen": -0.4296875,
"logps/rejected": -0.8671875,
"loss": 1.0995,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.3125,
"rewards/margins": 4.375,
"rewards/rejected": -8.6875,
"step": 240
},
{
"epoch": 0.9722222222222222,
"grad_norm": 38.0802575489829,
"learning_rate": 7.095736976853894e-10,
"logits/chosen": -1.2578125,
"logits/rejected": -1.2578125,
"logps/chosen": -0.48046875,
"logps/rejected": -0.890625,
"loss": 1.3758,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.78125,
"rewards/margins": 4.125,
"rewards/rejected": -8.875,
"step": 245
},
{
"epoch": 0.9920634920634921,
"grad_norm": 54.22110017894445,
"learning_rate": 5.796637581689245e-11,
"logits/chosen": -1.484375,
"logits/rejected": -1.484375,
"logps/chosen": -0.44921875,
"logps/rejected": -0.6015625,
"loss": 1.466,
"rewards/accuracies": 1.0,
"rewards/chosen": -4.5,
"rewards/margins": 1.515625,
"rewards/rejected": -6.0,
"step": 250
},
{
"epoch": 1.0,
"step": 252,
"total_flos": 0.0,
"train_loss": 1.8338732454511855,
"train_runtime": 6099.2148,
"train_samples_per_second": 0.33,
"train_steps_per_second": 0.041
}
],
"logging_steps": 5,
"max_steps": 252,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}