{ "best_metric": null, "best_model_checkpoint": null, "epoch": 46.15384615384615, "eval_steps": 50, "global_step": 600, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.07692307692307693, "eval_loss": 3.9168343544006348, "eval_runtime": 5.6922, "eval_samples_per_second": 263.695, "eval_steps_per_second": 4.216, "step": 1 }, { "epoch": 0.7692307692307693, "grad_norm": 1.3671875, "learning_rate": 6.666666666666667e-05, "loss": 3.3633, "step": 10 }, { "epoch": 1.5384615384615383, "grad_norm": 1.09375, "learning_rate": 0.00013333333333333334, "loss": 3.1311, "step": 20 }, { "epoch": 2.3076923076923075, "grad_norm": 0.8203125, "learning_rate": 0.0002, "loss": 2.9153, "step": 30 }, { "epoch": 3.076923076923077, "grad_norm": 0.7265625, "learning_rate": 0.00019984815164333163, "loss": 2.7539, "step": 40 }, { "epoch": 3.8461538461538463, "grad_norm": 0.71875, "learning_rate": 0.00019939306773179497, "loss": 2.5978, "step": 50 }, { "epoch": 3.8461538461538463, "eval_loss": 2.814912796020508, "eval_runtime": 5.762, "eval_samples_per_second": 260.499, "eval_steps_per_second": 4.165, "step": 50 }, { "epoch": 4.615384615384615, "grad_norm": 0.71484375, "learning_rate": 0.00019863613034027224, "loss": 2.4698, "step": 60 }, { "epoch": 5.384615384615385, "grad_norm": 0.7734375, "learning_rate": 0.00019757963826274357, "loss": 2.3909, "step": 70 }, { "epoch": 6.153846153846154, "grad_norm": 0.765625, "learning_rate": 0.00019622680003092503, "loss": 2.2723, "step": 80 }, { "epoch": 6.923076923076923, "grad_norm": 0.796875, "learning_rate": 0.00019458172417006347, "loss": 2.1816, "step": 90 }, { "epoch": 7.6923076923076925, "grad_norm": 0.8203125, "learning_rate": 0.00019264940672148018, "loss": 2.0808, "step": 100 }, { "epoch": 7.6923076923076925, "eval_loss": 2.966378927230835, "eval_runtime": 5.6204, "eval_samples_per_second": 267.064, "eval_steps_per_second": 4.27, "step": 100 }, { "epoch": 8.461538461538462, "grad_norm": 0.83984375, "learning_rate": 0.00019043571606975777, "loss": 1.9902, "step": 110 }, { "epoch": 9.23076923076923, "grad_norm": 0.99609375, "learning_rate": 0.0001879473751206489, "loss": 1.9033, "step": 120 }, { "epoch": 10.0, "grad_norm": 0.83984375, "learning_rate": 0.00018519194088383273, "loss": 1.8173, "step": 130 }, { "epoch": 10.76923076923077, "grad_norm": 0.87890625, "learning_rate": 0.0001821777815225245, "loss": 1.7058, "step": 140 }, { "epoch": 11.538461538461538, "grad_norm": 0.9296875, "learning_rate": 0.00017891405093963938, "loss": 1.6294, "step": 150 }, { "epoch": 11.538461538461538, "eval_loss": 3.233675003051758, "eval_runtime": 5.5692, "eval_samples_per_second": 269.517, "eval_steps_per_second": 4.309, "step": 150 }, { "epoch": 12.307692307692308, "grad_norm": 0.9921875, "learning_rate": 0.00017541066097768963, "loss": 1.5563, "step": 160 }, { "epoch": 13.076923076923077, "grad_norm": 1.0546875, "learning_rate": 0.00017167825131684513, "loss": 1.4865, "step": 170 }, { "epoch": 13.846153846153847, "grad_norm": 0.94921875, "learning_rate": 0.00016772815716257412, "loss": 1.3938, "step": 180 }, { "epoch": 14.615384615384615, "grad_norm": 0.90625, "learning_rate": 0.00016357237482099684, "loss": 1.3247, "step": 190 }, { "epoch": 15.384615384615385, "grad_norm": 0.99609375, "learning_rate": 0.00015922352526649803, "loss": 1.2699, "step": 200 }, { "epoch": 15.384615384615385, "eval_loss": 3.521737813949585, "eval_runtime": 5.6468, "eval_samples_per_second": 265.815, "eval_steps_per_second": 4.25, "step": 200 }, { "epoch": 16.153846153846153, "grad_norm": 0.95703125, "learning_rate": 0.00015469481581224272, "loss": 1.2117, "step": 210 }, { "epoch": 16.923076923076923, "grad_norm": 1.0390625, "learning_rate": 0.00015000000000000001, "loss": 1.1498, "step": 220 }, { "epoch": 17.692307692307693, "grad_norm": 1.0078125, "learning_rate": 0.00014515333583108896, "loss": 1.0864, "step": 230 }, { "epoch": 18.46153846153846, "grad_norm": 0.81640625, "learning_rate": 0.00014016954246529696, "loss": 1.0441, "step": 240 }, { "epoch": 19.23076923076923, "grad_norm": 0.98828125, "learning_rate": 0.00013506375551927547, "loss": 1.0092, "step": 250 }, { "epoch": 19.23076923076923, "eval_loss": 3.726165533065796, "eval_runtime": 5.8207, "eval_samples_per_second": 257.872, "eval_steps_per_second": 4.123, "step": 250 }, { "epoch": 20.0, "grad_norm": 0.8359375, "learning_rate": 0.00012985148110016947, "loss": 0.9671, "step": 260 }, { "epoch": 20.76923076923077, "grad_norm": 0.8203125, "learning_rate": 0.00012454854871407994, "loss": 0.9182, "step": 270 }, { "epoch": 21.53846153846154, "grad_norm": 0.796875, "learning_rate": 0.00011917106319237386, "loss": 0.8857, "step": 280 }, { "epoch": 22.307692307692307, "grad_norm": 0.79296875, "learning_rate": 0.00011373535578184082, "loss": 0.8623, "step": 290 }, { "epoch": 23.076923076923077, "grad_norm": 0.7421875, "learning_rate": 0.00010825793454723325, "loss": 0.8392, "step": 300 }, { "epoch": 23.076923076923077, "eval_loss": 3.868284225463867, "eval_runtime": 5.7191, "eval_samples_per_second": 262.455, "eval_steps_per_second": 4.196, "step": 300 }, { "epoch": 23.846153846153847, "grad_norm": 0.75, "learning_rate": 0.00010275543423681621, "loss": 0.8103, "step": 310 }, { "epoch": 24.615384615384617, "grad_norm": 0.75390625, "learning_rate": 9.724456576318381e-05, "loss": 0.7828, "step": 320 }, { "epoch": 25.384615384615383, "grad_norm": 0.7265625, "learning_rate": 9.174206545276677e-05, "loss": 0.7749, "step": 330 }, { "epoch": 26.153846153846153, "grad_norm": 0.74609375, "learning_rate": 8.626464421815919e-05, "loss": 0.7574, "step": 340 }, { "epoch": 26.923076923076923, "grad_norm": 0.6875, "learning_rate": 8.082893680762619e-05, "loss": 0.7428, "step": 350 }, { "epoch": 26.923076923076923, "eval_loss": 3.9434773921966553, "eval_runtime": 5.6325, "eval_samples_per_second": 266.49, "eval_steps_per_second": 4.261, "step": 350 }, { "epoch": 27.692307692307693, "grad_norm": 0.6953125, "learning_rate": 7.54514512859201e-05, "loss": 0.7298, "step": 360 }, { "epoch": 28.46153846153846, "grad_norm": 0.66015625, "learning_rate": 7.014851889983057e-05, "loss": 0.7167, "step": 370 }, { "epoch": 29.23076923076923, "grad_norm": 0.65625, "learning_rate": 6.493624448072457e-05, "loss": 0.7147, "step": 380 }, { "epoch": 30.0, "grad_norm": 0.69140625, "learning_rate": 5.983045753470308e-05, "loss": 0.7019, "step": 390 }, { "epoch": 30.76923076923077, "grad_norm": 0.66796875, "learning_rate": 5.484666416891109e-05, "loss": 0.6952, "step": 400 }, { "epoch": 30.76923076923077, "eval_loss": 3.985978841781616, "eval_runtime": 5.6438, "eval_samples_per_second": 265.955, "eval_steps_per_second": 4.252, "step": 400 }, { "epoch": 31.53846153846154, "grad_norm": 0.6484375, "learning_rate": 5.000000000000002e-05, "loss": 0.6901, "step": 410 }, { "epoch": 32.30769230769231, "grad_norm": 0.640625, "learning_rate": 4.530518418775733e-05, "loss": 0.685, "step": 420 }, { "epoch": 33.07692307692308, "grad_norm": 0.62890625, "learning_rate": 4.077647473350201e-05, "loss": 0.6851, "step": 430 }, { "epoch": 33.84615384615385, "grad_norm": 0.62109375, "learning_rate": 3.642762517900322e-05, "loss": 0.6782, "step": 440 }, { "epoch": 34.61538461538461, "grad_norm": 0.6171875, "learning_rate": 3.227184283742591e-05, "loss": 0.6762, "step": 450 }, { "epoch": 34.61538461538461, "eval_loss": 3.998966932296753, "eval_runtime": 5.7156, "eval_samples_per_second": 262.613, "eval_steps_per_second": 4.199, "step": 450 }, { "epoch": 35.38461538461539, "grad_norm": 0.62109375, "learning_rate": 2.8321748683154893e-05, "loss": 0.6742, "step": 460 }, { "epoch": 36.15384615384615, "grad_norm": 0.6171875, "learning_rate": 2.4589339022310386e-05, "loss": 0.674, "step": 470 }, { "epoch": 36.92307692307692, "grad_norm": 0.62109375, "learning_rate": 2.1085949060360654e-05, "loss": 0.6728, "step": 480 }, { "epoch": 37.69230769230769, "grad_norm": 0.6171875, "learning_rate": 1.7822218477475494e-05, "loss": 0.6681, "step": 490 }, { "epoch": 38.46153846153846, "grad_norm": 0.62109375, "learning_rate": 1.4808059116167305e-05, "loss": 0.6739, "step": 500 }, { "epoch": 38.46153846153846, "eval_loss": 4.016704559326172, "eval_runtime": 5.9697, "eval_samples_per_second": 251.437, "eval_steps_per_second": 4.02, "step": 500 }, { "epoch": 39.23076923076923, "grad_norm": 0.61328125, "learning_rate": 1.2052624879351104e-05, "loss": 0.6685, "step": 510 }, { "epoch": 40.0, "grad_norm": 0.625, "learning_rate": 9.564283930242257e-06, "loss": 0.6697, "step": 520 }, { "epoch": 40.76923076923077, "grad_norm": 0.60546875, "learning_rate": 7.350593278519824e-06, "loss": 0.6691, "step": 530 }, { "epoch": 41.53846153846154, "grad_norm": 0.609375, "learning_rate": 5.418275829936537e-06, "loss": 0.6709, "step": 540 }, { "epoch": 42.30769230769231, "grad_norm": 0.62109375, "learning_rate": 3.7731999690749585e-06, "loss": 0.6691, "step": 550 }, { "epoch": 42.30769230769231, "eval_loss": 4.020811080932617, "eval_runtime": 5.6558, "eval_samples_per_second": 265.392, "eval_steps_per_second": 4.243, "step": 550 }, { "epoch": 43.07692307692308, "grad_norm": 0.61328125, "learning_rate": 2.420361737256438e-06, "loss": 0.6671, "step": 560 }, { "epoch": 43.84615384615385, "grad_norm": 0.62890625, "learning_rate": 1.3638696597277679e-06, "loss": 0.6683, "step": 570 }, { "epoch": 44.61538461538461, "grad_norm": 0.59375, "learning_rate": 6.069322682050516e-07, "loss": 0.6714, "step": 580 }, { "epoch": 45.38461538461539, "grad_norm": 0.6015625, "learning_rate": 1.518483566683826e-07, "loss": 0.6695, "step": 590 }, { "epoch": 46.15384615384615, "grad_norm": 0.62109375, "learning_rate": 0.0, "loss": 0.6667, "step": 600 }, { "epoch": 46.15384615384615, "eval_loss": 4.010259628295898, "eval_runtime": 5.6787, "eval_samples_per_second": 264.32, "eval_steps_per_second": 4.226, "step": 600 } ], "logging_steps": 10, "max_steps": 600, "num_input_tokens_seen": 0, "num_train_epochs": 47, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.05042678235136e+16, "train_batch_size": 32, "trial_name": null, "trial_params": null }