|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.11726078799249531, |
|
"eval_steps": 500, |
|
"global_step": 3000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.003908692933083177, |
|
"grad_norm": 0.47716134786605835, |
|
"learning_rate": 4.98240400406663e-05, |
|
"loss": 1.3868, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.007817385866166354, |
|
"grad_norm": 0.28900113701820374, |
|
"learning_rate": 4.9628528974739973e-05, |
|
"loss": 1.3348, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01172607879924953, |
|
"grad_norm": 0.37652796506881714, |
|
"learning_rate": 4.943301790881364e-05, |
|
"loss": 1.3298, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.015634771732332707, |
|
"grad_norm": 0.24992702901363373, |
|
"learning_rate": 4.923750684288731e-05, |
|
"loss": 1.3309, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.019543464665415886, |
|
"grad_norm": 0.3237004280090332, |
|
"learning_rate": 4.904199577696098e-05, |
|
"loss": 1.3228, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.02345215759849906, |
|
"grad_norm": 0.259011447429657, |
|
"learning_rate": 4.884648471103465e-05, |
|
"loss": 1.3166, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.02736085053158224, |
|
"grad_norm": 0.34121841192245483, |
|
"learning_rate": 4.8650973645108314e-05, |
|
"loss": 1.3145, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.031269543464665414, |
|
"grad_norm": 0.41398781538009644, |
|
"learning_rate": 4.845546257918198e-05, |
|
"loss": 1.3081, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.03517823639774859, |
|
"grad_norm": 0.4289904832839966, |
|
"learning_rate": 4.825995151325565e-05, |
|
"loss": 1.3073, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.03908692933083177, |
|
"grad_norm": 0.21226634085178375, |
|
"learning_rate": 4.8064440447329324e-05, |
|
"loss": 1.3057, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.04299562226391495, |
|
"grad_norm": 0.2929254174232483, |
|
"learning_rate": 4.786892938140299e-05, |
|
"loss": 1.3034, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.04690431519699812, |
|
"grad_norm": 0.22025109827518463, |
|
"learning_rate": 4.767341831547666e-05, |
|
"loss": 1.3048, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.0508130081300813, |
|
"grad_norm": 0.36455097794532776, |
|
"learning_rate": 4.747790724955033e-05, |
|
"loss": 1.3028, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.05472170106316448, |
|
"grad_norm": 0.24527208507061005, |
|
"learning_rate": 4.728239618362399e-05, |
|
"loss": 1.2996, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.05863039399624766, |
|
"grad_norm": 0.33187615871429443, |
|
"learning_rate": 4.7086885117697664e-05, |
|
"loss": 1.3001, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.06253908692933083, |
|
"grad_norm": 0.502117931842804, |
|
"learning_rate": 4.689137405177133e-05, |
|
"loss": 1.2954, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.06644777986241401, |
|
"grad_norm": 0.289112389087677, |
|
"learning_rate": 4.6695862985845e-05, |
|
"loss": 1.2969, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.07035647279549719, |
|
"grad_norm": 0.24350133538246155, |
|
"learning_rate": 4.6500351919918674e-05, |
|
"loss": 1.2929, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.07426516572858036, |
|
"grad_norm": 0.2194579690694809, |
|
"learning_rate": 4.630484085399234e-05, |
|
"loss": 1.2913, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.07817385866166354, |
|
"grad_norm": 0.3578532636165619, |
|
"learning_rate": 4.6109329788066005e-05, |
|
"loss": 1.2946, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.08208255159474671, |
|
"grad_norm": 0.22988037765026093, |
|
"learning_rate": 4.591381872213967e-05, |
|
"loss": 1.2906, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.0859912445278299, |
|
"grad_norm": 0.44038262963294983, |
|
"learning_rate": 4.571830765621334e-05, |
|
"loss": 1.2941, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.08989993746091307, |
|
"grad_norm": 0.27333030104637146, |
|
"learning_rate": 4.5522796590287015e-05, |
|
"loss": 1.2903, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.09380863039399624, |
|
"grad_norm": 0.23196421563625336, |
|
"learning_rate": 4.532728552436068e-05, |
|
"loss": 1.2897, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.09771732332707943, |
|
"grad_norm": 0.25054216384887695, |
|
"learning_rate": 4.513177445843435e-05, |
|
"loss": 1.2891, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.1016260162601626, |
|
"grad_norm": 0.17129705846309662, |
|
"learning_rate": 4.493626339250802e-05, |
|
"loss": 1.2874, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.10553470919324578, |
|
"grad_norm": 0.5260087251663208, |
|
"learning_rate": 4.474075232658168e-05, |
|
"loss": 1.289, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.10944340212632896, |
|
"grad_norm": 0.2251499593257904, |
|
"learning_rate": 4.4545241260655355e-05, |
|
"loss": 1.2882, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.11335209505941213, |
|
"grad_norm": 0.24193763732910156, |
|
"learning_rate": 4.434973019472902e-05, |
|
"loss": 1.2886, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.11726078799249531, |
|
"grad_norm": 0.2361423224210739, |
|
"learning_rate": 4.415421912880269e-05, |
|
"loss": 1.2863, |
|
"step": 3000 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 25584, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 8389279088640000.0, |
|
"train_batch_size": 48, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|