|
{ |
|
"best_metric": 3.989581823348999, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-600", |
|
"epoch": 1.1320754716981132, |
|
"eval_steps": 50, |
|
"global_step": 600, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0018867924528301887, |
|
"eval_loss": 4.477752685546875, |
|
"eval_runtime": 40.7428, |
|
"eval_samples_per_second": 21.918, |
|
"eval_steps_per_second": 5.498, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.018867924528301886, |
|
"grad_norm": 15.21529483795166, |
|
"learning_rate": 0.0002, |
|
"loss": 4.2379, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03773584905660377, |
|
"grad_norm": 54.98305130004883, |
|
"learning_rate": 0.0001998582695676762, |
|
"loss": 27.8114, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.05660377358490566, |
|
"grad_norm": 16.903823852539062, |
|
"learning_rate": 0.00019943348002101371, |
|
"loss": 25.2762, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.07547169811320754, |
|
"grad_norm": 172.9484405517578, |
|
"learning_rate": 0.00019872683547213446, |
|
"loss": 23.4547, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.09433962264150944, |
|
"grad_norm": 164.54025268554688, |
|
"learning_rate": 0.00019774033898178667, |
|
"loss": 8.2119, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.09433962264150944, |
|
"eval_loss": 7.419450759887695, |
|
"eval_runtime": 41.389, |
|
"eval_samples_per_second": 21.576, |
|
"eval_steps_per_second": 5.412, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.11320754716981132, |
|
"grad_norm": 44.91653060913086, |
|
"learning_rate": 0.0001964767868814516, |
|
"loss": 30.0243, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.1320754716981132, |
|
"grad_norm": 9.489733695983887, |
|
"learning_rate": 0.00019493976084683813, |
|
"loss": 25.1776, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.1509433962264151, |
|
"grad_norm": 47.95170211791992, |
|
"learning_rate": 0.00019313361774523385, |
|
"loss": 24.4991, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.16981132075471697, |
|
"grad_norm": 166.98104858398438, |
|
"learning_rate": 0.00019106347728549135, |
|
"loss": 16.642, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.18867924528301888, |
|
"grad_norm": 155.18824768066406, |
|
"learning_rate": 0.00018873520750565718, |
|
"loss": 6.9458, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.18867924528301888, |
|
"eval_loss": 14.900463104248047, |
|
"eval_runtime": 41.4137, |
|
"eval_samples_per_second": 21.563, |
|
"eval_steps_per_second": 5.409, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.20754716981132076, |
|
"grad_norm": 31.870784759521484, |
|
"learning_rate": 0.0001861554081393806, |
|
"loss": 55.9772, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.22641509433962265, |
|
"grad_norm": 19.77384376525879, |
|
"learning_rate": 0.0001833313919082515, |
|
"loss": 25.2894, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.24528301886792453, |
|
"grad_norm": 71.0095443725586, |
|
"learning_rate": 0.00018027116379309638, |
|
"loss": 24.5163, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.2641509433962264, |
|
"grad_norm": 328.7481384277344, |
|
"learning_rate": 0.00017698339834299061, |
|
"loss": 15.8981, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.2830188679245283, |
|
"grad_norm": 242.60081481933594, |
|
"learning_rate": 0.00017347741508630672, |
|
"loss": 6.2023, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2830188679245283, |
|
"eval_loss": 11.450271606445312, |
|
"eval_runtime": 41.4022, |
|
"eval_samples_per_second": 21.569, |
|
"eval_steps_per_second": 5.41, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.3018867924528302, |
|
"grad_norm": 31.868478775024414, |
|
"learning_rate": 0.0001697631521134985, |
|
"loss": 38.3331, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.32075471698113206, |
|
"grad_norm": 25.038724899291992, |
|
"learning_rate": 0.00016585113790650388, |
|
"loss": 24.7657, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.33962264150943394, |
|
"grad_norm": 53.333683013916016, |
|
"learning_rate": 0.0001617524614946192, |
|
"loss": 23.8673, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.3584905660377358, |
|
"grad_norm": 183.1505889892578, |
|
"learning_rate": 0.0001574787410214407, |
|
"loss": 12.851, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.37735849056603776, |
|
"grad_norm": 199.62698364257812, |
|
"learning_rate": 0.00015304209081197425, |
|
"loss": 3.7041, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.37735849056603776, |
|
"eval_loss": 10.561553001403809, |
|
"eval_runtime": 41.4742, |
|
"eval_samples_per_second": 21.531, |
|
"eval_steps_per_second": 5.401, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.39622641509433965, |
|
"grad_norm": 30.381412506103516, |
|
"learning_rate": 0.00014845508703326504, |
|
"loss": 36.4437, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.41509433962264153, |
|
"grad_norm": 25.108572006225586, |
|
"learning_rate": 0.00014373073204588556, |
|
"loss": 24.811, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.4339622641509434, |
|
"grad_norm": 25.86770248413086, |
|
"learning_rate": 0.00013888241754733208, |
|
"loss": 22.9303, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.4528301886792453, |
|
"grad_norm": 83.9566650390625, |
|
"learning_rate": 0.00013392388661180303, |
|
"loss": 14.6497, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.4716981132075472, |
|
"grad_norm": 46.80604934692383, |
|
"learning_rate": 0.0001288691947339621, |
|
"loss": 3.5489, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.4716981132075472, |
|
"eval_loss": 13.115553855895996, |
|
"eval_runtime": 41.4443, |
|
"eval_samples_per_second": 21.547, |
|
"eval_steps_per_second": 5.405, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.49056603773584906, |
|
"grad_norm": 60.517303466796875, |
|
"learning_rate": 0.0001237326699871115, |
|
"loss": 52.7818, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.5094339622641509, |
|
"grad_norm": 88.43193817138672, |
|
"learning_rate": 0.00011852887240871145, |
|
"loss": 28.1595, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.5283018867924528, |
|
"grad_norm": 39.37873458862305, |
|
"learning_rate": 0.00011327255272837221, |
|
"loss": 23.8529, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.5471698113207547, |
|
"grad_norm": 118.77758026123047, |
|
"learning_rate": 0.00010797861055530831, |
|
"loss": 14.4378, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.5660377358490566, |
|
"grad_norm": 136.2257537841797, |
|
"learning_rate": 0.00010266205214377748, |
|
"loss": 4.6485, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.5660377358490566, |
|
"eval_loss": 15.816683769226074, |
|
"eval_runtime": 41.4418, |
|
"eval_samples_per_second": 21.548, |
|
"eval_steps_per_second": 5.405, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.5849056603773585, |
|
"grad_norm": 54.89567947387695, |
|
"learning_rate": 9.733794785622253e-05, |
|
"loss": 53.7256, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.6037735849056604, |
|
"grad_norm": 32.46332931518555, |
|
"learning_rate": 9.202138944469168e-05, |
|
"loss": 24.7777, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.6226415094339622, |
|
"grad_norm": 23.78044319152832, |
|
"learning_rate": 8.672744727162781e-05, |
|
"loss": 23.3428, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.6415094339622641, |
|
"grad_norm": 223.9010467529297, |
|
"learning_rate": 8.147112759128859e-05, |
|
"loss": 17.2469, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.660377358490566, |
|
"grad_norm": 55.03976821899414, |
|
"learning_rate": 7.626733001288851e-05, |
|
"loss": 1.8853, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.660377358490566, |
|
"eval_loss": 12.180185317993164, |
|
"eval_runtime": 41.5042, |
|
"eval_samples_per_second": 21.516, |
|
"eval_steps_per_second": 5.397, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.6792452830188679, |
|
"grad_norm": 707.2769775390625, |
|
"learning_rate": 7.113080526603792e-05, |
|
"loss": 68.5596, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.6981132075471698, |
|
"grad_norm": 32.202911376953125, |
|
"learning_rate": 6.607611338819697e-05, |
|
"loss": 27.5274, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.7169811320754716, |
|
"grad_norm": 32.115299224853516, |
|
"learning_rate": 6.111758245266794e-05, |
|
"loss": 22.6553, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.7358490566037735, |
|
"grad_norm": 260.97906494140625, |
|
"learning_rate": 5.626926795411447e-05, |
|
"loss": 17.7957, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.7547169811320755, |
|
"grad_norm": 19.01742935180664, |
|
"learning_rate": 5.1544912966734994e-05, |
|
"loss": 1.821, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.7547169811320755, |
|
"eval_loss": 10.776637077331543, |
|
"eval_runtime": 41.3976, |
|
"eval_samples_per_second": 21.571, |
|
"eval_steps_per_second": 5.411, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.7735849056603774, |
|
"grad_norm": 814.9197998046875, |
|
"learning_rate": 4.695790918802576e-05, |
|
"loss": 66.7377, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.7924528301886793, |
|
"grad_norm": 30.05072784423828, |
|
"learning_rate": 4.252125897855932e-05, |
|
"loss": 30.2499, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.8113207547169812, |
|
"grad_norm": 24.943675994873047, |
|
"learning_rate": 3.824753850538082e-05, |
|
"loss": 23.6266, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.8301886792452831, |
|
"grad_norm": 178.87713623046875, |
|
"learning_rate": 3.414886209349615e-05, |
|
"loss": 11.4625, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.8490566037735849, |
|
"grad_norm": 0.000857223873026669, |
|
"learning_rate": 3.0236847886501542e-05, |
|
"loss": 1.0218, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.8490566037735849, |
|
"eval_loss": 20.48834991455078, |
|
"eval_runtime": 41.5021, |
|
"eval_samples_per_second": 21.517, |
|
"eval_steps_per_second": 5.397, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.8679245283018868, |
|
"grad_norm": 695.1412353515625, |
|
"learning_rate": 2.6522584913693294e-05, |
|
"loss": 73.7925, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.8867924528301887, |
|
"grad_norm": 118.63365936279297, |
|
"learning_rate": 2.301660165700936e-05, |
|
"loss": 34.6368, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.9056603773584906, |
|
"grad_norm": 50.097171783447266, |
|
"learning_rate": 1.9728836206903656e-05, |
|
"loss": 23.91, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.9245283018867925, |
|
"grad_norm": 127.65890502929688, |
|
"learning_rate": 1.6668608091748495e-05, |
|
"loss": 9.4513, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.9433962264150944, |
|
"grad_norm": 3.7929797172546387, |
|
"learning_rate": 1.3844591860619383e-05, |
|
"loss": 1.6077, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.9433962264150944, |
|
"eval_loss": 4.622971534729004, |
|
"eval_runtime": 41.3713, |
|
"eval_samples_per_second": 21.585, |
|
"eval_steps_per_second": 5.414, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.9622641509433962, |
|
"grad_norm": 320.5762634277344, |
|
"learning_rate": 1.1264792494342857e-05, |
|
"loss": 31.4437, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.9811320754716981, |
|
"grad_norm": 46.00068664550781, |
|
"learning_rate": 8.936522714508678e-06, |
|
"loss": 23.0031, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 4.2149858474731445, |
|
"learning_rate": 6.866382254766157e-06, |
|
"loss": 0.5552, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.0188679245283019, |
|
"grad_norm": 88.85747528076172, |
|
"learning_rate": 5.060239153161872e-06, |
|
"loss": 27.1639, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.0377358490566038, |
|
"grad_norm": 45.806617736816406, |
|
"learning_rate": 3.5232131185484076e-06, |
|
"loss": 24.8611, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.0377358490566038, |
|
"eval_loss": 3.9979488849639893, |
|
"eval_runtime": 41.4471, |
|
"eval_samples_per_second": 21.546, |
|
"eval_steps_per_second": 5.404, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.0566037735849056, |
|
"grad_norm": 22.6519832611084, |
|
"learning_rate": 2.259661018213333e-06, |
|
"loss": 22.9177, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.0754716981132075, |
|
"grad_norm": 48.2337760925293, |
|
"learning_rate": 1.2731645278655445e-06, |
|
"loss": 3.576, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.0943396226415094, |
|
"grad_norm": 13.664042472839355, |
|
"learning_rate": 5.665199789862907e-07, |
|
"loss": 0.382, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.1132075471698113, |
|
"grad_norm": 96.89643096923828, |
|
"learning_rate": 1.4173043232380557e-07, |
|
"loss": 25.7203, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.1320754716981132, |
|
"grad_norm": 19.166614532470703, |
|
"learning_rate": 0.0, |
|
"loss": 24.3324, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.1320754716981132, |
|
"eval_loss": 3.989581823348999, |
|
"eval_runtime": 41.5204, |
|
"eval_samples_per_second": 21.507, |
|
"eval_steps_per_second": 5.395, |
|
"step": 600 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 600, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 150, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.6255532515157606e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|