|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 78.07620237351655, |
|
"global_step": 125000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.832e-06, |
|
"loss": 2.2981, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.5832e-05, |
|
"loss": 1.7902, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 2.3832e-05, |
|
"loss": 1.6143, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.1832e-05, |
|
"loss": 1.5059, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 3.9832e-05, |
|
"loss": 1.4265, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 4.7824e-05, |
|
"loss": 1.3666, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 5.5808000000000005e-05, |
|
"loss": 1.3195, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 6.38e-05, |
|
"loss": 1.2827, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"learning_rate": 7.18e-05, |
|
"loss": 1.2519, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 7.9792e-05, |
|
"loss": 1.2251, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"learning_rate": 7.93224347826087e-05, |
|
"loss": 1.2031, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 7.862747826086958e-05, |
|
"loss": 1.1824, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"learning_rate": 7.793252173913044e-05, |
|
"loss": 1.1646, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 8.74, |
|
"learning_rate": 7.723756521739132e-05, |
|
"loss": 1.1514, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 9.37, |
|
"learning_rate": 7.654260869565218e-05, |
|
"loss": 1.1388, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 9.99, |
|
"learning_rate": 7.584765217391305e-05, |
|
"loss": 1.1281, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 10.62, |
|
"learning_rate": 7.5152e-05, |
|
"loss": 1.1184, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 11.24, |
|
"learning_rate": 7.445773913043479e-05, |
|
"loss": 1.11, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 11.87, |
|
"learning_rate": 7.376208695652175e-05, |
|
"loss": 1.1023, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 12.49, |
|
"learning_rate": 7.306713043478261e-05, |
|
"loss": 1.0945, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 13.12, |
|
"learning_rate": 7.237217391304349e-05, |
|
"loss": 1.0871, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 13.74, |
|
"learning_rate": 7.167791304347826e-05, |
|
"loss": 1.081, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 14.37, |
|
"learning_rate": 7.098226086956523e-05, |
|
"loss": 1.075, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 14.99, |
|
"learning_rate": 7.028730434782609e-05, |
|
"loss": 1.0698, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 15.62, |
|
"learning_rate": 6.959234782608696e-05, |
|
"loss": 1.0647, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 16.24, |
|
"learning_rate": 6.889739130434783e-05, |
|
"loss": 1.0585, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 16.86, |
|
"learning_rate": 6.820173913043479e-05, |
|
"loss": 1.0556, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 17.49, |
|
"learning_rate": 6.750678260869565e-05, |
|
"loss": 1.0504, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 18.11, |
|
"learning_rate": 6.681182608695653e-05, |
|
"loss": 1.0471, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 18.74, |
|
"learning_rate": 6.61168695652174e-05, |
|
"loss": 1.0426, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 19.36, |
|
"learning_rate": 6.542191304347826e-05, |
|
"loss": 1.0392, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 19.99, |
|
"learning_rate": 6.472626086956522e-05, |
|
"loss": 1.0364, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 20.61, |
|
"learning_rate": 6.4032e-05, |
|
"loss": 1.0324, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 21.24, |
|
"learning_rate": 6.333634782608696e-05, |
|
"loss": 1.0302, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 21.86, |
|
"learning_rate": 6.264139130434782e-05, |
|
"loss": 1.027, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 22.49, |
|
"learning_rate": 6.19464347826087e-05, |
|
"loss": 1.0234, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 23.11, |
|
"learning_rate": 6.125147826086957e-05, |
|
"loss": 1.0204, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 23.74, |
|
"learning_rate": 6.0556521739130436e-05, |
|
"loss": 1.0178, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 24.36, |
|
"learning_rate": 5.986156521739131e-05, |
|
"loss": 1.0162, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 24.98, |
|
"learning_rate": 5.9165913043478267e-05, |
|
"loss": 1.0136, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 25.61, |
|
"learning_rate": 5.847095652173914e-05, |
|
"loss": 1.0112, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 26.23, |
|
"learning_rate": 5.7776000000000004e-05, |
|
"loss": 1.0089, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 26.86, |
|
"learning_rate": 5.708104347826087e-05, |
|
"loss": 1.0069, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 27.48, |
|
"learning_rate": 5.6385391304347834e-05, |
|
"loss": 1.005, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 28.11, |
|
"learning_rate": 5.56904347826087e-05, |
|
"loss": 1.0026, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 28.73, |
|
"learning_rate": 5.499478260869565e-05, |
|
"loss": 1.0008, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 29.36, |
|
"learning_rate": 5.429982608695653e-05, |
|
"loss": 0.9986, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 29.98, |
|
"learning_rate": 5.3604869565217396e-05, |
|
"loss": 0.9967, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 30.61, |
|
"learning_rate": 5.2909913043478264e-05, |
|
"loss": 0.9949, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 31.23, |
|
"learning_rate": 5.2215652173913046e-05, |
|
"loss": 0.9928, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 31.86, |
|
"learning_rate": 5.1520695652173915e-05, |
|
"loss": 0.992, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 32.48, |
|
"learning_rate": 5.082504347826088e-05, |
|
"loss": 0.9902, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 33.1, |
|
"learning_rate": 5.0129391304347825e-05, |
|
"loss": 0.9877, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 33.73, |
|
"learning_rate": 4.9435130434782614e-05, |
|
"loss": 0.9865, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 34.35, |
|
"learning_rate": 4.873947826086957e-05, |
|
"loss": 0.9859, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 34.98, |
|
"learning_rate": 4.804452173913044e-05, |
|
"loss": 0.9839, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 35.6, |
|
"learning_rate": 4.7349565217391306e-05, |
|
"loss": 0.9825, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 36.23, |
|
"learning_rate": 4.665460869565218e-05, |
|
"loss": 0.9812, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 36.85, |
|
"learning_rate": 4.595965217391305e-05, |
|
"loss": 0.9797, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 37.48, |
|
"learning_rate": 4.526469565217392e-05, |
|
"loss": 0.9784, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 38.1, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.9733, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 38.73, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 0.9707, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 39.35, |
|
"learning_rate": 2.4e-05, |
|
"loss": 0.9714, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 39.98, |
|
"learning_rate": 3.2000000000000005e-05, |
|
"loss": 0.9716, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 40.6, |
|
"learning_rate": 4e-05, |
|
"loss": 0.9716, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 41.22, |
|
"learning_rate": 4.8e-05, |
|
"loss": 0.9734, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 41.85, |
|
"learning_rate": 5.6e-05, |
|
"loss": 0.9748, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 42.47, |
|
"learning_rate": 6.400000000000001e-05, |
|
"loss": 0.9757, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 43.1, |
|
"learning_rate": 7.2e-05, |
|
"loss": 0.9772, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 43.72, |
|
"learning_rate": 8e-05, |
|
"loss": 0.977, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 44.35, |
|
"learning_rate": 7.930434782608697e-05, |
|
"loss": 0.9782, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 44.97, |
|
"learning_rate": 7.860869565217392e-05, |
|
"loss": 0.977, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 45.6, |
|
"learning_rate": 7.791304347826088e-05, |
|
"loss": 0.9752, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 46.22, |
|
"learning_rate": 7.721739130434783e-05, |
|
"loss": 0.9747, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 46.85, |
|
"learning_rate": 7.652173913043479e-05, |
|
"loss": 0.9728, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 47.47, |
|
"learning_rate": 7.582608695652174e-05, |
|
"loss": 0.9701, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 48.09, |
|
"learning_rate": 7.51304347826087e-05, |
|
"loss": 0.9699, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 48.72, |
|
"learning_rate": 7.443478260869565e-05, |
|
"loss": 0.9681, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 49.34, |
|
"learning_rate": 7.373913043478261e-05, |
|
"loss": 0.9669, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 49.97, |
|
"learning_rate": 7.304347826086957e-05, |
|
"loss": 0.966, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 50.59, |
|
"learning_rate": 7.864000000000001e-06, |
|
"loss": 0.9577, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 51.22, |
|
"learning_rate": 1.5856e-05, |
|
"loss": 0.9535, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 51.84, |
|
"learning_rate": 2.3848e-05, |
|
"loss": 0.9518, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 52.47, |
|
"learning_rate": 3.184000000000001e-05, |
|
"loss": 0.9525, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 53.09, |
|
"learning_rate": 3.9832e-05, |
|
"loss": 0.9531, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 53.72, |
|
"learning_rate": 4.7824e-05, |
|
"loss": 0.9537, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 54.34, |
|
"learning_rate": 5.5816e-05, |
|
"loss": 0.9541, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 54.97, |
|
"learning_rate": 6.380800000000001e-05, |
|
"loss": 0.9561, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 55.59, |
|
"learning_rate": 7.1792e-05, |
|
"loss": 0.9562, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 56.21, |
|
"learning_rate": 7.9792e-05, |
|
"loss": 0.959, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 56.84, |
|
"learning_rate": 7.932313043478262e-05, |
|
"loss": 0.9589, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 57.46, |
|
"learning_rate": 7.862817391304348e-05, |
|
"loss": 0.9586, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 58.09, |
|
"learning_rate": 7.793252173913044e-05, |
|
"loss": 0.958, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 58.71, |
|
"learning_rate": 7.723756521739132e-05, |
|
"loss": 0.9558, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 59.34, |
|
"learning_rate": 7.654260869565218e-05, |
|
"loss": 0.9549, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 59.96, |
|
"learning_rate": 7.584695652173914e-05, |
|
"loss": 0.954, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 60.59, |
|
"learning_rate": 7.5152e-05, |
|
"loss": 0.9522, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 61.21, |
|
"learning_rate": 7.445704347826088e-05, |
|
"loss": 0.9517, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 61.84, |
|
"learning_rate": 7.376208695652175e-05, |
|
"loss": 0.9509, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 62.46, |
|
"learning_rate": 7.306852173913044e-05, |
|
"loss": 0.9497, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 63.09, |
|
"learning_rate": 7.23728695652174e-05, |
|
"loss": 0.9483, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 63.71, |
|
"learning_rate": 7.167721739130435e-05, |
|
"loss": 0.947, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 64.33, |
|
"learning_rate": 7.098226086956523e-05, |
|
"loss": 0.9463, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 64.96, |
|
"learning_rate": 7.028730434782609e-05, |
|
"loss": 0.9455, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 65.58, |
|
"learning_rate": 6.959165217391305e-05, |
|
"loss": 0.9437, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 66.21, |
|
"learning_rate": 6.889669565217393e-05, |
|
"loss": 0.9437, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 66.83, |
|
"learning_rate": 6.820173913043479e-05, |
|
"loss": 0.9429, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 67.46, |
|
"learning_rate": 6.750608695652175e-05, |
|
"loss": 0.9418, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 68.08, |
|
"learning_rate": 6.681113043478261e-05, |
|
"loss": 0.941, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 68.71, |
|
"learning_rate": 6.611617391304349e-05, |
|
"loss": 0.9397, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 69.33, |
|
"learning_rate": 6.542121739130435e-05, |
|
"loss": 0.939, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 69.96, |
|
"learning_rate": 6.472556521739131e-05, |
|
"loss": 0.9382, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 70.58, |
|
"learning_rate": 6.403060869565217e-05, |
|
"loss": 0.9375, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 71.21, |
|
"learning_rate": 6.333634782608696e-05, |
|
"loss": 0.9362, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 71.83, |
|
"learning_rate": 6.264069565217392e-05, |
|
"loss": 0.9362, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 72.45, |
|
"learning_rate": 6.194573913043479e-05, |
|
"loss": 0.9348, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 73.08, |
|
"learning_rate": 6.125078260869566e-05, |
|
"loss": 0.934, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 73.7, |
|
"learning_rate": 6.055513043478261e-05, |
|
"loss": 0.9325, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 74.33, |
|
"learning_rate": 5.986017391304348e-05, |
|
"loss": 0.9321, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 74.95, |
|
"learning_rate": 5.916521739130435e-05, |
|
"loss": 0.9315, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 75.58, |
|
"learning_rate": 5.847026086956522e-05, |
|
"loss": 0.9299, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 76.2, |
|
"learning_rate": 5.777530434782609e-05, |
|
"loss": 0.9291, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 76.83, |
|
"learning_rate": 5.708034782608696e-05, |
|
"loss": 0.9291, |
|
"step": 123000 |
|
}, |
|
{ |
|
"epoch": 77.45, |
|
"learning_rate": 5.6385391304347834e-05, |
|
"loss": 0.9277, |
|
"step": 124000 |
|
}, |
|
{ |
|
"epoch": 78.08, |
|
"learning_rate": 5.56904347826087e-05, |
|
"loss": 0.9278, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 78.08, |
|
"step": 125000, |
|
"total_flos": 6.73699192934362e+19, |
|
"train_loss": 0.18748452099609375, |
|
"train_runtime": 7317.8985, |
|
"train_samples_per_second": 34982.721, |
|
"train_steps_per_second": 17.081 |
|
} |
|
], |
|
"max_steps": 125000, |
|
"num_train_epochs": 79, |
|
"total_flos": 6.73699192934362e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|