{ "best_metric": null, "best_model_checkpoint": null, "epoch": 10.0, "eval_steps": 500, "global_step": 4920, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0020325203252032522, "grad_norm": 0.26638158174575627, "learning_rate": 1.3513513513513515e-07, "loss": 2.0675, "step": 1 }, { "epoch": 0.0040650406504065045, "grad_norm": 0.27310723901066764, "learning_rate": 2.702702702702703e-07, "loss": 1.9816, "step": 2 }, { "epoch": 0.006097560975609756, "grad_norm": 0.26694621856314155, "learning_rate": 4.0540540540540546e-07, "loss": 2.2887, "step": 3 }, { "epoch": 0.008130081300813009, "grad_norm": 0.2479619574080673, "learning_rate": 5.405405405405406e-07, "loss": 1.9933, "step": 4 }, { "epoch": 0.01016260162601626, "grad_norm": 0.22330215332953987, "learning_rate": 6.756756756756758e-07, "loss": 2.038, "step": 5 }, { "epoch": 0.012195121951219513, "grad_norm": 0.2494157711452319, "learning_rate": 8.108108108108109e-07, "loss": 2.0548, "step": 6 }, { "epoch": 0.014227642276422764, "grad_norm": 0.24594873531353217, "learning_rate": 9.459459459459461e-07, "loss": 2.0661, "step": 7 }, { "epoch": 0.016260162601626018, "grad_norm": 0.2660572965030069, "learning_rate": 1.0810810810810812e-06, "loss": 1.9151, "step": 8 }, { "epoch": 0.018292682926829267, "grad_norm": 0.24141645271110918, "learning_rate": 1.2162162162162164e-06, "loss": 1.9116, "step": 9 }, { "epoch": 0.02032520325203252, "grad_norm": 0.21248333572292508, "learning_rate": 1.3513513513513515e-06, "loss": 1.8348, "step": 10 }, { "epoch": 0.022357723577235773, "grad_norm": 0.2405131493577988, "learning_rate": 1.4864864864864868e-06, "loss": 1.9573, "step": 11 }, { "epoch": 0.024390243902439025, "grad_norm": 0.2713754906890974, "learning_rate": 1.6216216216216219e-06, "loss": 2.2145, "step": 12 }, { "epoch": 0.026422764227642278, "grad_norm": 0.24676476763973695, "learning_rate": 1.756756756756757e-06, "loss": 1.8709, "step": 13 }, { "epoch": 0.028455284552845527, "grad_norm": 0.2470353424240876, "learning_rate": 1.8918918918918922e-06, "loss": 1.8733, "step": 14 }, { "epoch": 0.03048780487804878, "grad_norm": 0.24054645879626352, "learning_rate": 2.0270270270270273e-06, "loss": 1.9181, "step": 15 }, { "epoch": 0.032520325203252036, "grad_norm": 0.2742836712893144, "learning_rate": 2.1621621621621623e-06, "loss": 2.0646, "step": 16 }, { "epoch": 0.034552845528455285, "grad_norm": 0.25807046239359827, "learning_rate": 2.297297297297298e-06, "loss": 2.1356, "step": 17 }, { "epoch": 0.036585365853658534, "grad_norm": 0.2758353954233353, "learning_rate": 2.432432432432433e-06, "loss": 2.0909, "step": 18 }, { "epoch": 0.03861788617886179, "grad_norm": 0.2888901320619277, "learning_rate": 2.5675675675675675e-06, "loss": 2.1493, "step": 19 }, { "epoch": 0.04065040650406504, "grad_norm": 0.34173192471639885, "learning_rate": 2.702702702702703e-06, "loss": 2.0987, "step": 20 }, { "epoch": 0.042682926829268296, "grad_norm": 0.26149149930755766, "learning_rate": 2.837837837837838e-06, "loss": 1.945, "step": 21 }, { "epoch": 0.044715447154471545, "grad_norm": 0.2760818358250014, "learning_rate": 2.9729729729729736e-06, "loss": 2.1805, "step": 22 }, { "epoch": 0.046747967479674794, "grad_norm": 0.21966106106546474, "learning_rate": 3.1081081081081082e-06, "loss": 1.6791, "step": 23 }, { "epoch": 0.04878048780487805, "grad_norm": 0.2842775595290317, "learning_rate": 3.2432432432432437e-06, "loss": 2.1092, "step": 24 }, { "epoch": 0.0508130081300813, "grad_norm": 0.22726816311145973, "learning_rate": 3.3783783783783788e-06, "loss": 1.7879, "step": 25 }, { "epoch": 0.052845528455284556, "grad_norm": 0.2634567668898366, "learning_rate": 3.513513513513514e-06, "loss": 2.2133, "step": 26 }, { "epoch": 0.054878048780487805, "grad_norm": 0.2934314502555563, "learning_rate": 3.648648648648649e-06, "loss": 2.116, "step": 27 }, { "epoch": 0.056910569105691054, "grad_norm": 0.2669352962249472, "learning_rate": 3.7837837837837844e-06, "loss": 1.9969, "step": 28 }, { "epoch": 0.05894308943089431, "grad_norm": 0.2946526020858884, "learning_rate": 3.918918918918919e-06, "loss": 1.9451, "step": 29 }, { "epoch": 0.06097560975609756, "grad_norm": 0.2788390716946643, "learning_rate": 4.0540540540540545e-06, "loss": 1.9784, "step": 30 }, { "epoch": 0.06300813008130081, "grad_norm": 0.2901731022333901, "learning_rate": 4.189189189189189e-06, "loss": 2.0166, "step": 31 }, { "epoch": 0.06504065040650407, "grad_norm": 0.2879289590944654, "learning_rate": 4.324324324324325e-06, "loss": 1.9265, "step": 32 }, { "epoch": 0.06707317073170732, "grad_norm": 0.30797150398271883, "learning_rate": 4.45945945945946e-06, "loss": 2.1205, "step": 33 }, { "epoch": 0.06910569105691057, "grad_norm": 0.23844150634738745, "learning_rate": 4.594594594594596e-06, "loss": 1.9885, "step": 34 }, { "epoch": 0.07113821138211382, "grad_norm": 0.2660286787888433, "learning_rate": 4.72972972972973e-06, "loss": 1.8769, "step": 35 }, { "epoch": 0.07317073170731707, "grad_norm": 0.328047597777894, "learning_rate": 4.864864864864866e-06, "loss": 2.0258, "step": 36 }, { "epoch": 0.07520325203252033, "grad_norm": 0.36701397762452254, "learning_rate": 5e-06, "loss": 1.9928, "step": 37 }, { "epoch": 0.07723577235772358, "grad_norm": 0.3201788817189425, "learning_rate": 5.135135135135135e-06, "loss": 1.9658, "step": 38 }, { "epoch": 0.07926829268292683, "grad_norm": 0.30154323972739955, "learning_rate": 5.2702702702702705e-06, "loss": 1.9894, "step": 39 }, { "epoch": 0.08130081300813008, "grad_norm": 0.3324136304476669, "learning_rate": 5.405405405405406e-06, "loss": 2.2665, "step": 40 }, { "epoch": 0.08333333333333333, "grad_norm": 0.34587067874298405, "learning_rate": 5.540540540540541e-06, "loss": 2.194, "step": 41 }, { "epoch": 0.08536585365853659, "grad_norm": 0.2750126715841735, "learning_rate": 5.675675675675676e-06, "loss": 2.0026, "step": 42 }, { "epoch": 0.08739837398373984, "grad_norm": 0.43757993223722025, "learning_rate": 5.810810810810811e-06, "loss": 2.2732, "step": 43 }, { "epoch": 0.08943089430894309, "grad_norm": 0.3021810760766025, "learning_rate": 5.945945945945947e-06, "loss": 1.9174, "step": 44 }, { "epoch": 0.09146341463414634, "grad_norm": 0.30196325740761853, "learning_rate": 6.081081081081082e-06, "loss": 1.7567, "step": 45 }, { "epoch": 0.09349593495934959, "grad_norm": 0.3159658734740649, "learning_rate": 6.2162162162162164e-06, "loss": 1.9686, "step": 46 }, { "epoch": 0.09552845528455285, "grad_norm": 0.33241023186086166, "learning_rate": 6.351351351351351e-06, "loss": 1.8695, "step": 47 }, { "epoch": 0.0975609756097561, "grad_norm": 0.332794490258164, "learning_rate": 6.486486486486487e-06, "loss": 1.7999, "step": 48 }, { "epoch": 0.09959349593495935, "grad_norm": 0.36174301339265025, "learning_rate": 6.621621621621622e-06, "loss": 1.9574, "step": 49 }, { "epoch": 0.1016260162601626, "grad_norm": 0.37172521009173376, "learning_rate": 6.7567567567567575e-06, "loss": 2.1735, "step": 50 }, { "epoch": 0.10365853658536585, "grad_norm": 0.31284873174038447, "learning_rate": 6.891891891891892e-06, "loss": 1.9748, "step": 51 }, { "epoch": 0.10569105691056911, "grad_norm": 0.41823869607911235, "learning_rate": 7.027027027027028e-06, "loss": 1.9962, "step": 52 }, { "epoch": 0.10772357723577236, "grad_norm": 0.4419760780095649, "learning_rate": 7.162162162162163e-06, "loss": 1.9109, "step": 53 }, { "epoch": 0.10975609756097561, "grad_norm": 0.3808547264421021, "learning_rate": 7.297297297297298e-06, "loss": 1.9526, "step": 54 }, { "epoch": 0.11178861788617886, "grad_norm": 0.36233012810526954, "learning_rate": 7.4324324324324324e-06, "loss": 1.7079, "step": 55 }, { "epoch": 0.11382113821138211, "grad_norm": 0.3877056017136688, "learning_rate": 7.567567567567569e-06, "loss": 1.967, "step": 56 }, { "epoch": 0.11585365853658537, "grad_norm": 0.41579607060864765, "learning_rate": 7.702702702702704e-06, "loss": 2.2282, "step": 57 }, { "epoch": 0.11788617886178862, "grad_norm": 0.43809590096012535, "learning_rate": 7.837837837837838e-06, "loss": 2.0285, "step": 58 }, { "epoch": 0.11991869918699187, "grad_norm": 0.46422984992796484, "learning_rate": 7.972972972972974e-06, "loss": 1.7735, "step": 59 }, { "epoch": 0.12195121951219512, "grad_norm": 0.4441856205669471, "learning_rate": 8.108108108108109e-06, "loss": 1.9783, "step": 60 }, { "epoch": 0.12398373983739837, "grad_norm": 0.42689456517904684, "learning_rate": 8.243243243243245e-06, "loss": 2.05, "step": 61 }, { "epoch": 0.12601626016260162, "grad_norm": 0.4488122609801876, "learning_rate": 8.378378378378378e-06, "loss": 2.0123, "step": 62 }, { "epoch": 0.12804878048780488, "grad_norm": 0.5152977416830085, "learning_rate": 8.513513513513514e-06, "loss": 2.1787, "step": 63 }, { "epoch": 0.13008130081300814, "grad_norm": 0.5126840380335577, "learning_rate": 8.64864864864865e-06, "loss": 1.8205, "step": 64 }, { "epoch": 0.13211382113821138, "grad_norm": 0.533714973212888, "learning_rate": 8.783783783783785e-06, "loss": 2.0384, "step": 65 }, { "epoch": 0.13414634146341464, "grad_norm": 0.44464369214899996, "learning_rate": 8.91891891891892e-06, "loss": 1.957, "step": 66 }, { "epoch": 0.13617886178861788, "grad_norm": 0.5030929824530757, "learning_rate": 9.054054054054054e-06, "loss": 1.9113, "step": 67 }, { "epoch": 0.13821138211382114, "grad_norm": 0.4748516722147689, "learning_rate": 9.189189189189191e-06, "loss": 1.7268, "step": 68 }, { "epoch": 0.1402439024390244, "grad_norm": 0.513340433666549, "learning_rate": 9.324324324324325e-06, "loss": 1.9849, "step": 69 }, { "epoch": 0.14227642276422764, "grad_norm": 0.48974582588313254, "learning_rate": 9.45945945945946e-06, "loss": 1.8533, "step": 70 }, { "epoch": 0.1443089430894309, "grad_norm": 0.5383912075966719, "learning_rate": 9.594594594594594e-06, "loss": 1.879, "step": 71 }, { "epoch": 0.14634146341463414, "grad_norm": 0.5780398496697261, "learning_rate": 9.729729729729732e-06, "loss": 1.8515, "step": 72 }, { "epoch": 0.1483739837398374, "grad_norm": 0.5792923140088602, "learning_rate": 9.864864864864865e-06, "loss": 1.7941, "step": 73 }, { "epoch": 0.15040650406504066, "grad_norm": 0.6103338636716018, "learning_rate": 1e-05, "loss": 1.8967, "step": 74 }, { "epoch": 0.1524390243902439, "grad_norm": 0.5521549627746493, "learning_rate": 1.0135135135135136e-05, "loss": 2.062, "step": 75 }, { "epoch": 0.15447154471544716, "grad_norm": 0.49667826178980906, "learning_rate": 1.027027027027027e-05, "loss": 1.8472, "step": 76 }, { "epoch": 0.1565040650406504, "grad_norm": 0.5407086436651087, "learning_rate": 1.0405405405405407e-05, "loss": 1.7384, "step": 77 }, { "epoch": 0.15853658536585366, "grad_norm": 0.5802202311742781, "learning_rate": 1.0540540540540541e-05, "loss": 1.9283, "step": 78 }, { "epoch": 0.16056910569105692, "grad_norm": 0.476967574495256, "learning_rate": 1.0675675675675677e-05, "loss": 1.8267, "step": 79 }, { "epoch": 0.16260162601626016, "grad_norm": 0.6001610186866417, "learning_rate": 1.0810810810810812e-05, "loss": 1.8288, "step": 80 }, { "epoch": 0.16463414634146342, "grad_norm": 0.5070823276209965, "learning_rate": 1.0945945945945946e-05, "loss": 1.7281, "step": 81 }, { "epoch": 0.16666666666666666, "grad_norm": 0.5602421231157648, "learning_rate": 1.1081081081081081e-05, "loss": 2.0092, "step": 82 }, { "epoch": 0.16869918699186992, "grad_norm": 0.5571854783072151, "learning_rate": 1.1216216216216219e-05, "loss": 1.843, "step": 83 }, { "epoch": 0.17073170731707318, "grad_norm": 0.6305360980370397, "learning_rate": 1.1351351351351352e-05, "loss": 1.7498, "step": 84 }, { "epoch": 0.17276422764227642, "grad_norm": 0.6332330762093754, "learning_rate": 1.1486486486486488e-05, "loss": 1.8271, "step": 85 }, { "epoch": 0.17479674796747968, "grad_norm": 0.5263379983638994, "learning_rate": 1.1621621621621622e-05, "loss": 1.7986, "step": 86 }, { "epoch": 0.17682926829268292, "grad_norm": 0.5699857243663454, "learning_rate": 1.1756756756756757e-05, "loss": 1.7866, "step": 87 }, { "epoch": 0.17886178861788618, "grad_norm": 0.5459562544641535, "learning_rate": 1.1891891891891894e-05, "loss": 1.8598, "step": 88 }, { "epoch": 0.18089430894308944, "grad_norm": 0.6269705097827972, "learning_rate": 1.2027027027027028e-05, "loss": 1.8462, "step": 89 }, { "epoch": 0.18292682926829268, "grad_norm": 0.5398025783285727, "learning_rate": 1.2162162162162164e-05, "loss": 1.8385, "step": 90 }, { "epoch": 0.18495934959349594, "grad_norm": 0.541866197969969, "learning_rate": 1.2297297297297299e-05, "loss": 1.562, "step": 91 }, { "epoch": 0.18699186991869918, "grad_norm": 0.5125397438719046, "learning_rate": 1.2432432432432433e-05, "loss": 1.6564, "step": 92 }, { "epoch": 0.18902439024390244, "grad_norm": 0.455991168484729, "learning_rate": 1.2567567567567568e-05, "loss": 1.6706, "step": 93 }, { "epoch": 0.1910569105691057, "grad_norm": 0.4913611513705073, "learning_rate": 1.2702702702702702e-05, "loss": 1.7659, "step": 94 }, { "epoch": 0.19308943089430894, "grad_norm": 0.641151947177122, "learning_rate": 1.283783783783784e-05, "loss": 1.7493, "step": 95 }, { "epoch": 0.1951219512195122, "grad_norm": 0.46109387574760535, "learning_rate": 1.2972972972972975e-05, "loss": 1.7085, "step": 96 }, { "epoch": 0.19715447154471544, "grad_norm": 0.4249797151022942, "learning_rate": 1.3108108108108109e-05, "loss": 1.62, "step": 97 }, { "epoch": 0.1991869918699187, "grad_norm": 0.5608746999580215, "learning_rate": 1.3243243243243244e-05, "loss": 1.729, "step": 98 }, { "epoch": 0.20121951219512196, "grad_norm": 0.4429128789898387, "learning_rate": 1.3378378378378381e-05, "loss": 1.5967, "step": 99 }, { "epoch": 0.2032520325203252, "grad_norm": 0.5012917466098897, "learning_rate": 1.3513513513513515e-05, "loss": 1.7438, "step": 100 }, { "epoch": 0.20528455284552846, "grad_norm": 0.502660790347558, "learning_rate": 1.364864864864865e-05, "loss": 1.7512, "step": 101 }, { "epoch": 0.2073170731707317, "grad_norm": 0.5041999333257511, "learning_rate": 1.3783783783783784e-05, "loss": 1.5514, "step": 102 }, { "epoch": 0.20934959349593496, "grad_norm": 0.5175550658947365, "learning_rate": 1.391891891891892e-05, "loss": 1.6419, "step": 103 }, { "epoch": 0.21138211382113822, "grad_norm": 0.5303810355493661, "learning_rate": 1.4054054054054055e-05, "loss": 1.5234, "step": 104 }, { "epoch": 0.21341463414634146, "grad_norm": 0.6116947870470436, "learning_rate": 1.4189189189189189e-05, "loss": 1.8669, "step": 105 }, { "epoch": 0.21544715447154472, "grad_norm": 0.528692442993893, "learning_rate": 1.4324324324324326e-05, "loss": 1.5288, "step": 106 }, { "epoch": 0.21747967479674796, "grad_norm": 0.673656796412201, "learning_rate": 1.4459459459459462e-05, "loss": 1.8058, "step": 107 }, { "epoch": 0.21951219512195122, "grad_norm": 0.5127811069450205, "learning_rate": 1.4594594594594596e-05, "loss": 1.2773, "step": 108 }, { "epoch": 0.22154471544715448, "grad_norm": 0.5283415775432352, "learning_rate": 1.4729729729729731e-05, "loss": 1.6658, "step": 109 }, { "epoch": 0.22357723577235772, "grad_norm": 0.5346526299161833, "learning_rate": 1.4864864864864865e-05, "loss": 1.4156, "step": 110 }, { "epoch": 0.22560975609756098, "grad_norm": 0.5275451268008083, "learning_rate": 1.5000000000000002e-05, "loss": 1.6986, "step": 111 }, { "epoch": 0.22764227642276422, "grad_norm": 0.5442878185251289, "learning_rate": 1.5135135135135138e-05, "loss": 1.7011, "step": 112 }, { "epoch": 0.22967479674796748, "grad_norm": 0.5957302970205195, "learning_rate": 1.527027027027027e-05, "loss": 1.507, "step": 113 }, { "epoch": 0.23170731707317074, "grad_norm": 0.4632091042318271, "learning_rate": 1.540540540540541e-05, "loss": 1.3852, "step": 114 }, { "epoch": 0.23373983739837398, "grad_norm": 0.406816778444053, "learning_rate": 1.554054054054054e-05, "loss": 1.4839, "step": 115 }, { "epoch": 0.23577235772357724, "grad_norm": 0.5403511190324622, "learning_rate": 1.5675675675675676e-05, "loss": 1.5465, "step": 116 }, { "epoch": 0.23780487804878048, "grad_norm": 0.47284155263410366, "learning_rate": 1.581081081081081e-05, "loss": 1.5257, "step": 117 }, { "epoch": 0.23983739837398374, "grad_norm": 0.5334205673370109, "learning_rate": 1.5945945945945947e-05, "loss": 1.365, "step": 118 }, { "epoch": 0.241869918699187, "grad_norm": 0.5196343977662706, "learning_rate": 1.6081081081081083e-05, "loss": 1.3581, "step": 119 }, { "epoch": 0.24390243902439024, "grad_norm": 0.5005555917464704, "learning_rate": 1.6216216216216218e-05, "loss": 1.5632, "step": 120 }, { "epoch": 0.2459349593495935, "grad_norm": 0.4520062619717936, "learning_rate": 1.6351351351351354e-05, "loss": 1.47, "step": 121 }, { "epoch": 0.24796747967479674, "grad_norm": 0.5682578714967847, "learning_rate": 1.648648648648649e-05, "loss": 1.5291, "step": 122 }, { "epoch": 0.25, "grad_norm": 0.49518081727668006, "learning_rate": 1.662162162162162e-05, "loss": 1.4649, "step": 123 }, { "epoch": 0.25203252032520324, "grad_norm": 0.5450397443916377, "learning_rate": 1.6756756756756757e-05, "loss": 1.613, "step": 124 }, { "epoch": 0.2540650406504065, "grad_norm": 0.5288824055439941, "learning_rate": 1.6891891891891896e-05, "loss": 1.5524, "step": 125 }, { "epoch": 0.25609756097560976, "grad_norm": 0.6273100391485248, "learning_rate": 1.7027027027027028e-05, "loss": 1.5758, "step": 126 }, { "epoch": 0.258130081300813, "grad_norm": 0.5660312655010844, "learning_rate": 1.7162162162162163e-05, "loss": 1.5333, "step": 127 }, { "epoch": 0.2601626016260163, "grad_norm": 0.5520814691629921, "learning_rate": 1.72972972972973e-05, "loss": 1.5449, "step": 128 }, { "epoch": 0.2621951219512195, "grad_norm": 0.45015957789545086, "learning_rate": 1.7432432432432434e-05, "loss": 1.3946, "step": 129 }, { "epoch": 0.26422764227642276, "grad_norm": 0.46517012017437076, "learning_rate": 1.756756756756757e-05, "loss": 1.3244, "step": 130 }, { "epoch": 0.266260162601626, "grad_norm": 0.5161467864727404, "learning_rate": 1.7702702702702702e-05, "loss": 1.3873, "step": 131 }, { "epoch": 0.2682926829268293, "grad_norm": 0.5548907182685687, "learning_rate": 1.783783783783784e-05, "loss": 1.5925, "step": 132 }, { "epoch": 0.2703252032520325, "grad_norm": 0.5081011334123496, "learning_rate": 1.7972972972972976e-05, "loss": 1.3991, "step": 133 }, { "epoch": 0.27235772357723576, "grad_norm": 0.4987265730840859, "learning_rate": 1.8108108108108108e-05, "loss": 1.4442, "step": 134 }, { "epoch": 0.27439024390243905, "grad_norm": 0.49614117194728724, "learning_rate": 1.8243243243243244e-05, "loss": 1.3452, "step": 135 }, { "epoch": 0.2764227642276423, "grad_norm": 0.5481594393974627, "learning_rate": 1.8378378378378383e-05, "loss": 1.4192, "step": 136 }, { "epoch": 0.2784552845528455, "grad_norm": 0.4212041763953864, "learning_rate": 1.8513513513513515e-05, "loss": 1.4477, "step": 137 }, { "epoch": 0.2804878048780488, "grad_norm": 0.45861856810336465, "learning_rate": 1.864864864864865e-05, "loss": 1.4169, "step": 138 }, { "epoch": 0.28252032520325204, "grad_norm": 0.5089270926808518, "learning_rate": 1.8783783783783786e-05, "loss": 1.4535, "step": 139 }, { "epoch": 0.2845528455284553, "grad_norm": 0.5378055644689794, "learning_rate": 1.891891891891892e-05, "loss": 1.1821, "step": 140 }, { "epoch": 0.2865853658536585, "grad_norm": 0.49182295085803684, "learning_rate": 1.9054054054054057e-05, "loss": 1.5199, "step": 141 }, { "epoch": 0.2886178861788618, "grad_norm": 0.5867085379487282, "learning_rate": 1.918918918918919e-05, "loss": 1.2422, "step": 142 }, { "epoch": 0.29065040650406504, "grad_norm": 0.5798084312570029, "learning_rate": 1.9324324324324328e-05, "loss": 1.4636, "step": 143 }, { "epoch": 0.2926829268292683, "grad_norm": 0.5553965865343059, "learning_rate": 1.9459459459459463e-05, "loss": 1.3106, "step": 144 }, { "epoch": 0.29471544715447157, "grad_norm": 0.65046137560153, "learning_rate": 1.9594594594594595e-05, "loss": 1.6027, "step": 145 }, { "epoch": 0.2967479674796748, "grad_norm": 0.5839047816217893, "learning_rate": 1.972972972972973e-05, "loss": 1.4144, "step": 146 }, { "epoch": 0.29878048780487804, "grad_norm": 0.5569596116488467, "learning_rate": 1.9864864864864866e-05, "loss": 1.1476, "step": 147 }, { "epoch": 0.3008130081300813, "grad_norm": 0.5416449094512765, "learning_rate": 2e-05, "loss": 1.4352, "step": 148 }, { "epoch": 0.30284552845528456, "grad_norm": 0.5506318392458708, "learning_rate": 1.999999783295034e-05, "loss": 1.273, "step": 149 }, { "epoch": 0.3048780487804878, "grad_norm": 0.5518893604343933, "learning_rate": 1.9999991331802294e-05, "loss": 1.2675, "step": 150 }, { "epoch": 0.30691056910569103, "grad_norm": 0.5167606317698892, "learning_rate": 1.9999980496558683e-05, "loss": 1.3724, "step": 151 }, { "epoch": 0.3089430894308943, "grad_norm": 0.4704404074896954, "learning_rate": 1.99999653272242e-05, "loss": 1.293, "step": 152 }, { "epoch": 0.31097560975609756, "grad_norm": 0.49273639890579335, "learning_rate": 1.999994582380542e-05, "loss": 1.1906, "step": 153 }, { "epoch": 0.3130081300813008, "grad_norm": 0.5483874588703126, "learning_rate": 1.99999219863108e-05, "loss": 1.1866, "step": 154 }, { "epoch": 0.3150406504065041, "grad_norm": 0.4693982795405603, "learning_rate": 1.9999893814750667e-05, "loss": 1.2755, "step": 155 }, { "epoch": 0.3170731707317073, "grad_norm": 0.5346163004909783, "learning_rate": 1.9999861309137232e-05, "loss": 1.2694, "step": 156 }, { "epoch": 0.31910569105691056, "grad_norm": 0.5063484990579301, "learning_rate": 1.9999824469484583e-05, "loss": 1.3244, "step": 157 }, { "epoch": 0.32113821138211385, "grad_norm": 0.5197807817774548, "learning_rate": 1.999978329580869e-05, "loss": 1.2746, "step": 158 }, { "epoch": 0.3231707317073171, "grad_norm": 0.5241698441816027, "learning_rate": 1.999973778812739e-05, "loss": 1.2228, "step": 159 }, { "epoch": 0.3252032520325203, "grad_norm": 0.5235866440740556, "learning_rate": 1.999968794646042e-05, "loss": 1.2699, "step": 160 }, { "epoch": 0.32723577235772355, "grad_norm": 0.6203596702273114, "learning_rate": 1.9999633770829368e-05, "loss": 1.2266, "step": 161 }, { "epoch": 0.32926829268292684, "grad_norm": 0.5715749137527399, "learning_rate": 1.9999575261257723e-05, "loss": 1.4359, "step": 162 }, { "epoch": 0.3313008130081301, "grad_norm": 0.5763384129195755, "learning_rate": 1.9999512417770836e-05, "loss": 1.289, "step": 163 }, { "epoch": 0.3333333333333333, "grad_norm": 0.6339400872600929, "learning_rate": 1.9999445240395953e-05, "loss": 1.2333, "step": 164 }, { "epoch": 0.3353658536585366, "grad_norm": 0.5778462075881274, "learning_rate": 1.9999373729162183e-05, "loss": 1.4093, "step": 165 }, { "epoch": 0.33739837398373984, "grad_norm": 0.5154746364254523, "learning_rate": 1.9999297884100524e-05, "loss": 1.2214, "step": 166 }, { "epoch": 0.3394308943089431, "grad_norm": 0.5418242108263562, "learning_rate": 1.999921770524384e-05, "loss": 1.2287, "step": 167 }, { "epoch": 0.34146341463414637, "grad_norm": 0.5847476266374432, "learning_rate": 1.9999133192626893e-05, "loss": 1.4373, "step": 168 }, { "epoch": 0.3434959349593496, "grad_norm": 0.586911035586581, "learning_rate": 1.9999044346286306e-05, "loss": 1.2608, "step": 169 }, { "epoch": 0.34552845528455284, "grad_norm": 0.5258198852133413, "learning_rate": 1.999895116626058e-05, "loss": 1.3411, "step": 170 }, { "epoch": 0.3475609756097561, "grad_norm": 0.6118600975287793, "learning_rate": 1.9998853652590108e-05, "loss": 1.2205, "step": 171 }, { "epoch": 0.34959349593495936, "grad_norm": 0.5048278110876353, "learning_rate": 1.9998751805317152e-05, "loss": 1.2134, "step": 172 }, { "epoch": 0.3516260162601626, "grad_norm": 0.5361967944671658, "learning_rate": 1.9998645624485854e-05, "loss": 1.1979, "step": 173 }, { "epoch": 0.35365853658536583, "grad_norm": 0.5529901115363403, "learning_rate": 1.999853511014223e-05, "loss": 1.1835, "step": 174 }, { "epoch": 0.3556910569105691, "grad_norm": 0.6102184016968556, "learning_rate": 1.999842026233418e-05, "loss": 1.4794, "step": 175 }, { "epoch": 0.35772357723577236, "grad_norm": 0.6000376990327756, "learning_rate": 1.999830108111148e-05, "loss": 1.356, "step": 176 }, { "epoch": 0.3597560975609756, "grad_norm": 0.5641935491601893, "learning_rate": 1.9998177566525788e-05, "loss": 1.2327, "step": 177 }, { "epoch": 0.3617886178861789, "grad_norm": 0.5820213889685879, "learning_rate": 1.999804971863063e-05, "loss": 1.1627, "step": 178 }, { "epoch": 0.3638211382113821, "grad_norm": 0.5789094111454401, "learning_rate": 1.9997917537481423e-05, "loss": 1.2597, "step": 179 }, { "epoch": 0.36585365853658536, "grad_norm": 0.5925739580980137, "learning_rate": 1.999778102313545e-05, "loss": 1.1533, "step": 180 }, { "epoch": 0.3678861788617886, "grad_norm": 0.5608178027883428, "learning_rate": 1.999764017565188e-05, "loss": 1.2241, "step": 181 }, { "epoch": 0.3699186991869919, "grad_norm": 0.6035987342394228, "learning_rate": 1.999749499509176e-05, "loss": 1.301, "step": 182 }, { "epoch": 0.3719512195121951, "grad_norm": 0.5628404819624865, "learning_rate": 1.9997345481518012e-05, "loss": 1.3171, "step": 183 }, { "epoch": 0.37398373983739835, "grad_norm": 0.6215429608817828, "learning_rate": 1.999719163499543e-05, "loss": 1.1118, "step": 184 }, { "epoch": 0.37601626016260165, "grad_norm": 0.6088293957795754, "learning_rate": 1.9997033455590703e-05, "loss": 1.2107, "step": 185 }, { "epoch": 0.3780487804878049, "grad_norm": 0.5657149852579519, "learning_rate": 1.999687094337238e-05, "loss": 1.1957, "step": 186 }, { "epoch": 0.3800813008130081, "grad_norm": 0.5871066375937471, "learning_rate": 1.99967040984109e-05, "loss": 1.2578, "step": 187 }, { "epoch": 0.3821138211382114, "grad_norm": 0.587720248187705, "learning_rate": 1.999653292077857e-05, "loss": 1.2805, "step": 188 }, { "epoch": 0.38414634146341464, "grad_norm": 0.5718468909660832, "learning_rate": 1.9996357410549585e-05, "loss": 1.1047, "step": 189 }, { "epoch": 0.3861788617886179, "grad_norm": 0.555992689982609, "learning_rate": 1.9996177567800012e-05, "loss": 1.1786, "step": 190 }, { "epoch": 0.3882113821138211, "grad_norm": 0.6408321910179687, "learning_rate": 1.9995993392607795e-05, "loss": 1.4004, "step": 191 }, { "epoch": 0.3902439024390244, "grad_norm": 0.5785957118274816, "learning_rate": 1.999580488505276e-05, "loss": 1.3679, "step": 192 }, { "epoch": 0.39227642276422764, "grad_norm": 0.5812267857394056, "learning_rate": 1.9995612045216602e-05, "loss": 1.2394, "step": 193 }, { "epoch": 0.3943089430894309, "grad_norm": 0.6504859365571795, "learning_rate": 1.9995414873182904e-05, "loss": 1.0817, "step": 194 }, { "epoch": 0.39634146341463417, "grad_norm": 0.6228843033176563, "learning_rate": 1.9995213369037124e-05, "loss": 1.3413, "step": 195 }, { "epoch": 0.3983739837398374, "grad_norm": 0.5784561498636087, "learning_rate": 1.9995007532866594e-05, "loss": 1.2489, "step": 196 }, { "epoch": 0.40040650406504064, "grad_norm": 0.7580973810417485, "learning_rate": 1.9994797364760528e-05, "loss": 1.4406, "step": 197 }, { "epoch": 0.4024390243902439, "grad_norm": 0.6034085204924278, "learning_rate": 1.9994582864810008e-05, "loss": 1.2351, "step": 198 }, { "epoch": 0.40447154471544716, "grad_norm": 0.5903327948473881, "learning_rate": 1.999436403310801e-05, "loss": 1.2987, "step": 199 }, { "epoch": 0.4065040650406504, "grad_norm": 0.5509718224062498, "learning_rate": 1.9994140869749366e-05, "loss": 1.1978, "step": 200 }, { "epoch": 0.40853658536585363, "grad_norm": 0.6437475577270788, "learning_rate": 1.9993913374830806e-05, "loss": 1.4371, "step": 201 }, { "epoch": 0.4105691056910569, "grad_norm": 0.5823108092522585, "learning_rate": 1.999368154845093e-05, "loss": 1.3512, "step": 202 }, { "epoch": 0.41260162601626016, "grad_norm": 0.703280478740523, "learning_rate": 1.9993445390710207e-05, "loss": 1.2723, "step": 203 }, { "epoch": 0.4146341463414634, "grad_norm": 0.584026207446822, "learning_rate": 1.9993204901710995e-05, "loss": 1.1542, "step": 204 }, { "epoch": 0.4166666666666667, "grad_norm": 0.6109212335010777, "learning_rate": 1.999296008155752e-05, "loss": 1.2073, "step": 205 }, { "epoch": 0.4186991869918699, "grad_norm": 0.601476418326994, "learning_rate": 1.9992710930355893e-05, "loss": 1.1988, "step": 206 }, { "epoch": 0.42073170731707316, "grad_norm": 0.6399992586681884, "learning_rate": 1.99924574482141e-05, "loss": 1.169, "step": 207 }, { "epoch": 0.42276422764227645, "grad_norm": 0.5472840954228271, "learning_rate": 1.9992199635241997e-05, "loss": 1.1829, "step": 208 }, { "epoch": 0.4247967479674797, "grad_norm": 0.6149740644563827, "learning_rate": 1.999193749155133e-05, "loss": 1.3061, "step": 209 }, { "epoch": 0.4268292682926829, "grad_norm": 0.6507625208374721, "learning_rate": 1.9991671017255712e-05, "loss": 1.2161, "step": 210 }, { "epoch": 0.42886178861788615, "grad_norm": 0.6552393793007679, "learning_rate": 1.9991400212470632e-05, "loss": 1.1706, "step": 211 }, { "epoch": 0.43089430894308944, "grad_norm": 0.6478314434988814, "learning_rate": 1.999112507731346e-05, "loss": 1.2408, "step": 212 }, { "epoch": 0.4329268292682927, "grad_norm": 0.6882511210450973, "learning_rate": 1.9990845611903447e-05, "loss": 1.3074, "step": 213 }, { "epoch": 0.4349593495934959, "grad_norm": 0.6712247726365308, "learning_rate": 1.999056181636171e-05, "loss": 1.2221, "step": 214 }, { "epoch": 0.4369918699186992, "grad_norm": 0.6225818065351404, "learning_rate": 1.9990273690811256e-05, "loss": 1.1912, "step": 215 }, { "epoch": 0.43902439024390244, "grad_norm": 0.6412314234579721, "learning_rate": 1.9989981235376956e-05, "loss": 1.2373, "step": 216 }, { "epoch": 0.4410569105691057, "grad_norm": 0.6608903378151971, "learning_rate": 1.9989684450185566e-05, "loss": 1.2317, "step": 217 }, { "epoch": 0.44308943089430897, "grad_norm": 0.6077628540327914, "learning_rate": 1.9989383335365713e-05, "loss": 1.2389, "step": 218 }, { "epoch": 0.4451219512195122, "grad_norm": 0.6445643051110905, "learning_rate": 1.9989077891047905e-05, "loss": 1.2015, "step": 219 }, { "epoch": 0.44715447154471544, "grad_norm": 0.6696203013363858, "learning_rate": 1.9988768117364526e-05, "loss": 1.0237, "step": 220 }, { "epoch": 0.4491869918699187, "grad_norm": 0.6857678118673014, "learning_rate": 1.998845401444983e-05, "loss": 1.1121, "step": 221 }, { "epoch": 0.45121951219512196, "grad_norm": 0.676249193447714, "learning_rate": 1.9988135582439955e-05, "loss": 1.3497, "step": 222 }, { "epoch": 0.4532520325203252, "grad_norm": 0.6588218055487259, "learning_rate": 1.9987812821472916e-05, "loss": 1.0491, "step": 223 }, { "epoch": 0.45528455284552843, "grad_norm": 0.6150055428034324, "learning_rate": 1.9987485731688595e-05, "loss": 1.1048, "step": 224 }, { "epoch": 0.4573170731707317, "grad_norm": 0.6697994458605332, "learning_rate": 1.998715431322876e-05, "loss": 1.2065, "step": 225 }, { "epoch": 0.45934959349593496, "grad_norm": 0.6279931468428332, "learning_rate": 1.9986818566237052e-05, "loss": 1.1322, "step": 226 }, { "epoch": 0.4613821138211382, "grad_norm": 0.6818965105332554, "learning_rate": 1.9986478490858983e-05, "loss": 1.1922, "step": 227 }, { "epoch": 0.4634146341463415, "grad_norm": 0.6233139491172669, "learning_rate": 1.998613408724195e-05, "loss": 1.0435, "step": 228 }, { "epoch": 0.4654471544715447, "grad_norm": 0.6002345686926185, "learning_rate": 1.9985785355535212e-05, "loss": 1.0931, "step": 229 }, { "epoch": 0.46747967479674796, "grad_norm": 0.6540728374079714, "learning_rate": 1.998543229588992e-05, "loss": 1.1357, "step": 230 }, { "epoch": 0.4695121951219512, "grad_norm": 0.633216634482548, "learning_rate": 1.9985074908459094e-05, "loss": 1.2458, "step": 231 }, { "epoch": 0.4715447154471545, "grad_norm": 0.6312740216296445, "learning_rate": 1.998471319339763e-05, "loss": 1.2057, "step": 232 }, { "epoch": 0.4735772357723577, "grad_norm": 0.7370884578177695, "learning_rate": 1.9984347150862293e-05, "loss": 1.1477, "step": 233 }, { "epoch": 0.47560975609756095, "grad_norm": 0.722757839095584, "learning_rate": 1.9983976781011735e-05, "loss": 1.0165, "step": 234 }, { "epoch": 0.47764227642276424, "grad_norm": 0.652657682373598, "learning_rate": 1.9983602084006476e-05, "loss": 1.2242, "step": 235 }, { "epoch": 0.4796747967479675, "grad_norm": 0.662434335579235, "learning_rate": 1.9983223060008908e-05, "loss": 1.1721, "step": 236 }, { "epoch": 0.4817073170731707, "grad_norm": 0.7746233407283346, "learning_rate": 1.9982839709183313e-05, "loss": 1.2792, "step": 237 }, { "epoch": 0.483739837398374, "grad_norm": 0.6403613852584241, "learning_rate": 1.9982452031695837e-05, "loss": 1.2405, "step": 238 }, { "epoch": 0.48577235772357724, "grad_norm": 0.7476615203324409, "learning_rate": 1.99820600277145e-05, "loss": 1.2114, "step": 239 }, { "epoch": 0.4878048780487805, "grad_norm": 0.7827606613997342, "learning_rate": 1.9981663697409203e-05, "loss": 1.3453, "step": 240 }, { "epoch": 0.4898373983739837, "grad_norm": 0.737197209949893, "learning_rate": 1.9981263040951715e-05, "loss": 1.1417, "step": 241 }, { "epoch": 0.491869918699187, "grad_norm": 0.6672504613890761, "learning_rate": 1.998085805851569e-05, "loss": 1.242, "step": 242 }, { "epoch": 0.49390243902439024, "grad_norm": 0.7888447840557603, "learning_rate": 1.998044875027665e-05, "loss": 1.0458, "step": 243 }, { "epoch": 0.4959349593495935, "grad_norm": 0.6982785976081, "learning_rate": 1.998003511641199e-05, "loss": 1.1137, "step": 244 }, { "epoch": 0.49796747967479676, "grad_norm": 0.6230319737138204, "learning_rate": 1.9979617157100987e-05, "loss": 1.1018, "step": 245 }, { "epoch": 0.5, "grad_norm": 0.630396792214378, "learning_rate": 1.9979194872524787e-05, "loss": 1.0491, "step": 246 }, { "epoch": 0.5020325203252033, "grad_norm": 0.6881572392364095, "learning_rate": 1.9978768262866414e-05, "loss": 1.0789, "step": 247 }, { "epoch": 0.5040650406504065, "grad_norm": 0.7160769076823188, "learning_rate": 1.997833732831076e-05, "loss": 1.0322, "step": 248 }, { "epoch": 0.5060975609756098, "grad_norm": 0.683251991412851, "learning_rate": 1.9977902069044603e-05, "loss": 1.1722, "step": 249 }, { "epoch": 0.508130081300813, "grad_norm": 0.734137901963406, "learning_rate": 1.9977462485256582e-05, "loss": 1.0926, "step": 250 }, { "epoch": 0.5101626016260162, "grad_norm": 0.7775526726428674, "learning_rate": 1.9977018577137223e-05, "loss": 1.3298, "step": 251 }, { "epoch": 0.5121951219512195, "grad_norm": 0.6461456175382727, "learning_rate": 1.9976570344878916e-05, "loss": 1.0067, "step": 252 }, { "epoch": 0.5142276422764228, "grad_norm": 0.6357310562029466, "learning_rate": 1.997611778867593e-05, "loss": 1.0114, "step": 253 }, { "epoch": 0.516260162601626, "grad_norm": 0.7173954109698393, "learning_rate": 1.9975660908724413e-05, "loss": 1.0988, "step": 254 }, { "epoch": 0.5182926829268293, "grad_norm": 0.6398154537011658, "learning_rate": 1.997519970522237e-05, "loss": 1.1044, "step": 255 }, { "epoch": 0.5203252032520326, "grad_norm": 0.6296013689116963, "learning_rate": 1.9974734178369702e-05, "loss": 0.9837, "step": 256 }, { "epoch": 0.5223577235772358, "grad_norm": 0.6418928926965674, "learning_rate": 1.997426432836817e-05, "loss": 1.1465, "step": 257 }, { "epoch": 0.524390243902439, "grad_norm": 0.7792633437003351, "learning_rate": 1.9973790155421406e-05, "loss": 1.1129, "step": 258 }, { "epoch": 0.5264227642276422, "grad_norm": 0.7042425720071207, "learning_rate": 1.997331165973493e-05, "loss": 1.2392, "step": 259 }, { "epoch": 0.5284552845528455, "grad_norm": 0.7333277098114007, "learning_rate": 1.997282884151612e-05, "loss": 1.1913, "step": 260 }, { "epoch": 0.5304878048780488, "grad_norm": 0.6863703682592543, "learning_rate": 1.9972341700974236e-05, "loss": 1.2162, "step": 261 }, { "epoch": 0.532520325203252, "grad_norm": 0.71605882945189, "learning_rate": 1.9971850238320408e-05, "loss": 1.2232, "step": 262 }, { "epoch": 0.5345528455284553, "grad_norm": 0.6917077564906295, "learning_rate": 1.997135445376765e-05, "loss": 0.9333, "step": 263 }, { "epoch": 0.5365853658536586, "grad_norm": 0.6496181773344989, "learning_rate": 1.9970854347530828e-05, "loss": 1.1762, "step": 264 }, { "epoch": 0.5386178861788617, "grad_norm": 0.7419651600944303, "learning_rate": 1.9970349919826696e-05, "loss": 1.2856, "step": 265 }, { "epoch": 0.540650406504065, "grad_norm": 0.7826806603406501, "learning_rate": 1.9969841170873883e-05, "loss": 1.1031, "step": 266 }, { "epoch": 0.5426829268292683, "grad_norm": 0.6929478640057137, "learning_rate": 1.9969328100892885e-05, "loss": 1.075, "step": 267 }, { "epoch": 0.5447154471544715, "grad_norm": 0.7066729630460149, "learning_rate": 1.9968810710106065e-05, "loss": 1.2669, "step": 268 }, { "epoch": 0.5467479674796748, "grad_norm": 0.7258330804554547, "learning_rate": 1.996828899873767e-05, "loss": 1.2179, "step": 269 }, { "epoch": 0.5487804878048781, "grad_norm": 0.6643787604833694, "learning_rate": 1.9967762967013814e-05, "loss": 1.1391, "step": 270 }, { "epoch": 0.5508130081300813, "grad_norm": 0.6753902249487399, "learning_rate": 1.9967232615162487e-05, "loss": 0.9714, "step": 271 }, { "epoch": 0.5528455284552846, "grad_norm": 0.6664964586548208, "learning_rate": 1.9966697943413548e-05, "loss": 0.9469, "step": 272 }, { "epoch": 0.5548780487804879, "grad_norm": 0.8146684056687993, "learning_rate": 1.9966158951998724e-05, "loss": 1.1138, "step": 273 }, { "epoch": 0.556910569105691, "grad_norm": 0.6704360299259119, "learning_rate": 1.9965615641151627e-05, "loss": 1.0823, "step": 274 }, { "epoch": 0.5589430894308943, "grad_norm": 0.7923153355302501, "learning_rate": 1.9965068011107728e-05, "loss": 1.1495, "step": 275 }, { "epoch": 0.5609756097560976, "grad_norm": 0.7082779585015615, "learning_rate": 1.9964516062104377e-05, "loss": 1.2264, "step": 276 }, { "epoch": 0.5630081300813008, "grad_norm": 0.6810548850238901, "learning_rate": 1.9963959794380788e-05, "loss": 1.0713, "step": 277 }, { "epoch": 0.5650406504065041, "grad_norm": 0.7962841068770551, "learning_rate": 1.9963399208178066e-05, "loss": 0.9788, "step": 278 }, { "epoch": 0.5670731707317073, "grad_norm": 0.6768397222950926, "learning_rate": 1.9962834303739162e-05, "loss": 0.9887, "step": 279 }, { "epoch": 0.5691056910569106, "grad_norm": 0.7253944492110169, "learning_rate": 1.996226508130892e-05, "loss": 0.9384, "step": 280 }, { "epoch": 0.5711382113821138, "grad_norm": 0.7441344630650087, "learning_rate": 1.9961691541134043e-05, "loss": 1.1031, "step": 281 }, { "epoch": 0.573170731707317, "grad_norm": 0.8050508766942358, "learning_rate": 1.996111368346311e-05, "loss": 1.266, "step": 282 }, { "epoch": 0.5752032520325203, "grad_norm": 0.6955642968618274, "learning_rate": 1.996053150854657e-05, "loss": 1.0933, "step": 283 }, { "epoch": 0.5772357723577236, "grad_norm": 0.8045191863960803, "learning_rate": 1.995994501663674e-05, "loss": 1.1552, "step": 284 }, { "epoch": 0.5792682926829268, "grad_norm": 0.8089317596135961, "learning_rate": 1.995935420798781e-05, "loss": 1.2533, "step": 285 }, { "epoch": 0.5813008130081301, "grad_norm": 0.7945719588441591, "learning_rate": 1.9958759082855855e-05, "loss": 1.1706, "step": 286 }, { "epoch": 0.5833333333333334, "grad_norm": 0.7814968034643627, "learning_rate": 1.9958159641498796e-05, "loss": 1.2515, "step": 287 }, { "epoch": 0.5853658536585366, "grad_norm": 0.6893564793597616, "learning_rate": 1.995755588417644e-05, "loss": 1.0151, "step": 288 }, { "epoch": 0.5873983739837398, "grad_norm": 0.7486695224712384, "learning_rate": 1.9956947811150464e-05, "loss": 1.1222, "step": 289 }, { "epoch": 0.5894308943089431, "grad_norm": 0.6633913412328797, "learning_rate": 1.9956335422684406e-05, "loss": 1.0056, "step": 290 }, { "epoch": 0.5914634146341463, "grad_norm": 0.8391787055728057, "learning_rate": 1.9955718719043693e-05, "loss": 1.0028, "step": 291 }, { "epoch": 0.5934959349593496, "grad_norm": 0.7456901867501453, "learning_rate": 1.99550977004956e-05, "loss": 1.2999, "step": 292 }, { "epoch": 0.5955284552845529, "grad_norm": 0.6471778037679606, "learning_rate": 1.995447236730929e-05, "loss": 1.1561, "step": 293 }, { "epoch": 0.5975609756097561, "grad_norm": 0.7781734988901607, "learning_rate": 1.9953842719755778e-05, "loss": 1.0572, "step": 294 }, { "epoch": 0.5995934959349594, "grad_norm": 0.7437538883499519, "learning_rate": 1.995320875810797e-05, "loss": 1.1162, "step": 295 }, { "epoch": 0.6016260162601627, "grad_norm": 0.7512987639710623, "learning_rate": 1.9952570482640628e-05, "loss": 1.2067, "step": 296 }, { "epoch": 0.6036585365853658, "grad_norm": 0.7387274430746452, "learning_rate": 1.9951927893630386e-05, "loss": 1.2129, "step": 297 }, { "epoch": 0.6056910569105691, "grad_norm": 0.6886161529518747, "learning_rate": 1.995128099135575e-05, "loss": 0.9931, "step": 298 }, { "epoch": 0.6077235772357723, "grad_norm": 0.9337498201171143, "learning_rate": 1.995062977609709e-05, "loss": 1.1844, "step": 299 }, { "epoch": 0.6097560975609756, "grad_norm": 0.7534108571809767, "learning_rate": 1.9949974248136655e-05, "loss": 0.9298, "step": 300 }, { "epoch": 0.6117886178861789, "grad_norm": 0.8006782589151503, "learning_rate": 1.9949314407758554e-05, "loss": 1.2175, "step": 301 }, { "epoch": 0.6138211382113821, "grad_norm": 0.7231237903621146, "learning_rate": 1.9948650255248768e-05, "loss": 1.1178, "step": 302 }, { "epoch": 0.6158536585365854, "grad_norm": 0.6538987690278102, "learning_rate": 1.9947981790895148e-05, "loss": 1.0094, "step": 303 }, { "epoch": 0.6178861788617886, "grad_norm": 0.7105197258184757, "learning_rate": 1.9947309014987414e-05, "loss": 1.0997, "step": 304 }, { "epoch": 0.6199186991869918, "grad_norm": 0.6595564188693971, "learning_rate": 1.9946631927817147e-05, "loss": 0.9506, "step": 305 }, { "epoch": 0.6219512195121951, "grad_norm": 0.7282023799928583, "learning_rate": 1.994595052967782e-05, "loss": 1.206, "step": 306 }, { "epoch": 0.6239837398373984, "grad_norm": 0.6105700212394325, "learning_rate": 1.9945264820864737e-05, "loss": 1.0077, "step": 307 }, { "epoch": 0.6260162601626016, "grad_norm": 0.7366229672955289, "learning_rate": 1.9944574801675106e-05, "loss": 1.2344, "step": 308 }, { "epoch": 0.6280487804878049, "grad_norm": 0.7550402489521852, "learning_rate": 1.9943880472407984e-05, "loss": 1.1313, "step": 309 }, { "epoch": 0.6300813008130082, "grad_norm": 0.7335721110721907, "learning_rate": 1.9943181833364296e-05, "loss": 1.0534, "step": 310 }, { "epoch": 0.6321138211382114, "grad_norm": 0.7862953496941809, "learning_rate": 1.9942478884846843e-05, "loss": 0.9994, "step": 311 }, { "epoch": 0.6341463414634146, "grad_norm": 0.6490177904509872, "learning_rate": 1.9941771627160287e-05, "loss": 1.001, "step": 312 }, { "epoch": 0.6361788617886179, "grad_norm": 0.7760110544141132, "learning_rate": 1.994106006061117e-05, "loss": 1.2654, "step": 313 }, { "epoch": 0.6382113821138211, "grad_norm": 0.8261555241390679, "learning_rate": 1.994034418550788e-05, "loss": 1.2691, "step": 314 }, { "epoch": 0.6402439024390244, "grad_norm": 0.8242461328923543, "learning_rate": 1.993962400216069e-05, "loss": 0.9361, "step": 315 }, { "epoch": 0.6422764227642277, "grad_norm": 0.7784050224657479, "learning_rate": 1.9938899510881732e-05, "loss": 1.3881, "step": 316 }, { "epoch": 0.6443089430894309, "grad_norm": 0.7655064453165692, "learning_rate": 1.9938170711985012e-05, "loss": 0.969, "step": 317 }, { "epoch": 0.6463414634146342, "grad_norm": 0.6093603866187096, "learning_rate": 1.9937437605786395e-05, "loss": 1.1047, "step": 318 }, { "epoch": 0.6483739837398373, "grad_norm": 0.695675927345526, "learning_rate": 1.993670019260362e-05, "loss": 1.0072, "step": 319 }, { "epoch": 0.6504065040650406, "grad_norm": 0.8308660510744199, "learning_rate": 1.9935958472756283e-05, "loss": 1.1061, "step": 320 }, { "epoch": 0.6524390243902439, "grad_norm": 0.7359503826354641, "learning_rate": 1.9935212446565858e-05, "loss": 1.1301, "step": 321 }, { "epoch": 0.6544715447154471, "grad_norm": 0.7357907459651524, "learning_rate": 1.993446211435568e-05, "loss": 1.1282, "step": 322 }, { "epoch": 0.6565040650406504, "grad_norm": 0.6977548126465809, "learning_rate": 1.993370747645095e-05, "loss": 0.9635, "step": 323 }, { "epoch": 0.6585365853658537, "grad_norm": 0.7759267874799558, "learning_rate": 1.993294853317873e-05, "loss": 1.0304, "step": 324 }, { "epoch": 0.6605691056910569, "grad_norm": 0.7046721057726196, "learning_rate": 1.9932185284867962e-05, "loss": 0.9107, "step": 325 }, { "epoch": 0.6626016260162602, "grad_norm": 0.8145655286668039, "learning_rate": 1.993141773184944e-05, "loss": 1.0994, "step": 326 }, { "epoch": 0.6646341463414634, "grad_norm": 0.7115204931225575, "learning_rate": 1.9930645874455834e-05, "loss": 1.0652, "step": 327 }, { "epoch": 0.6666666666666666, "grad_norm": 0.8899492423125045, "learning_rate": 1.9929869713021668e-05, "loss": 1.1785, "step": 328 }, { "epoch": 0.6686991869918699, "grad_norm": 0.7879185381160279, "learning_rate": 1.9929089247883344e-05, "loss": 1.1128, "step": 329 }, { "epoch": 0.6707317073170732, "grad_norm": 0.8769844369992154, "learning_rate": 1.9928304479379118e-05, "loss": 1.1305, "step": 330 }, { "epoch": 0.6727642276422764, "grad_norm": 0.8687446983119437, "learning_rate": 1.992751540784912e-05, "loss": 1.2483, "step": 331 }, { "epoch": 0.6747967479674797, "grad_norm": 0.8364485747510623, "learning_rate": 1.9926722033635343e-05, "loss": 1.2804, "step": 332 }, { "epoch": 0.676829268292683, "grad_norm": 0.7524840668857417, "learning_rate": 1.992592435708164e-05, "loss": 1.0503, "step": 333 }, { "epoch": 0.6788617886178862, "grad_norm": 0.8759455196518872, "learning_rate": 1.9925122378533735e-05, "loss": 1.1792, "step": 334 }, { "epoch": 0.6808943089430894, "grad_norm": 0.83774530232434, "learning_rate": 1.992431609833921e-05, "loss": 1.3306, "step": 335 }, { "epoch": 0.6829268292682927, "grad_norm": 0.8836861103382394, "learning_rate": 1.9923505516847514e-05, "loss": 1.2272, "step": 336 }, { "epoch": 0.6849593495934959, "grad_norm": 0.8209543849265472, "learning_rate": 1.9922690634409965e-05, "loss": 1.1447, "step": 337 }, { "epoch": 0.6869918699186992, "grad_norm": 0.9246060744382218, "learning_rate": 1.992187145137974e-05, "loss": 1.1337, "step": 338 }, { "epoch": 0.6890243902439024, "grad_norm": 0.72313512655042, "learning_rate": 1.9921047968111882e-05, "loss": 1.2611, "step": 339 }, { "epoch": 0.6910569105691057, "grad_norm": 0.7129447553079935, "learning_rate": 1.9920220184963296e-05, "loss": 1.0408, "step": 340 }, { "epoch": 0.693089430894309, "grad_norm": 0.8479948802045845, "learning_rate": 1.991938810229275e-05, "loss": 1.255, "step": 341 }, { "epoch": 0.6951219512195121, "grad_norm": 0.7712579457815385, "learning_rate": 1.9918551720460874e-05, "loss": 0.8778, "step": 342 }, { "epoch": 0.6971544715447154, "grad_norm": 0.8713207868002114, "learning_rate": 1.991771103983017e-05, "loss": 1.0737, "step": 343 }, { "epoch": 0.6991869918699187, "grad_norm": 0.7849843964311328, "learning_rate": 1.9916866060764994e-05, "loss": 1.1461, "step": 344 }, { "epoch": 0.7012195121951219, "grad_norm": 0.7893248208932844, "learning_rate": 1.991601678363157e-05, "loss": 1.0177, "step": 345 }, { "epoch": 0.7032520325203252, "grad_norm": 0.8461423402217931, "learning_rate": 1.9915163208797986e-05, "loss": 1.1233, "step": 346 }, { "epoch": 0.7052845528455285, "grad_norm": 0.6527702137350551, "learning_rate": 1.9914305336634187e-05, "loss": 1.0204, "step": 347 }, { "epoch": 0.7073170731707317, "grad_norm": 0.7466305361439416, "learning_rate": 1.991344316751198e-05, "loss": 1.0347, "step": 348 }, { "epoch": 0.709349593495935, "grad_norm": 0.7403559161275431, "learning_rate": 1.991257670180504e-05, "loss": 0.9934, "step": 349 }, { "epoch": 0.7113821138211383, "grad_norm": 0.8757238282481509, "learning_rate": 1.9911705939888905e-05, "loss": 0.929, "step": 350 }, { "epoch": 0.7134146341463414, "grad_norm": 0.8464033800891151, "learning_rate": 1.9910830882140967e-05, "loss": 0.9744, "step": 351 }, { "epoch": 0.7154471544715447, "grad_norm": 0.9041960406003268, "learning_rate": 1.9909951528940485e-05, "loss": 1.0844, "step": 352 }, { "epoch": 0.717479674796748, "grad_norm": 0.7718255643186065, "learning_rate": 1.9909067880668583e-05, "loss": 1.0546, "step": 353 }, { "epoch": 0.7195121951219512, "grad_norm": 0.8559886670637414, "learning_rate": 1.9908179937708244e-05, "loss": 1.1614, "step": 354 }, { "epoch": 0.7215447154471545, "grad_norm": 0.8793626877908183, "learning_rate": 1.990728770044431e-05, "loss": 1.1303, "step": 355 }, { "epoch": 0.7235772357723578, "grad_norm": 0.709460532953605, "learning_rate": 1.990639116926348e-05, "loss": 1.0501, "step": 356 }, { "epoch": 0.725609756097561, "grad_norm": 0.8398701083774753, "learning_rate": 1.9905490344554325e-05, "loss": 0.9864, "step": 357 }, { "epoch": 0.7276422764227642, "grad_norm": 0.8683835809740145, "learning_rate": 1.990458522670727e-05, "loss": 1.0017, "step": 358 }, { "epoch": 0.7296747967479674, "grad_norm": 0.9093979910852322, "learning_rate": 1.99036758161146e-05, "loss": 1.1866, "step": 359 }, { "epoch": 0.7317073170731707, "grad_norm": 0.7862220594658448, "learning_rate": 1.9902762113170467e-05, "loss": 1.0232, "step": 360 }, { "epoch": 0.733739837398374, "grad_norm": 0.8512543420004076, "learning_rate": 1.9901844118270875e-05, "loss": 1.1265, "step": 361 }, { "epoch": 0.7357723577235772, "grad_norm": 0.9359099639950121, "learning_rate": 1.9900921831813695e-05, "loss": 0.9303, "step": 362 }, { "epoch": 0.7378048780487805, "grad_norm": 0.7268312961357214, "learning_rate": 1.9899995254198655e-05, "loss": 1.0529, "step": 363 }, { "epoch": 0.7398373983739838, "grad_norm": 0.8039946599556934, "learning_rate": 1.989906438582734e-05, "loss": 0.965, "step": 364 }, { "epoch": 0.741869918699187, "grad_norm": 0.8444207131888133, "learning_rate": 1.98981292271032e-05, "loss": 1.2866, "step": 365 }, { "epoch": 0.7439024390243902, "grad_norm": 0.8542704797481094, "learning_rate": 1.9897189778431542e-05, "loss": 1.2246, "step": 366 }, { "epoch": 0.7459349593495935, "grad_norm": 0.76548113628278, "learning_rate": 1.989624604021953e-05, "loss": 1.1666, "step": 367 }, { "epoch": 0.7479674796747967, "grad_norm": 0.8261987698490619, "learning_rate": 1.9895298012876192e-05, "loss": 0.9967, "step": 368 }, { "epoch": 0.75, "grad_norm": 0.8763250741843557, "learning_rate": 1.9894345696812413e-05, "loss": 1.1903, "step": 369 }, { "epoch": 0.7520325203252033, "grad_norm": 0.8551737355763933, "learning_rate": 1.9893389092440932e-05, "loss": 1.0636, "step": 370 }, { "epoch": 0.7540650406504065, "grad_norm": 0.8687165307764337, "learning_rate": 1.989242820017636e-05, "loss": 1.1209, "step": 371 }, { "epoch": 0.7560975609756098, "grad_norm": 0.8986269816037802, "learning_rate": 1.9891463020435144e-05, "loss": 1.0994, "step": 372 }, { "epoch": 0.758130081300813, "grad_norm": 0.9128601794164192, "learning_rate": 1.989049355363561e-05, "loss": 0.9773, "step": 373 }, { "epoch": 0.7601626016260162, "grad_norm": 0.8859899190733247, "learning_rate": 1.988951980019794e-05, "loss": 1.1162, "step": 374 }, { "epoch": 0.7621951219512195, "grad_norm": 0.7911240823451815, "learning_rate": 1.9888541760544154e-05, "loss": 1.1656, "step": 375 }, { "epoch": 0.7642276422764228, "grad_norm": 0.8117680606213679, "learning_rate": 1.9887559435098162e-05, "loss": 1.0229, "step": 376 }, { "epoch": 0.766260162601626, "grad_norm": 0.9832847811984806, "learning_rate": 1.9886572824285698e-05, "loss": 1.2616, "step": 377 }, { "epoch": 0.7682926829268293, "grad_norm": 0.8707182915756151, "learning_rate": 1.988558192853438e-05, "loss": 1.0617, "step": 378 }, { "epoch": 0.7703252032520326, "grad_norm": 0.730639019068089, "learning_rate": 1.9884586748273664e-05, "loss": 0.9824, "step": 379 }, { "epoch": 0.7723577235772358, "grad_norm": 0.8491160020626655, "learning_rate": 1.9883587283934875e-05, "loss": 1.1243, "step": 380 }, { "epoch": 0.774390243902439, "grad_norm": 0.8758504389440154, "learning_rate": 1.988258353595119e-05, "loss": 1.1837, "step": 381 }, { "epoch": 0.7764227642276422, "grad_norm": 0.8984507124702282, "learning_rate": 1.988157550475764e-05, "loss": 1.0556, "step": 382 }, { "epoch": 0.7784552845528455, "grad_norm": 0.865512629824748, "learning_rate": 1.9880563190791125e-05, "loss": 1.0094, "step": 383 }, { "epoch": 0.7804878048780488, "grad_norm": 0.8719091025330833, "learning_rate": 1.9879546594490383e-05, "loss": 1.0989, "step": 384 }, { "epoch": 0.782520325203252, "grad_norm": 0.8443421671113395, "learning_rate": 1.987852571629602e-05, "loss": 1.0766, "step": 385 }, { "epoch": 0.7845528455284553, "grad_norm": 0.7801184751508234, "learning_rate": 1.9877500556650494e-05, "loss": 1.0505, "step": 386 }, { "epoch": 0.7865853658536586, "grad_norm": 0.8947502640696742, "learning_rate": 1.9876471115998122e-05, "loss": 1.3102, "step": 387 }, { "epoch": 0.7886178861788617, "grad_norm": 0.7259560312278743, "learning_rate": 1.987543739478507e-05, "loss": 1.0932, "step": 388 }, { "epoch": 0.790650406504065, "grad_norm": 0.812712031940616, "learning_rate": 1.9874399393459365e-05, "loss": 1.0595, "step": 389 }, { "epoch": 0.7926829268292683, "grad_norm": 0.8574089878289958, "learning_rate": 1.987335711247089e-05, "loss": 0.9476, "step": 390 }, { "epoch": 0.7947154471544715, "grad_norm": 0.9079669578420158, "learning_rate": 1.9872310552271374e-05, "loss": 0.9934, "step": 391 }, { "epoch": 0.7967479674796748, "grad_norm": 0.9418588852393975, "learning_rate": 1.987125971331441e-05, "loss": 1.1037, "step": 392 }, { "epoch": 0.7987804878048781, "grad_norm": 0.8588984811467736, "learning_rate": 1.987020459605544e-05, "loss": 0.989, "step": 393 }, { "epoch": 0.8008130081300813, "grad_norm": 0.8378124737523079, "learning_rate": 1.9869145200951766e-05, "loss": 0.9201, "step": 394 }, { "epoch": 0.8028455284552846, "grad_norm": 0.9979608204270138, "learning_rate": 1.986808152846254e-05, "loss": 1.0191, "step": 395 }, { "epoch": 0.8048780487804879, "grad_norm": 0.89781681406106, "learning_rate": 1.9867013579048765e-05, "loss": 1.2184, "step": 396 }, { "epoch": 0.806910569105691, "grad_norm": 0.7886510521239992, "learning_rate": 1.9865941353173302e-05, "loss": 1.0385, "step": 397 }, { "epoch": 0.8089430894308943, "grad_norm": 0.8327183691819595, "learning_rate": 1.9864864851300863e-05, "loss": 1.0465, "step": 398 }, { "epoch": 0.8109756097560976, "grad_norm": 0.7853127856207369, "learning_rate": 1.9863784073898018e-05, "loss": 1.155, "step": 399 }, { "epoch": 0.8130081300813008, "grad_norm": 0.8409050861873923, "learning_rate": 1.9862699021433186e-05, "loss": 1.0399, "step": 400 }, { "epoch": 0.8150406504065041, "grad_norm": 0.7424180594441359, "learning_rate": 1.9861609694376635e-05, "loss": 1.1304, "step": 401 }, { "epoch": 0.8170731707317073, "grad_norm": 0.8020892439617656, "learning_rate": 1.9860516093200493e-05, "loss": 1.0076, "step": 402 }, { "epoch": 0.8191056910569106, "grad_norm": 0.8062601762244481, "learning_rate": 1.9859418218378742e-05, "loss": 1.1278, "step": 403 }, { "epoch": 0.8211382113821138, "grad_norm": 0.8235535650407905, "learning_rate": 1.9858316070387208e-05, "loss": 1.0584, "step": 404 }, { "epoch": 0.823170731707317, "grad_norm": 0.8286517930336054, "learning_rate": 1.9857209649703573e-05, "loss": 1.1391, "step": 405 }, { "epoch": 0.8252032520325203, "grad_norm": 0.8898545782379115, "learning_rate": 1.985609895680737e-05, "loss": 0.9472, "step": 406 }, { "epoch": 0.8272357723577236, "grad_norm": 0.9288637237327725, "learning_rate": 1.9854983992179987e-05, "loss": 1.0112, "step": 407 }, { "epoch": 0.8292682926829268, "grad_norm": 0.8750047969552921, "learning_rate": 1.9853864756304654e-05, "loss": 0.9476, "step": 408 }, { "epoch": 0.8313008130081301, "grad_norm": 0.9127082380849462, "learning_rate": 1.9852741249666464e-05, "loss": 1.3042, "step": 409 }, { "epoch": 0.8333333333333334, "grad_norm": 0.8962346633719827, "learning_rate": 1.985161347275236e-05, "loss": 1.0546, "step": 410 }, { "epoch": 0.8353658536585366, "grad_norm": 0.8380458894075802, "learning_rate": 1.9850481426051125e-05, "loss": 1.0214, "step": 411 }, { "epoch": 0.8373983739837398, "grad_norm": 0.8883777730504038, "learning_rate": 1.9849345110053405e-05, "loss": 0.8893, "step": 412 }, { "epoch": 0.8394308943089431, "grad_norm": 0.8935903053211901, "learning_rate": 1.984820452525168e-05, "loss": 1.0352, "step": 413 }, { "epoch": 0.8414634146341463, "grad_norm": 0.8128931099950174, "learning_rate": 1.9847059672140305e-05, "loss": 1.0146, "step": 414 }, { "epoch": 0.8434959349593496, "grad_norm": 0.7480300679619891, "learning_rate": 1.9845910551215462e-05, "loss": 1.0525, "step": 415 }, { "epoch": 0.8455284552845529, "grad_norm": 0.8805237760347523, "learning_rate": 1.984475716297519e-05, "loss": 1.0843, "step": 416 }, { "epoch": 0.8475609756097561, "grad_norm": 0.7928507222795159, "learning_rate": 1.9843599507919385e-05, "loss": 1.1023, "step": 417 }, { "epoch": 0.8495934959349594, "grad_norm": 0.771311805449204, "learning_rate": 1.9842437586549783e-05, "loss": 0.8653, "step": 418 }, { "epoch": 0.8516260162601627, "grad_norm": 0.8560345845917903, "learning_rate": 1.9841271399369973e-05, "loss": 1.2467, "step": 419 }, { "epoch": 0.8536585365853658, "grad_norm": 0.9336520493233321, "learning_rate": 1.984010094688539e-05, "loss": 1.1441, "step": 420 }, { "epoch": 0.8556910569105691, "grad_norm": 0.8587387628931449, "learning_rate": 1.9838926229603322e-05, "loss": 1.1753, "step": 421 }, { "epoch": 0.8577235772357723, "grad_norm": 0.8970493549861975, "learning_rate": 1.9837747248032904e-05, "loss": 1.2841, "step": 422 }, { "epoch": 0.8597560975609756, "grad_norm": 0.7943026017030438, "learning_rate": 1.9836564002685115e-05, "loss": 1.017, "step": 423 }, { "epoch": 0.8617886178861789, "grad_norm": 0.8254146263379378, "learning_rate": 1.9835376494072788e-05, "loss": 0.9192, "step": 424 }, { "epoch": 0.8638211382113821, "grad_norm": 0.949237257438604, "learning_rate": 1.98341847227106e-05, "loss": 1.2524, "step": 425 }, { "epoch": 0.8658536585365854, "grad_norm": 0.9055662476130731, "learning_rate": 1.9832988689115078e-05, "loss": 0.9752, "step": 426 }, { "epoch": 0.8678861788617886, "grad_norm": 0.8815100211159871, "learning_rate": 1.9831788393804593e-05, "loss": 1.1699, "step": 427 }, { "epoch": 0.8699186991869918, "grad_norm": 0.8400228167321123, "learning_rate": 1.9830583837299363e-05, "loss": 1.0792, "step": 428 }, { "epoch": 0.8719512195121951, "grad_norm": 0.9318773526574476, "learning_rate": 1.982937502012146e-05, "loss": 1.1205, "step": 429 }, { "epoch": 0.8739837398373984, "grad_norm": 0.842683442661975, "learning_rate": 1.9828161942794793e-05, "loss": 1.1944, "step": 430 }, { "epoch": 0.8760162601626016, "grad_norm": 0.7792062442706137, "learning_rate": 1.9826944605845125e-05, "loss": 0.9745, "step": 431 }, { "epoch": 0.8780487804878049, "grad_norm": 0.9589991861871685, "learning_rate": 1.9825723009800058e-05, "loss": 1.2528, "step": 432 }, { "epoch": 0.8800813008130082, "grad_norm": 0.8484464229328488, "learning_rate": 1.9824497155189048e-05, "loss": 1.0932, "step": 433 }, { "epoch": 0.8821138211382114, "grad_norm": 0.7774659504672015, "learning_rate": 1.9823267042543392e-05, "loss": 1.2028, "step": 434 }, { "epoch": 0.8841463414634146, "grad_norm": 1.0179275123489693, "learning_rate": 1.9822032672396226e-05, "loss": 1.4034, "step": 435 }, { "epoch": 0.8861788617886179, "grad_norm": 0.8394780231481216, "learning_rate": 1.9820794045282553e-05, "loss": 0.902, "step": 436 }, { "epoch": 0.8882113821138211, "grad_norm": 0.8515218134925033, "learning_rate": 1.981955116173919e-05, "loss": 1.2325, "step": 437 }, { "epoch": 0.8902439024390244, "grad_norm": 0.9287920221146229, "learning_rate": 1.9818304022304824e-05, "loss": 1.0282, "step": 438 }, { "epoch": 0.8922764227642277, "grad_norm": 0.834638455935727, "learning_rate": 1.981705262751998e-05, "loss": 1.1342, "step": 439 }, { "epoch": 0.8943089430894309, "grad_norm": 0.9983290397384613, "learning_rate": 1.9815796977927015e-05, "loss": 1.077, "step": 440 }, { "epoch": 0.8963414634146342, "grad_norm": 0.8364522708708451, "learning_rate": 1.981453707407015e-05, "loss": 1.2097, "step": 441 }, { "epoch": 0.8983739837398373, "grad_norm": 0.8993441329241969, "learning_rate": 1.9813272916495435e-05, "loss": 1.1354, "step": 442 }, { "epoch": 0.9004065040650406, "grad_norm": 0.8917915652079444, "learning_rate": 1.981200450575077e-05, "loss": 1.0096, "step": 443 }, { "epoch": 0.9024390243902439, "grad_norm": 0.8230494563099897, "learning_rate": 1.9810731842385892e-05, "loss": 1.0393, "step": 444 }, { "epoch": 0.9044715447154471, "grad_norm": 0.8217744498279704, "learning_rate": 1.9809454926952395e-05, "loss": 1.0265, "step": 445 }, { "epoch": 0.9065040650406504, "grad_norm": 0.8647768600783248, "learning_rate": 1.98081737600037e-05, "loss": 0.9744, "step": 446 }, { "epoch": 0.9085365853658537, "grad_norm": 0.9312158964921499, "learning_rate": 1.9806888342095078e-05, "loss": 1.1097, "step": 447 }, { "epoch": 0.9105691056910569, "grad_norm": 0.9848245150123199, "learning_rate": 1.9805598673783644e-05, "loss": 1.033, "step": 448 }, { "epoch": 0.9126016260162602, "grad_norm": 0.8237848135264456, "learning_rate": 1.9804304755628352e-05, "loss": 1.0903, "step": 449 }, { "epoch": 0.9146341463414634, "grad_norm": 0.8801877589335505, "learning_rate": 1.9803006588190002e-05, "loss": 1.1874, "step": 450 }, { "epoch": 0.9166666666666666, "grad_norm": 0.8764161996402368, "learning_rate": 1.9801704172031228e-05, "loss": 1.1432, "step": 451 }, { "epoch": 0.9186991869918699, "grad_norm": 0.8614998148744878, "learning_rate": 1.980039750771651e-05, "loss": 1.0555, "step": 452 }, { "epoch": 0.9207317073170732, "grad_norm": 0.8000409333103937, "learning_rate": 1.9799086595812174e-05, "loss": 1.1637, "step": 453 }, { "epoch": 0.9227642276422764, "grad_norm": 0.813661251580359, "learning_rate": 1.979777143688638e-05, "loss": 0.9032, "step": 454 }, { "epoch": 0.9247967479674797, "grad_norm": 0.7661037712971139, "learning_rate": 1.979645203150913e-05, "loss": 0.9553, "step": 455 }, { "epoch": 0.926829268292683, "grad_norm": 0.8798543374146954, "learning_rate": 1.9795128380252263e-05, "loss": 1.0171, "step": 456 }, { "epoch": 0.9288617886178862, "grad_norm": 0.9764498879223148, "learning_rate": 1.979380048368947e-05, "loss": 1.1928, "step": 457 }, { "epoch": 0.9308943089430894, "grad_norm": 1.1097447292506373, "learning_rate": 1.9792468342396277e-05, "loss": 1.2134, "step": 458 }, { "epoch": 0.9329268292682927, "grad_norm": 0.9711741882637649, "learning_rate": 1.9791131956950037e-05, "loss": 1.1271, "step": 459 }, { "epoch": 0.9349593495934959, "grad_norm": 0.8801632107258242, "learning_rate": 1.978979132792996e-05, "loss": 0.9238, "step": 460 }, { "epoch": 0.9369918699186992, "grad_norm": 0.9045546101227948, "learning_rate": 1.9788446455917087e-05, "loss": 0.9944, "step": 461 }, { "epoch": 0.9390243902439024, "grad_norm": 0.8207212373128288, "learning_rate": 1.9787097341494296e-05, "loss": 1.007, "step": 462 }, { "epoch": 0.9410569105691057, "grad_norm": 0.8647651589403151, "learning_rate": 1.978574398524631e-05, "loss": 1.0948, "step": 463 }, { "epoch": 0.943089430894309, "grad_norm": 0.8573886973280551, "learning_rate": 1.9784386387759684e-05, "loss": 1.2171, "step": 464 }, { "epoch": 0.9451219512195121, "grad_norm": 0.9242064557261793, "learning_rate": 1.9783024549622816e-05, "loss": 0.9022, "step": 465 }, { "epoch": 0.9471544715447154, "grad_norm": 1.0226618825537657, "learning_rate": 1.978165847142594e-05, "loss": 1.1819, "step": 466 }, { "epoch": 0.9491869918699187, "grad_norm": 0.7757185672443746, "learning_rate": 1.978028815376113e-05, "loss": 0.9813, "step": 467 }, { "epoch": 0.9512195121951219, "grad_norm": 0.8994770307054246, "learning_rate": 1.977891359722229e-05, "loss": 0.9417, "step": 468 }, { "epoch": 0.9532520325203252, "grad_norm": 0.8093898820250115, "learning_rate": 1.977753480240517e-05, "loss": 0.9101, "step": 469 }, { "epoch": 0.9552845528455285, "grad_norm": 0.8740581782309595, "learning_rate": 1.9776151769907353e-05, "loss": 1.2183, "step": 470 }, { "epoch": 0.9573170731707317, "grad_norm": 0.937506828643832, "learning_rate": 1.977476450032826e-05, "loss": 1.0675, "step": 471 }, { "epoch": 0.959349593495935, "grad_norm": 0.8522982483891116, "learning_rate": 1.9773372994269147e-05, "loss": 1.2072, "step": 472 }, { "epoch": 0.9613821138211383, "grad_norm": 0.9439955765636268, "learning_rate": 1.9771977252333104e-05, "loss": 1.0035, "step": 473 }, { "epoch": 0.9634146341463414, "grad_norm": 0.7998856634468354, "learning_rate": 1.9770577275125063e-05, "loss": 1.0851, "step": 474 }, { "epoch": 0.9654471544715447, "grad_norm": 0.9571030002483849, "learning_rate": 1.9769173063251782e-05, "loss": 1.1612, "step": 475 }, { "epoch": 0.967479674796748, "grad_norm": 0.8908700702869584, "learning_rate": 1.976776461732187e-05, "loss": 1.1408, "step": 476 }, { "epoch": 0.9695121951219512, "grad_norm": 0.8728997982294588, "learning_rate": 1.9766351937945753e-05, "loss": 1.0041, "step": 477 }, { "epoch": 0.9715447154471545, "grad_norm": 0.870660804424435, "learning_rate": 1.9764935025735704e-05, "loss": 1.1166, "step": 478 }, { "epoch": 0.9735772357723578, "grad_norm": 0.955940340885727, "learning_rate": 1.976351388130583e-05, "loss": 1.1196, "step": 479 }, { "epoch": 0.975609756097561, "grad_norm": 0.7866372824902711, "learning_rate": 1.976208850527206e-05, "loss": 0.9699, "step": 480 }, { "epoch": 0.9776422764227642, "grad_norm": 0.99353371788088, "learning_rate": 1.976065889825217e-05, "loss": 1.0134, "step": 481 }, { "epoch": 0.9796747967479674, "grad_norm": 1.0177510139022083, "learning_rate": 1.9759225060865772e-05, "loss": 1.2094, "step": 482 }, { "epoch": 0.9817073170731707, "grad_norm": 0.9439025257140534, "learning_rate": 1.97577869937343e-05, "loss": 1.4114, "step": 483 }, { "epoch": 0.983739837398374, "grad_norm": 0.9297009457711639, "learning_rate": 1.9756344697481027e-05, "loss": 1.0473, "step": 484 }, { "epoch": 0.9857723577235772, "grad_norm": 0.9311138528482534, "learning_rate": 1.9754898172731056e-05, "loss": 0.9335, "step": 485 }, { "epoch": 0.9878048780487805, "grad_norm": 0.8750141287257764, "learning_rate": 1.975344742011133e-05, "loss": 1.0733, "step": 486 }, { "epoch": 0.9898373983739838, "grad_norm": 0.8393224270687004, "learning_rate": 1.975199244025062e-05, "loss": 1.0896, "step": 487 }, { "epoch": 0.991869918699187, "grad_norm": 0.9165402838801273, "learning_rate": 1.975053323377952e-05, "loss": 1.108, "step": 488 }, { "epoch": 0.9939024390243902, "grad_norm": 0.8341401767477676, "learning_rate": 1.9749069801330477e-05, "loss": 1.0481, "step": 489 }, { "epoch": 0.9959349593495935, "grad_norm": 0.9763340591101795, "learning_rate": 1.9747602143537752e-05, "loss": 1.1283, "step": 490 }, { "epoch": 0.9979674796747967, "grad_norm": 0.9906176820498767, "learning_rate": 1.9746130261037437e-05, "loss": 0.8376, "step": 491 }, { "epoch": 1.0, "grad_norm": 1.010416819082582, "learning_rate": 1.9744654154467468e-05, "loss": 1.1198, "step": 492 }, { "epoch": 1.0, "eval_loss": 1.0683757066726685, "eval_runtime": 133.6246, "eval_samples_per_second": 7.364, "eval_steps_per_second": 0.92, "step": 492 }, { "epoch": 1.0020325203252032, "grad_norm": 0.8656439459456532, "learning_rate": 1.9743173824467597e-05, "loss": 0.9182, "step": 493 }, { "epoch": 1.0040650406504066, "grad_norm": 0.945814123346113, "learning_rate": 1.9741689271679422e-05, "loss": 0.9677, "step": 494 }, { "epoch": 1.0060975609756098, "grad_norm": 0.9270700282516309, "learning_rate": 1.974020049674636e-05, "loss": 0.9522, "step": 495 }, { "epoch": 1.008130081300813, "grad_norm": 0.8565476969124097, "learning_rate": 1.9738707500313655e-05, "loss": 1.0552, "step": 496 }, { "epoch": 1.0101626016260163, "grad_norm": 0.9541026156007423, "learning_rate": 1.9737210283028397e-05, "loss": 1.0402, "step": 497 }, { "epoch": 1.0121951219512195, "grad_norm": 0.9001216650614674, "learning_rate": 1.9735708845539486e-05, "loss": 0.9542, "step": 498 }, { "epoch": 1.0142276422764227, "grad_norm": 0.918499826223804, "learning_rate": 1.973420318849766e-05, "loss": 0.922, "step": 499 }, { "epoch": 1.016260162601626, "grad_norm": 0.9937922900277678, "learning_rate": 1.9732693312555492e-05, "loss": 1.0634, "step": 500 }, { "epoch": 1.0182926829268293, "grad_norm": 0.9342471787637066, "learning_rate": 1.9731179218367375e-05, "loss": 1.0121, "step": 501 }, { "epoch": 1.0203252032520325, "grad_norm": 0.9408831275042376, "learning_rate": 1.972966090658953e-05, "loss": 1.1629, "step": 502 }, { "epoch": 1.0223577235772359, "grad_norm": 1.0218109059500948, "learning_rate": 1.9728138377880008e-05, "loss": 1.0242, "step": 503 }, { "epoch": 1.024390243902439, "grad_norm": 0.9215693945728668, "learning_rate": 1.9726611632898693e-05, "loss": 1.0532, "step": 504 }, { "epoch": 1.0264227642276422, "grad_norm": 0.8962329735165209, "learning_rate": 1.9725080672307286e-05, "loss": 0.998, "step": 505 }, { "epoch": 1.0284552845528456, "grad_norm": 0.8966170734296421, "learning_rate": 1.9723545496769324e-05, "loss": 0.9778, "step": 506 }, { "epoch": 1.0304878048780488, "grad_norm": 0.9856619528664992, "learning_rate": 1.9722006106950166e-05, "loss": 0.9943, "step": 507 }, { "epoch": 1.032520325203252, "grad_norm": 0.9245060179608685, "learning_rate": 1.9720462503517e-05, "loss": 1.0569, "step": 508 }, { "epoch": 1.0345528455284554, "grad_norm": 0.8338561515252424, "learning_rate": 1.9718914687138836e-05, "loss": 0.892, "step": 509 }, { "epoch": 1.0365853658536586, "grad_norm": 1.048434684293557, "learning_rate": 1.9717362658486515e-05, "loss": 1.2267, "step": 510 }, { "epoch": 1.0386178861788617, "grad_norm": 0.8090002608809052, "learning_rate": 1.9715806418232707e-05, "loss": 1.0331, "step": 511 }, { "epoch": 1.040650406504065, "grad_norm": 1.0038519126076604, "learning_rate": 1.971424596705189e-05, "loss": 0.9952, "step": 512 }, { "epoch": 1.0426829268292683, "grad_norm": 0.9477979117971985, "learning_rate": 1.971268130562039e-05, "loss": 0.9445, "step": 513 }, { "epoch": 1.0447154471544715, "grad_norm": 0.769797140690112, "learning_rate": 1.971111243461634e-05, "loss": 0.8902, "step": 514 }, { "epoch": 1.0467479674796747, "grad_norm": 0.866960560433393, "learning_rate": 1.9709539354719707e-05, "loss": 0.9437, "step": 515 }, { "epoch": 1.048780487804878, "grad_norm": 1.0068508727432885, "learning_rate": 1.9707962066612278e-05, "loss": 1.1071, "step": 516 }, { "epoch": 1.0508130081300813, "grad_norm": 0.9143179652902571, "learning_rate": 1.970638057097767e-05, "loss": 1.0065, "step": 517 }, { "epoch": 1.0528455284552845, "grad_norm": 1.0081539268983915, "learning_rate": 1.9704794868501314e-05, "loss": 1.0687, "step": 518 }, { "epoch": 1.0548780487804879, "grad_norm": 0.9814922751010605, "learning_rate": 1.970320495987047e-05, "loss": 0.9504, "step": 519 }, { "epoch": 1.056910569105691, "grad_norm": 0.9227882785986544, "learning_rate": 1.970161084577422e-05, "loss": 1.0029, "step": 520 }, { "epoch": 1.0589430894308942, "grad_norm": 0.9476431219479937, "learning_rate": 1.9700012526903473e-05, "loss": 0.8949, "step": 521 }, { "epoch": 1.0609756097560976, "grad_norm": 0.9476858671380127, "learning_rate": 1.9698410003950953e-05, "loss": 0.9543, "step": 522 }, { "epoch": 1.0630081300813008, "grad_norm": 0.9105724148517178, "learning_rate": 1.9696803277611205e-05, "loss": 0.9033, "step": 523 }, { "epoch": 1.065040650406504, "grad_norm": 0.959095627376021, "learning_rate": 1.9695192348580606e-05, "loss": 1.3785, "step": 524 }, { "epoch": 1.0670731707317074, "grad_norm": 0.9382991240239209, "learning_rate": 1.969357721755735e-05, "loss": 1.0604, "step": 525 }, { "epoch": 1.0691056910569106, "grad_norm": 1.3265490121342782, "learning_rate": 1.9691957885241448e-05, "loss": 0.9654, "step": 526 }, { "epoch": 1.0711382113821137, "grad_norm": 1.0392400977234846, "learning_rate": 1.969033435233473e-05, "loss": 1.1916, "step": 527 }, { "epoch": 1.0731707317073171, "grad_norm": 0.9502670525463409, "learning_rate": 1.9688706619540863e-05, "loss": 1.033, "step": 528 }, { "epoch": 1.0752032520325203, "grad_norm": 1.0509213439699032, "learning_rate": 1.968707468756531e-05, "loss": 1.2185, "step": 529 }, { "epoch": 1.0772357723577235, "grad_norm": 0.9519964272124978, "learning_rate": 1.9685438557115376e-05, "loss": 0.9599, "step": 530 }, { "epoch": 1.079268292682927, "grad_norm": 1.1401137359669156, "learning_rate": 1.968379822890017e-05, "loss": 1.206, "step": 531 }, { "epoch": 1.08130081300813, "grad_norm": 0.9254403893118026, "learning_rate": 1.968215370363063e-05, "loss": 0.7942, "step": 532 }, { "epoch": 1.0833333333333333, "grad_norm": 0.9048771320631627, "learning_rate": 1.9680504982019506e-05, "loss": 0.968, "step": 533 }, { "epoch": 1.0853658536585367, "grad_norm": 0.872598238446131, "learning_rate": 1.9678852064781372e-05, "loss": 0.8925, "step": 534 }, { "epoch": 1.0873983739837398, "grad_norm": 0.8200198308916853, "learning_rate": 1.9677194952632625e-05, "loss": 0.857, "step": 535 }, { "epoch": 1.089430894308943, "grad_norm": 0.9955307930230053, "learning_rate": 1.9675533646291463e-05, "loss": 1.0388, "step": 536 }, { "epoch": 1.0914634146341464, "grad_norm": 0.8685497526163544, "learning_rate": 1.967386814647792e-05, "loss": 0.9, "step": 537 }, { "epoch": 1.0934959349593496, "grad_norm": 1.0073252163726403, "learning_rate": 1.967219845391384e-05, "loss": 0.9064, "step": 538 }, { "epoch": 1.0955284552845528, "grad_norm": 1.008977386626071, "learning_rate": 1.967052456932288e-05, "loss": 1.1986, "step": 539 }, { "epoch": 1.0975609756097562, "grad_norm": 0.9742079904509064, "learning_rate": 1.9668846493430522e-05, "loss": 1.1208, "step": 540 }, { "epoch": 1.0995934959349594, "grad_norm": 1.030740688551819, "learning_rate": 1.966716422696406e-05, "loss": 1.1988, "step": 541 }, { "epoch": 1.1016260162601625, "grad_norm": 1.0866307086362366, "learning_rate": 1.9665477770652602e-05, "loss": 1.1048, "step": 542 }, { "epoch": 1.103658536585366, "grad_norm": 0.976062938832535, "learning_rate": 1.966378712522708e-05, "loss": 0.9779, "step": 543 }, { "epoch": 1.1056910569105691, "grad_norm": 1.0579837362616067, "learning_rate": 1.9662092291420233e-05, "loss": 1.1631, "step": 544 }, { "epoch": 1.1077235772357723, "grad_norm": 0.9029208862109198, "learning_rate": 1.966039326996662e-05, "loss": 0.9237, "step": 545 }, { "epoch": 1.1097560975609757, "grad_norm": 0.8727714487853463, "learning_rate": 1.9658690061602614e-05, "loss": 0.771, "step": 546 }, { "epoch": 1.1117886178861789, "grad_norm": 1.2040815359999504, "learning_rate": 1.96569826670664e-05, "loss": 1.1337, "step": 547 }, { "epoch": 1.113821138211382, "grad_norm": 1.1092878734674854, "learning_rate": 1.965527108709798e-05, "loss": 1.1023, "step": 548 }, { "epoch": 1.1158536585365855, "grad_norm": 0.9966645539399892, "learning_rate": 1.9653555322439176e-05, "loss": 1.04, "step": 549 }, { "epoch": 1.1178861788617886, "grad_norm": 1.031778600515744, "learning_rate": 1.9651835373833613e-05, "loss": 1.1368, "step": 550 }, { "epoch": 1.1199186991869918, "grad_norm": 0.8364382956840528, "learning_rate": 1.9650111242026732e-05, "loss": 0.8988, "step": 551 }, { "epoch": 1.1219512195121952, "grad_norm": 1.0638249134468187, "learning_rate": 1.964838292776579e-05, "loss": 1.0758, "step": 552 }, { "epoch": 1.1239837398373984, "grad_norm": 1.0870780327598886, "learning_rate": 1.9646650431799855e-05, "loss": 0.8828, "step": 553 }, { "epoch": 1.1260162601626016, "grad_norm": 0.9781677489862137, "learning_rate": 1.964491375487981e-05, "loss": 0.8184, "step": 554 }, { "epoch": 1.1280487804878048, "grad_norm": 1.0607590974961458, "learning_rate": 1.964317289775835e-05, "loss": 0.8548, "step": 555 }, { "epoch": 1.1300813008130082, "grad_norm": 1.1911593440825776, "learning_rate": 1.9641427861189973e-05, "loss": 1.1225, "step": 556 }, { "epoch": 1.1321138211382114, "grad_norm": 0.9390414818503613, "learning_rate": 1.9639678645931002e-05, "loss": 1.0706, "step": 557 }, { "epoch": 1.1341463414634148, "grad_norm": 0.8952157776483269, "learning_rate": 1.963792525273956e-05, "loss": 0.9529, "step": 558 }, { "epoch": 1.136178861788618, "grad_norm": 0.8324089284511068, "learning_rate": 1.9636167682375587e-05, "loss": 0.8206, "step": 559 }, { "epoch": 1.1382113821138211, "grad_norm": 0.8191790623570732, "learning_rate": 1.963440593560083e-05, "loss": 0.8329, "step": 560 }, { "epoch": 1.1402439024390243, "grad_norm": 0.8515089784839401, "learning_rate": 1.963264001317885e-05, "loss": 0.8944, "step": 561 }, { "epoch": 1.1422764227642277, "grad_norm": 0.9473263735500067, "learning_rate": 1.9630869915875012e-05, "loss": 1.0415, "step": 562 }, { "epoch": 1.1443089430894309, "grad_norm": 1.12059280726806, "learning_rate": 1.9629095644456495e-05, "loss": 1.0907, "step": 563 }, { "epoch": 1.146341463414634, "grad_norm": 1.0434975977525258, "learning_rate": 1.9627317199692287e-05, "loss": 0.9394, "step": 564 }, { "epoch": 1.1483739837398375, "grad_norm": 1.065971800176704, "learning_rate": 1.962553458235318e-05, "loss": 1.2149, "step": 565 }, { "epoch": 1.1504065040650406, "grad_norm": 1.1476608601955756, "learning_rate": 1.9623747793211784e-05, "loss": 1.4029, "step": 566 }, { "epoch": 1.1524390243902438, "grad_norm": 1.0621530314545733, "learning_rate": 1.9621956833042507e-05, "loss": 1.1128, "step": 567 }, { "epoch": 1.1544715447154472, "grad_norm": 1.1022028809109568, "learning_rate": 1.962016170262157e-05, "loss": 0.8697, "step": 568 }, { "epoch": 1.1565040650406504, "grad_norm": 0.9379519622199853, "learning_rate": 1.9618362402727003e-05, "loss": 0.8749, "step": 569 }, { "epoch": 1.1585365853658536, "grad_norm": 1.0752567366854306, "learning_rate": 1.9616558934138637e-05, "loss": 1.2139, "step": 570 }, { "epoch": 1.160569105691057, "grad_norm": 1.0452815959225727, "learning_rate": 1.961475129763811e-05, "loss": 1.071, "step": 571 }, { "epoch": 1.1626016260162602, "grad_norm": 0.9876431278867317, "learning_rate": 1.961293949400888e-05, "loss": 0.9854, "step": 572 }, { "epoch": 1.1646341463414633, "grad_norm": 0.9804891003173283, "learning_rate": 1.961112352403619e-05, "loss": 1.1502, "step": 573 }, { "epoch": 1.1666666666666667, "grad_norm": 0.986950862310023, "learning_rate": 1.9609303388507105e-05, "loss": 0.9665, "step": 574 }, { "epoch": 1.16869918699187, "grad_norm": 0.8406390928230718, "learning_rate": 1.960747908821049e-05, "loss": 1.0098, "step": 575 }, { "epoch": 1.170731707317073, "grad_norm": 1.0210553286567723, "learning_rate": 1.960565062393701e-05, "loss": 0.9317, "step": 576 }, { "epoch": 1.1727642276422765, "grad_norm": 0.9494676956402708, "learning_rate": 1.9603817996479148e-05, "loss": 0.8168, "step": 577 }, { "epoch": 1.1747967479674797, "grad_norm": 1.302035182688394, "learning_rate": 1.960198120663117e-05, "loss": 1.0575, "step": 578 }, { "epoch": 1.1768292682926829, "grad_norm": 1.0695758976820913, "learning_rate": 1.9600140255189175e-05, "loss": 1.0516, "step": 579 }, { "epoch": 1.1788617886178863, "grad_norm": 0.8241341182269302, "learning_rate": 1.9598295142951035e-05, "loss": 0.8644, "step": 580 }, { "epoch": 1.1808943089430894, "grad_norm": 1.1389369297183656, "learning_rate": 1.9596445870716453e-05, "loss": 1.1174, "step": 581 }, { "epoch": 1.1829268292682926, "grad_norm": 1.0339138588051022, "learning_rate": 1.959459243928691e-05, "loss": 1.0757, "step": 582 }, { "epoch": 1.184959349593496, "grad_norm": 0.8382639817204828, "learning_rate": 1.959273484946571e-05, "loss": 0.9571, "step": 583 }, { "epoch": 1.1869918699186992, "grad_norm": 1.0154339088294115, "learning_rate": 1.9590873102057948e-05, "loss": 0.8301, "step": 584 }, { "epoch": 1.1890243902439024, "grad_norm": 0.9492385770700219, "learning_rate": 1.9589007197870523e-05, "loss": 0.9154, "step": 585 }, { "epoch": 1.1910569105691058, "grad_norm": 1.053042113971433, "learning_rate": 1.9587137137712137e-05, "loss": 1.1256, "step": 586 }, { "epoch": 1.193089430894309, "grad_norm": 0.9527759402604629, "learning_rate": 1.9585262922393292e-05, "loss": 1.1042, "step": 587 }, { "epoch": 1.1951219512195121, "grad_norm": 0.9440031399792729, "learning_rate": 1.9583384552726294e-05, "loss": 1.0881, "step": 588 }, { "epoch": 1.1971544715447155, "grad_norm": 1.023661085980655, "learning_rate": 1.9581502029525245e-05, "loss": 0.966, "step": 589 }, { "epoch": 1.1991869918699187, "grad_norm": 1.0062449810172498, "learning_rate": 1.9579615353606053e-05, "loss": 1.0622, "step": 590 }, { "epoch": 1.201219512195122, "grad_norm": 0.9220838532115284, "learning_rate": 1.957772452578641e-05, "loss": 0.8577, "step": 591 }, { "epoch": 1.203252032520325, "grad_norm": 0.9627483191157311, "learning_rate": 1.957582954688584e-05, "loss": 1.0463, "step": 592 }, { "epoch": 1.2052845528455285, "grad_norm": 1.0255637969900555, "learning_rate": 1.9573930417725626e-05, "loss": 1.1338, "step": 593 }, { "epoch": 1.2073170731707317, "grad_norm": 0.8883701805842332, "learning_rate": 1.9572027139128877e-05, "loss": 0.9152, "step": 594 }, { "epoch": 1.209349593495935, "grad_norm": 0.9919089554823538, "learning_rate": 1.9570119711920495e-05, "loss": 1.0828, "step": 595 }, { "epoch": 1.2113821138211383, "grad_norm": 0.9965065358314751, "learning_rate": 1.9568208136927177e-05, "loss": 0.9222, "step": 596 }, { "epoch": 1.2134146341463414, "grad_norm": 0.9297541298358124, "learning_rate": 1.956629241497742e-05, "loss": 0.9957, "step": 597 }, { "epoch": 1.2154471544715446, "grad_norm": 1.0789015703383775, "learning_rate": 1.9564372546901512e-05, "loss": 1.165, "step": 598 }, { "epoch": 1.217479674796748, "grad_norm": 0.9706984629071367, "learning_rate": 1.9562448533531548e-05, "loss": 0.8826, "step": 599 }, { "epoch": 1.2195121951219512, "grad_norm": 1.0002290912162703, "learning_rate": 1.9560520375701408e-05, "loss": 0.9033, "step": 600 }, { "epoch": 1.2215447154471546, "grad_norm": 1.082574459096385, "learning_rate": 1.955858807424678e-05, "loss": 1.0247, "step": 601 }, { "epoch": 1.2235772357723578, "grad_norm": 0.9721018065651923, "learning_rate": 1.9556651630005144e-05, "loss": 1.0158, "step": 602 }, { "epoch": 1.225609756097561, "grad_norm": 1.09733529375878, "learning_rate": 1.9554711043815772e-05, "loss": 0.9, "step": 603 }, { "epoch": 1.2276422764227641, "grad_norm": 1.0008777893720482, "learning_rate": 1.9552766316519726e-05, "loss": 1.1968, "step": 604 }, { "epoch": 1.2296747967479675, "grad_norm": 1.1105654710565063, "learning_rate": 1.9550817448959886e-05, "loss": 1.1366, "step": 605 }, { "epoch": 1.2317073170731707, "grad_norm": 0.984235675450752, "learning_rate": 1.9548864441980897e-05, "loss": 1.0297, "step": 606 }, { "epoch": 1.233739837398374, "grad_norm": 0.8474922634461918, "learning_rate": 1.9546907296429216e-05, "loss": 0.7547, "step": 607 }, { "epoch": 1.2357723577235773, "grad_norm": 1.0187314568588217, "learning_rate": 1.9544946013153093e-05, "loss": 1.0329, "step": 608 }, { "epoch": 1.2378048780487805, "grad_norm": 1.0033636041182572, "learning_rate": 1.954298059300256e-05, "loss": 1.0439, "step": 609 }, { "epoch": 1.2398373983739837, "grad_norm": 1.056705416849351, "learning_rate": 1.9541011036829455e-05, "loss": 0.9923, "step": 610 }, { "epoch": 1.241869918699187, "grad_norm": 0.8843439547543105, "learning_rate": 1.9539037345487404e-05, "loss": 0.8476, "step": 611 }, { "epoch": 1.2439024390243902, "grad_norm": 1.0028235978210454, "learning_rate": 1.9537059519831822e-05, "loss": 0.8438, "step": 612 }, { "epoch": 1.2459349593495934, "grad_norm": 1.0197955038755855, "learning_rate": 1.953507756071992e-05, "loss": 1.1247, "step": 613 }, { "epoch": 1.2479674796747968, "grad_norm": 0.9831074752358715, "learning_rate": 1.9533091469010694e-05, "loss": 0.9378, "step": 614 }, { "epoch": 1.25, "grad_norm": 0.9704036322083621, "learning_rate": 1.9531101245564944e-05, "loss": 0.9341, "step": 615 }, { "epoch": 1.2520325203252032, "grad_norm": 1.1760280523602702, "learning_rate": 1.9529106891245244e-05, "loss": 0.9379, "step": 616 }, { "epoch": 1.2540650406504066, "grad_norm": 1.1125939551634654, "learning_rate": 1.952710840691597e-05, "loss": 0.9081, "step": 617 }, { "epoch": 1.2560975609756098, "grad_norm": 0.9608442934011571, "learning_rate": 1.9525105793443288e-05, "loss": 0.9555, "step": 618 }, { "epoch": 1.258130081300813, "grad_norm": 0.9147128934010139, "learning_rate": 1.9523099051695153e-05, "loss": 1.0244, "step": 619 }, { "epoch": 1.2601626016260163, "grad_norm": 0.964970705695258, "learning_rate": 1.9521088182541298e-05, "loss": 0.968, "step": 620 }, { "epoch": 1.2621951219512195, "grad_norm": 1.123968155861131, "learning_rate": 1.9519073186853256e-05, "loss": 1.1213, "step": 621 }, { "epoch": 1.2642276422764227, "grad_norm": 0.9658759701321707, "learning_rate": 1.9517054065504352e-05, "loss": 1.0832, "step": 622 }, { "epoch": 1.2662601626016259, "grad_norm": 1.093747624916482, "learning_rate": 1.9515030819369685e-05, "loss": 1.1348, "step": 623 }, { "epoch": 1.2682926829268293, "grad_norm": 1.042388013059673, "learning_rate": 1.951300344932616e-05, "loss": 0.9884, "step": 624 }, { "epoch": 1.2703252032520325, "grad_norm": 1.1844257266586038, "learning_rate": 1.9510971956252453e-05, "loss": 0.8726, "step": 625 }, { "epoch": 1.2723577235772359, "grad_norm": 1.04138434556784, "learning_rate": 1.950893634102903e-05, "loss": 0.9228, "step": 626 }, { "epoch": 1.274390243902439, "grad_norm": 1.0772773977245416, "learning_rate": 1.9506896604538155e-05, "loss": 1.0471, "step": 627 }, { "epoch": 1.2764227642276422, "grad_norm": 1.143037689587824, "learning_rate": 1.9504852747663862e-05, "loss": 1.0451, "step": 628 }, { "epoch": 1.2784552845528454, "grad_norm": 1.0442638963539712, "learning_rate": 1.9502804771291983e-05, "loss": 0.8657, "step": 629 }, { "epoch": 1.2804878048780488, "grad_norm": 1.2626354692994493, "learning_rate": 1.9500752676310136e-05, "loss": 1.3271, "step": 630 }, { "epoch": 1.282520325203252, "grad_norm": 1.022835441699893, "learning_rate": 1.9498696463607708e-05, "loss": 1.0009, "step": 631 }, { "epoch": 1.2845528455284554, "grad_norm": 1.0238161393090686, "learning_rate": 1.9496636134075894e-05, "loss": 0.8619, "step": 632 }, { "epoch": 1.2865853658536586, "grad_norm": 1.107116810938377, "learning_rate": 1.9494571688607654e-05, "loss": 1.1449, "step": 633 }, { "epoch": 1.2886178861788617, "grad_norm": 1.041941537630491, "learning_rate": 1.949250312809774e-05, "loss": 0.8005, "step": 634 }, { "epoch": 1.290650406504065, "grad_norm": 1.0379472860617085, "learning_rate": 1.9490430453442687e-05, "loss": 1.0341, "step": 635 }, { "epoch": 1.2926829268292683, "grad_norm": 1.0053247382715982, "learning_rate": 1.9488353665540813e-05, "loss": 0.9834, "step": 636 }, { "epoch": 1.2947154471544715, "grad_norm": 1.2007403975305015, "learning_rate": 1.948627276529222e-05, "loss": 1.0624, "step": 637 }, { "epoch": 1.296747967479675, "grad_norm": 1.0363561898451492, "learning_rate": 1.948418775359879e-05, "loss": 1.0069, "step": 638 }, { "epoch": 1.298780487804878, "grad_norm": 0.9997224499933353, "learning_rate": 1.9482098631364185e-05, "loss": 1.0971, "step": 639 }, { "epoch": 1.3008130081300813, "grad_norm": 1.1837351377939331, "learning_rate": 1.9480005399493857e-05, "loss": 1.0018, "step": 640 }, { "epoch": 1.3028455284552845, "grad_norm": 1.0517284524924406, "learning_rate": 1.9477908058895028e-05, "loss": 1.0196, "step": 641 }, { "epoch": 1.3048780487804879, "grad_norm": 0.9366704251546321, "learning_rate": 1.947580661047671e-05, "loss": 0.8126, "step": 642 }, { "epoch": 1.306910569105691, "grad_norm": 1.013495176742294, "learning_rate": 1.9473701055149687e-05, "loss": 0.9414, "step": 643 }, { "epoch": 1.3089430894308944, "grad_norm": 1.171023232646797, "learning_rate": 1.9471591393826536e-05, "loss": 0.9219, "step": 644 }, { "epoch": 1.3109756097560976, "grad_norm": 1.0277426238393792, "learning_rate": 1.9469477627421595e-05, "loss": 0.8156, "step": 645 }, { "epoch": 1.3130081300813008, "grad_norm": 1.0591793976945811, "learning_rate": 1.9467359756851e-05, "loss": 0.9114, "step": 646 }, { "epoch": 1.315040650406504, "grad_norm": 1.120861708187247, "learning_rate": 1.946523778303265e-05, "loss": 0.9285, "step": 647 }, { "epoch": 1.3170731707317074, "grad_norm": 1.1534333803993144, "learning_rate": 1.9463111706886234e-05, "loss": 1.073, "step": 648 }, { "epoch": 1.3191056910569106, "grad_norm": 1.2712466106290885, "learning_rate": 1.9460981529333213e-05, "loss": 1.0078, "step": 649 }, { "epoch": 1.321138211382114, "grad_norm": 1.0164344258902422, "learning_rate": 1.945884725129683e-05, "loss": 0.9894, "step": 650 }, { "epoch": 1.3231707317073171, "grad_norm": 1.1295294772630893, "learning_rate": 1.94567088737021e-05, "loss": 1.0257, "step": 651 }, { "epoch": 1.3252032520325203, "grad_norm": 1.0664453181992402, "learning_rate": 1.9454566397475813e-05, "loss": 0.9191, "step": 652 }, { "epoch": 1.3272357723577235, "grad_norm": 1.0596887343878676, "learning_rate": 1.9452419823546545e-05, "loss": 1.0329, "step": 653 }, { "epoch": 1.329268292682927, "grad_norm": 1.159435298114323, "learning_rate": 1.9450269152844643e-05, "loss": 0.8711, "step": 654 }, { "epoch": 1.33130081300813, "grad_norm": 1.2349359190547953, "learning_rate": 1.9448114386302226e-05, "loss": 1.0545, "step": 655 }, { "epoch": 1.3333333333333333, "grad_norm": 1.1745215690108461, "learning_rate": 1.944595552485319e-05, "loss": 0.9755, "step": 656 }, { "epoch": 1.3353658536585367, "grad_norm": 1.082095987678102, "learning_rate": 1.944379256943321e-05, "loss": 0.9763, "step": 657 }, { "epoch": 1.3373983739837398, "grad_norm": 1.0150599994830385, "learning_rate": 1.9441625520979736e-05, "loss": 1.0419, "step": 658 }, { "epoch": 1.339430894308943, "grad_norm": 1.083300054409599, "learning_rate": 1.9439454380431977e-05, "loss": 0.9919, "step": 659 }, { "epoch": 1.3414634146341464, "grad_norm": 0.9851111639626815, "learning_rate": 1.943727914873094e-05, "loss": 0.9542, "step": 660 }, { "epoch": 1.3434959349593496, "grad_norm": 1.0931397764257782, "learning_rate": 1.943509982681938e-05, "loss": 1.0049, "step": 661 }, { "epoch": 1.3455284552845528, "grad_norm": 1.0703975133012424, "learning_rate": 1.9432916415641844e-05, "loss": 0.9306, "step": 662 }, { "epoch": 1.3475609756097562, "grad_norm": 0.9823975231774701, "learning_rate": 1.9430728916144642e-05, "loss": 0.7659, "step": 663 }, { "epoch": 1.3495934959349594, "grad_norm": 1.190984746547742, "learning_rate": 1.9428537329275862e-05, "loss": 1.101, "step": 664 }, { "epoch": 1.3516260162601625, "grad_norm": 1.008249663136625, "learning_rate": 1.9426341655985353e-05, "loss": 0.9798, "step": 665 }, { "epoch": 1.3536585365853657, "grad_norm": 1.1507486017457853, "learning_rate": 1.9424141897224747e-05, "loss": 0.9403, "step": 666 }, { "epoch": 1.3556910569105691, "grad_norm": 1.1249448259640789, "learning_rate": 1.9421938053947435e-05, "loss": 1.1113, "step": 667 }, { "epoch": 1.3577235772357723, "grad_norm": 1.1335636893481633, "learning_rate": 1.941973012710859e-05, "loss": 1.1532, "step": 668 }, { "epoch": 1.3597560975609757, "grad_norm": 1.1196141823600272, "learning_rate": 1.941751811766515e-05, "loss": 0.9871, "step": 669 }, { "epoch": 1.3617886178861789, "grad_norm": 1.0186532591203394, "learning_rate": 1.941530202657582e-05, "loss": 0.9549, "step": 670 }, { "epoch": 1.363821138211382, "grad_norm": 0.9948615109990332, "learning_rate": 1.9413081854801076e-05, "loss": 1.0043, "step": 671 }, { "epoch": 1.3658536585365852, "grad_norm": 1.0945118168634727, "learning_rate": 1.941085760330316e-05, "loss": 1.1693, "step": 672 }, { "epoch": 1.3678861788617886, "grad_norm": 0.9125525287811888, "learning_rate": 1.940862927304609e-05, "loss": 0.7902, "step": 673 }, { "epoch": 1.3699186991869918, "grad_norm": 1.348836625621657, "learning_rate": 1.940639686499564e-05, "loss": 1.1989, "step": 674 }, { "epoch": 1.3719512195121952, "grad_norm": 1.1099198382724187, "learning_rate": 1.9404160380119364e-05, "loss": 1.0604, "step": 675 }, { "epoch": 1.3739837398373984, "grad_norm": 1.068946163441042, "learning_rate": 1.940191981938657e-05, "loss": 0.9821, "step": 676 }, { "epoch": 1.3760162601626016, "grad_norm": 1.1645513839572994, "learning_rate": 1.9399675183768346e-05, "loss": 0.976, "step": 677 }, { "epoch": 1.3780487804878048, "grad_norm": 1.1514256623189216, "learning_rate": 1.9397426474237538e-05, "loss": 0.8568, "step": 678 }, { "epoch": 1.3800813008130082, "grad_norm": 1.096095914205792, "learning_rate": 1.9395173691768755e-05, "loss": 0.9302, "step": 679 }, { "epoch": 1.3821138211382114, "grad_norm": 1.1640570527245329, "learning_rate": 1.9392916837338376e-05, "loss": 1.1458, "step": 680 }, { "epoch": 1.3841463414634148, "grad_norm": 1.0499222290608226, "learning_rate": 1.939065591192455e-05, "loss": 1.163, "step": 681 }, { "epoch": 1.386178861788618, "grad_norm": 1.1295664095423157, "learning_rate": 1.938839091650718e-05, "loss": 1.0541, "step": 682 }, { "epoch": 1.3882113821138211, "grad_norm": 0.916988980990881, "learning_rate": 1.938612185206794e-05, "loss": 0.8477, "step": 683 }, { "epoch": 1.3902439024390243, "grad_norm": 1.1984418933365262, "learning_rate": 1.9383848719590257e-05, "loss": 1.0074, "step": 684 }, { "epoch": 1.3922764227642277, "grad_norm": 1.2012097242070081, "learning_rate": 1.9381571520059342e-05, "loss": 1.0192, "step": 685 }, { "epoch": 1.3943089430894309, "grad_norm": 1.1434606700810577, "learning_rate": 1.9379290254462147e-05, "loss": 1.0225, "step": 686 }, { "epoch": 1.3963414634146343, "grad_norm": 1.0353212534916705, "learning_rate": 1.9377004923787395e-05, "loss": 0.933, "step": 687 }, { "epoch": 1.3983739837398375, "grad_norm": 1.0660891794611684, "learning_rate": 1.9374715529025575e-05, "loss": 1.041, "step": 688 }, { "epoch": 1.4004065040650406, "grad_norm": 0.9614885483600116, "learning_rate": 1.9372422071168936e-05, "loss": 1.0076, "step": 689 }, { "epoch": 1.4024390243902438, "grad_norm": 1.072143548546644, "learning_rate": 1.9370124551211476e-05, "loss": 1.227, "step": 690 }, { "epoch": 1.4044715447154472, "grad_norm": 1.038558920372387, "learning_rate": 1.9367822970148967e-05, "loss": 1.0234, "step": 691 }, { "epoch": 1.4065040650406504, "grad_norm": 1.134090339335001, "learning_rate": 1.9365517328978943e-05, "loss": 1.0311, "step": 692 }, { "epoch": 1.4085365853658536, "grad_norm": 1.083363414946034, "learning_rate": 1.9363207628700685e-05, "loss": 1.0503, "step": 693 }, { "epoch": 1.410569105691057, "grad_norm": 1.2156586258358328, "learning_rate": 1.9360893870315244e-05, "loss": 1.041, "step": 694 }, { "epoch": 1.4126016260162602, "grad_norm": 1.0820964972826081, "learning_rate": 1.9358576054825423e-05, "loss": 1.1187, "step": 695 }, { "epoch": 1.4146341463414633, "grad_norm": 1.1066380306039862, "learning_rate": 1.9356254183235785e-05, "loss": 0.9687, "step": 696 }, { "epoch": 1.4166666666666667, "grad_norm": 1.0065568739239037, "learning_rate": 1.9353928256552657e-05, "loss": 0.9254, "step": 697 }, { "epoch": 1.41869918699187, "grad_norm": 1.0870235046508003, "learning_rate": 1.9351598275784116e-05, "loss": 0.9345, "step": 698 }, { "epoch": 1.420731707317073, "grad_norm": 1.0688726064964853, "learning_rate": 1.934926424194e-05, "loss": 0.9643, "step": 699 }, { "epoch": 1.4227642276422765, "grad_norm": 1.1908383786196635, "learning_rate": 1.93469261560319e-05, "loss": 1.0289, "step": 700 }, { "epoch": 1.4247967479674797, "grad_norm": 0.9981510578882552, "learning_rate": 1.9344584019073167e-05, "loss": 0.9982, "step": 701 }, { "epoch": 1.4268292682926829, "grad_norm": 0.976517837465255, "learning_rate": 1.9342237832078906e-05, "loss": 0.8425, "step": 702 }, { "epoch": 1.428861788617886, "grad_norm": 1.3276268057483074, "learning_rate": 1.933988759606598e-05, "loss": 0.9002, "step": 703 }, { "epoch": 1.4308943089430894, "grad_norm": 1.0611075708377762, "learning_rate": 1.9337533312053002e-05, "loss": 0.9867, "step": 704 }, { "epoch": 1.4329268292682926, "grad_norm": 1.1703020078334052, "learning_rate": 1.9335174981060342e-05, "loss": 1.1129, "step": 705 }, { "epoch": 1.434959349593496, "grad_norm": 1.019181774087124, "learning_rate": 1.9332812604110127e-05, "loss": 0.8931, "step": 706 }, { "epoch": 1.4369918699186992, "grad_norm": 1.1641644605847221, "learning_rate": 1.9330446182226233e-05, "loss": 1.0461, "step": 707 }, { "epoch": 1.4390243902439024, "grad_norm": 1.1138133850112382, "learning_rate": 1.9328075716434287e-05, "loss": 0.9523, "step": 708 }, { "epoch": 1.4410569105691056, "grad_norm": 1.2884117424532884, "learning_rate": 1.932570120776168e-05, "loss": 1.0415, "step": 709 }, { "epoch": 1.443089430894309, "grad_norm": 1.1265683925149688, "learning_rate": 1.932332265723754e-05, "loss": 0.9241, "step": 710 }, { "epoch": 1.4451219512195121, "grad_norm": 1.0598615485334557, "learning_rate": 1.9320940065892762e-05, "loss": 1.0097, "step": 711 }, { "epoch": 1.4471544715447155, "grad_norm": 0.9867906053681914, "learning_rate": 1.931855343475998e-05, "loss": 0.9631, "step": 712 }, { "epoch": 1.4491869918699187, "grad_norm": 1.0342383124783099, "learning_rate": 1.931616276487358e-05, "loss": 1.0871, "step": 713 }, { "epoch": 1.451219512195122, "grad_norm": 1.2616061968493317, "learning_rate": 1.9313768057269714e-05, "loss": 0.8755, "step": 714 }, { "epoch": 1.453252032520325, "grad_norm": 1.0318168369430845, "learning_rate": 1.9311369312986258e-05, "loss": 1.013, "step": 715 }, { "epoch": 1.4552845528455285, "grad_norm": 1.2387777003645628, "learning_rate": 1.930896653306286e-05, "loss": 1.0653, "step": 716 }, { "epoch": 1.4573170731707317, "grad_norm": 1.1054984446251568, "learning_rate": 1.9306559718540906e-05, "loss": 1.0905, "step": 717 }, { "epoch": 1.459349593495935, "grad_norm": 1.023825079312879, "learning_rate": 1.9304148870463534e-05, "loss": 0.9382, "step": 718 }, { "epoch": 1.4613821138211383, "grad_norm": 1.1236077329422962, "learning_rate": 1.930173398987563e-05, "loss": 1.0731, "step": 719 }, { "epoch": 1.4634146341463414, "grad_norm": 0.9776379444866141, "learning_rate": 1.929931507782383e-05, "loss": 1.044, "step": 720 }, { "epoch": 1.4654471544715446, "grad_norm": 1.1836184549311881, "learning_rate": 1.929689213535651e-05, "loss": 0.964, "step": 721 }, { "epoch": 1.467479674796748, "grad_norm": 1.1489549254510127, "learning_rate": 1.9294465163523794e-05, "loss": 0.9768, "step": 722 }, { "epoch": 1.4695121951219512, "grad_norm": 1.1910177823556818, "learning_rate": 1.9292034163377565e-05, "loss": 1.0084, "step": 723 }, { "epoch": 1.4715447154471546, "grad_norm": 1.2687094958335101, "learning_rate": 1.9289599135971437e-05, "loss": 1.1006, "step": 724 }, { "epoch": 1.4735772357723578, "grad_norm": 1.01861158776981, "learning_rate": 1.9287160082360775e-05, "loss": 0.85, "step": 725 }, { "epoch": 1.475609756097561, "grad_norm": 0.9982150398218703, "learning_rate": 1.928471700360269e-05, "loss": 0.8813, "step": 726 }, { "epoch": 1.4776422764227641, "grad_norm": 1.1831757316713678, "learning_rate": 1.9282269900756038e-05, "loss": 1.0445, "step": 727 }, { "epoch": 1.4796747967479675, "grad_norm": 1.0944237246681818, "learning_rate": 1.9279818774881418e-05, "loss": 1.0032, "step": 728 }, { "epoch": 1.4817073170731707, "grad_norm": 1.1426797452105049, "learning_rate": 1.927736362704117e-05, "loss": 1.0752, "step": 729 }, { "epoch": 1.4837398373983741, "grad_norm": 1.1116660012480626, "learning_rate": 1.927490445829938e-05, "loss": 1.0341, "step": 730 }, { "epoch": 1.4857723577235773, "grad_norm": 1.1050059881837964, "learning_rate": 1.927244126972187e-05, "loss": 0.7792, "step": 731 }, { "epoch": 1.4878048780487805, "grad_norm": 1.1427379499865755, "learning_rate": 1.9269974062376224e-05, "loss": 1.0364, "step": 732 }, { "epoch": 1.4898373983739837, "grad_norm": 1.080054966814705, "learning_rate": 1.926750283733174e-05, "loss": 1.011, "step": 733 }, { "epoch": 1.491869918699187, "grad_norm": 1.044562397889363, "learning_rate": 1.9265027595659488e-05, "loss": 0.9735, "step": 734 }, { "epoch": 1.4939024390243902, "grad_norm": 1.5488639746092212, "learning_rate": 1.9262548338432246e-05, "loss": 1.1137, "step": 735 }, { "epoch": 1.4959349593495934, "grad_norm": 1.2831988072139413, "learning_rate": 1.926006506672456e-05, "loss": 0.8967, "step": 736 }, { "epoch": 1.4979674796747968, "grad_norm": 1.1336153623576628, "learning_rate": 1.9257577781612693e-05, "loss": 0.7672, "step": 737 }, { "epoch": 1.5, "grad_norm": 1.252226484335923, "learning_rate": 1.925508648417467e-05, "loss": 1.1063, "step": 738 }, { "epoch": 1.5020325203252032, "grad_norm": 1.2497850636030459, "learning_rate": 1.925259117549024e-05, "loss": 0.9332, "step": 739 }, { "epoch": 1.5040650406504064, "grad_norm": 1.2545128967308559, "learning_rate": 1.9250091856640895e-05, "loss": 0.9206, "step": 740 }, { "epoch": 1.5060975609756098, "grad_norm": 1.079972532992215, "learning_rate": 1.9247588528709867e-05, "loss": 1.0332, "step": 741 }, { "epoch": 1.5081300813008132, "grad_norm": 1.083962142048209, "learning_rate": 1.9245081192782115e-05, "loss": 1.0287, "step": 742 }, { "epoch": 1.5101626016260163, "grad_norm": 1.0361252167647437, "learning_rate": 1.924256984994435e-05, "loss": 0.8339, "step": 743 }, { "epoch": 1.5121951219512195, "grad_norm": 1.3413296494946658, "learning_rate": 1.9240054501285015e-05, "loss": 1.2411, "step": 744 }, { "epoch": 1.5142276422764227, "grad_norm": 1.1236852337170489, "learning_rate": 1.9237535147894283e-05, "loss": 1.0234, "step": 745 }, { "epoch": 1.5162601626016259, "grad_norm": 0.9607332230503219, "learning_rate": 1.9235011790864066e-05, "loss": 0.8482, "step": 746 }, { "epoch": 1.5182926829268293, "grad_norm": 0.9813166841505963, "learning_rate": 1.9232484431288013e-05, "loss": 0.8124, "step": 747 }, { "epoch": 1.5203252032520327, "grad_norm": 1.0527562508960404, "learning_rate": 1.922995307026151e-05, "loss": 0.8786, "step": 748 }, { "epoch": 1.5223577235772359, "grad_norm": 1.111037098897001, "learning_rate": 1.9227417708881672e-05, "loss": 0.8828, "step": 749 }, { "epoch": 1.524390243902439, "grad_norm": 1.064708398715043, "learning_rate": 1.9224878348247343e-05, "loss": 1.0935, "step": 750 }, { "epoch": 1.5264227642276422, "grad_norm": 1.15670941459941, "learning_rate": 1.922233498945912e-05, "loss": 0.9851, "step": 751 }, { "epoch": 1.5284552845528454, "grad_norm": 1.0286800908435323, "learning_rate": 1.921978763361931e-05, "loss": 0.9287, "step": 752 }, { "epoch": 1.5304878048780488, "grad_norm": 1.1495893832226514, "learning_rate": 1.9217236281831963e-05, "loss": 0.8368, "step": 753 }, { "epoch": 1.532520325203252, "grad_norm": 1.2268481101627984, "learning_rate": 1.9214680935202866e-05, "loss": 1.0195, "step": 754 }, { "epoch": 1.5345528455284554, "grad_norm": 0.9924161711075202, "learning_rate": 1.921212159483953e-05, "loss": 0.7649, "step": 755 }, { "epoch": 1.5365853658536586, "grad_norm": 1.2158634645015989, "learning_rate": 1.9209558261851194e-05, "loss": 1.0439, "step": 756 }, { "epoch": 1.5386178861788617, "grad_norm": 1.2160506755965959, "learning_rate": 1.9206990937348835e-05, "loss": 0.8386, "step": 757 }, { "epoch": 1.540650406504065, "grad_norm": 1.1405547568290486, "learning_rate": 1.9204419622445157e-05, "loss": 1.0551, "step": 758 }, { "epoch": 1.5426829268292683, "grad_norm": 1.1566193686301058, "learning_rate": 1.9201844318254595e-05, "loss": 0.9831, "step": 759 }, { "epoch": 1.5447154471544715, "grad_norm": 1.1618170321514478, "learning_rate": 1.919926502589331e-05, "loss": 0.8595, "step": 760 }, { "epoch": 1.546747967479675, "grad_norm": 1.2443876917507757, "learning_rate": 1.919668174647919e-05, "loss": 0.9547, "step": 761 }, { "epoch": 1.548780487804878, "grad_norm": 1.1912178078726239, "learning_rate": 1.9194094481131858e-05, "loss": 1.0762, "step": 762 }, { "epoch": 1.5508130081300813, "grad_norm": 1.0463945551031764, "learning_rate": 1.919150323097266e-05, "loss": 0.6292, "step": 763 }, { "epoch": 1.5528455284552845, "grad_norm": 1.1937972109258725, "learning_rate": 1.9188907997124666e-05, "loss": 0.9759, "step": 764 }, { "epoch": 1.5548780487804879, "grad_norm": 1.1051184816123683, "learning_rate": 1.9186308780712678e-05, "loss": 1.0126, "step": 765 }, { "epoch": 1.556910569105691, "grad_norm": 1.0381991042450514, "learning_rate": 1.9183705582863226e-05, "loss": 0.9572, "step": 766 }, { "epoch": 1.5589430894308944, "grad_norm": 1.2509483959649366, "learning_rate": 1.918109840470456e-05, "loss": 1.0103, "step": 767 }, { "epoch": 1.5609756097560976, "grad_norm": 1.094876896518942, "learning_rate": 1.9178487247366652e-05, "loss": 0.9254, "step": 768 }, { "epoch": 1.5630081300813008, "grad_norm": 1.2368881584508455, "learning_rate": 1.917587211198121e-05, "loss": 0.9304, "step": 769 }, { "epoch": 1.565040650406504, "grad_norm": 1.2510337619941885, "learning_rate": 1.9173252999681654e-05, "loss": 0.901, "step": 770 }, { "epoch": 1.5670731707317072, "grad_norm": 1.1270737435717393, "learning_rate": 1.9170629911603135e-05, "loss": 0.8238, "step": 771 }, { "epoch": 1.5691056910569106, "grad_norm": 1.1919184178918836, "learning_rate": 1.916800284888253e-05, "loss": 0.9403, "step": 772 }, { "epoch": 1.571138211382114, "grad_norm": 1.2414671727004003, "learning_rate": 1.9165371812658428e-05, "loss": 1.0079, "step": 773 }, { "epoch": 1.5731707317073171, "grad_norm": 1.1096966831709332, "learning_rate": 1.9162736804071152e-05, "loss": 1.0377, "step": 774 }, { "epoch": 1.5752032520325203, "grad_norm": 1.225037537067534, "learning_rate": 1.9160097824262735e-05, "loss": 1.074, "step": 775 }, { "epoch": 1.5772357723577235, "grad_norm": 1.3788412915494717, "learning_rate": 1.915745487437694e-05, "loss": 1.0751, "step": 776 }, { "epoch": 1.5792682926829267, "grad_norm": 1.0578507114242872, "learning_rate": 1.9154807955559248e-05, "loss": 0.9446, "step": 777 }, { "epoch": 1.58130081300813, "grad_norm": 1.2482358567301979, "learning_rate": 1.9152157068956863e-05, "loss": 1.0847, "step": 778 }, { "epoch": 1.5833333333333335, "grad_norm": 1.159306282539715, "learning_rate": 1.9149502215718698e-05, "loss": 1.0262, "step": 779 }, { "epoch": 1.5853658536585367, "grad_norm": 1.1680390598680543, "learning_rate": 1.9146843396995396e-05, "loss": 1.1474, "step": 780 }, { "epoch": 1.5873983739837398, "grad_norm": 1.0940226950812795, "learning_rate": 1.9144180613939317e-05, "loss": 0.9471, "step": 781 }, { "epoch": 1.589430894308943, "grad_norm": 1.1068357344292954, "learning_rate": 1.914151386770454e-05, "loss": 1.0332, "step": 782 }, { "epoch": 1.5914634146341462, "grad_norm": 1.1487387020684523, "learning_rate": 1.9138843159446852e-05, "loss": 0.8886, "step": 783 }, { "epoch": 1.5934959349593496, "grad_norm": 1.1537545646796203, "learning_rate": 1.9136168490323772e-05, "loss": 0.8807, "step": 784 }, { "epoch": 1.595528455284553, "grad_norm": 1.034079387261029, "learning_rate": 1.9133489861494523e-05, "loss": 0.7828, "step": 785 }, { "epoch": 1.5975609756097562, "grad_norm": 1.2838146395984527, "learning_rate": 1.913080727412005e-05, "loss": 1.0194, "step": 786 }, { "epoch": 1.5995934959349594, "grad_norm": 1.1758100612910858, "learning_rate": 1.9128120729363017e-05, "loss": 0.9901, "step": 787 }, { "epoch": 1.6016260162601625, "grad_norm": 1.2184559864225222, "learning_rate": 1.9125430228387794e-05, "loss": 0.8812, "step": 788 }, { "epoch": 1.6036585365853657, "grad_norm": 1.5691671095430229, "learning_rate": 1.9122735772360476e-05, "loss": 0.8655, "step": 789 }, { "epoch": 1.6056910569105691, "grad_norm": 1.3578266386264912, "learning_rate": 1.912003736244886e-05, "loss": 1.1502, "step": 790 }, { "epoch": 1.6077235772357723, "grad_norm": 1.3275856052114903, "learning_rate": 1.911733499982247e-05, "loss": 1.1362, "step": 791 }, { "epoch": 1.6097560975609757, "grad_norm": 1.1542451815911996, "learning_rate": 1.9114628685652535e-05, "loss": 0.9057, "step": 792 }, { "epoch": 1.6117886178861789, "grad_norm": 1.221684135563438, "learning_rate": 1.9111918421111996e-05, "loss": 1.0517, "step": 793 }, { "epoch": 1.613821138211382, "grad_norm": 1.1996150667241723, "learning_rate": 1.910920420737551e-05, "loss": 0.9357, "step": 794 }, { "epoch": 1.6158536585365852, "grad_norm": 1.19587390545998, "learning_rate": 1.910648604561945e-05, "loss": 1.005, "step": 795 }, { "epoch": 1.6178861788617886, "grad_norm": 1.1406860671442027, "learning_rate": 1.9103763937021887e-05, "loss": 0.8309, "step": 796 }, { "epoch": 1.6199186991869918, "grad_norm": 1.2397434173674955, "learning_rate": 1.9101037882762613e-05, "loss": 0.9774, "step": 797 }, { "epoch": 1.6219512195121952, "grad_norm": 1.2583453968333393, "learning_rate": 1.9098307884023122e-05, "loss": 0.9241, "step": 798 }, { "epoch": 1.6239837398373984, "grad_norm": 1.1374730564812578, "learning_rate": 1.9095573941986628e-05, "loss": 1.1404, "step": 799 }, { "epoch": 1.6260162601626016, "grad_norm": 1.3019404302999533, "learning_rate": 1.909283605783805e-05, "loss": 0.8861, "step": 800 }, { "epoch": 1.6280487804878048, "grad_norm": 1.0959972860584506, "learning_rate": 1.909009423276401e-05, "loss": 0.8371, "step": 801 }, { "epoch": 1.6300813008130082, "grad_norm": 1.163694111345936, "learning_rate": 1.9087348467952843e-05, "loss": 0.8681, "step": 802 }, { "epoch": 1.6321138211382114, "grad_norm": 1.272385329112214, "learning_rate": 1.908459876459459e-05, "loss": 0.9866, "step": 803 }, { "epoch": 1.6341463414634148, "grad_norm": 1.2833260785098346, "learning_rate": 1.9081845123881002e-05, "loss": 0.9185, "step": 804 }, { "epoch": 1.636178861788618, "grad_norm": 1.3558713778317453, "learning_rate": 1.9079087547005535e-05, "loss": 0.9368, "step": 805 }, { "epoch": 1.6382113821138211, "grad_norm": 1.2406664190418282, "learning_rate": 1.907632603516335e-05, "loss": 0.9657, "step": 806 }, { "epoch": 1.6402439024390243, "grad_norm": 1.124019471454537, "learning_rate": 1.9073560589551307e-05, "loss": 0.9692, "step": 807 }, { "epoch": 1.6422764227642277, "grad_norm": 1.2698137677552594, "learning_rate": 1.9070791211367984e-05, "loss": 0.9503, "step": 808 }, { "epoch": 1.6443089430894309, "grad_norm": 1.152895610644723, "learning_rate": 1.906801790181366e-05, "loss": 0.9588, "step": 809 }, { "epoch": 1.6463414634146343, "grad_norm": 1.2096230644363575, "learning_rate": 1.9065240662090308e-05, "loss": 1.0514, "step": 810 }, { "epoch": 1.6483739837398375, "grad_norm": 1.1389604371863957, "learning_rate": 1.9062459493401615e-05, "loss": 0.9016, "step": 811 }, { "epoch": 1.6504065040650406, "grad_norm": 1.294654144993417, "learning_rate": 1.9059674396952963e-05, "loss": 1.0059, "step": 812 }, { "epoch": 1.6524390243902438, "grad_norm": 1.1320304743789562, "learning_rate": 1.9056885373951446e-05, "loss": 0.8736, "step": 813 }, { "epoch": 1.654471544715447, "grad_norm": 1.1375532319708506, "learning_rate": 1.9054092425605852e-05, "loss": 1.1284, "step": 814 }, { "epoch": 1.6565040650406504, "grad_norm": 1.1738401591224243, "learning_rate": 1.905129555312667e-05, "loss": 0.8658, "step": 815 }, { "epoch": 1.6585365853658538, "grad_norm": 1.199471503177632, "learning_rate": 1.90484947577261e-05, "loss": 0.8175, "step": 816 }, { "epoch": 1.660569105691057, "grad_norm": 1.2415028369551975, "learning_rate": 1.9045690040618024e-05, "loss": 1.0914, "step": 817 }, { "epoch": 1.6626016260162602, "grad_norm": 1.3615379689487086, "learning_rate": 1.9042881403018044e-05, "loss": 1.064, "step": 818 }, { "epoch": 1.6646341463414633, "grad_norm": 1.4991965734955044, "learning_rate": 1.9040068846143444e-05, "loss": 0.9285, "step": 819 }, { "epoch": 1.6666666666666665, "grad_norm": 1.2812866370312144, "learning_rate": 1.903725237121322e-05, "loss": 0.8727, "step": 820 }, { "epoch": 1.66869918699187, "grad_norm": 1.1606130029528967, "learning_rate": 1.9034431979448053e-05, "loss": 1.05, "step": 821 }, { "epoch": 1.6707317073170733, "grad_norm": 1.080206009065487, "learning_rate": 1.903160767207034e-05, "loss": 0.8346, "step": 822 }, { "epoch": 1.6727642276422765, "grad_norm": 1.1255889694194527, "learning_rate": 1.902877945030415e-05, "loss": 0.8859, "step": 823 }, { "epoch": 1.6747967479674797, "grad_norm": 1.2030861976266671, "learning_rate": 1.902594731537527e-05, "loss": 0.8309, "step": 824 }, { "epoch": 1.6768292682926829, "grad_norm": 1.0704878165618659, "learning_rate": 1.902311126851118e-05, "loss": 1.0136, "step": 825 }, { "epoch": 1.678861788617886, "grad_norm": 1.38718671438824, "learning_rate": 1.902027131094104e-05, "loss": 1.064, "step": 826 }, { "epoch": 1.6808943089430894, "grad_norm": 1.265189034705007, "learning_rate": 1.9017427443895723e-05, "loss": 0.9452, "step": 827 }, { "epoch": 1.6829268292682928, "grad_norm": 1.282156786472457, "learning_rate": 1.901457966860779e-05, "loss": 0.7882, "step": 828 }, { "epoch": 1.684959349593496, "grad_norm": 1.2196070599337463, "learning_rate": 1.9011727986311486e-05, "loss": 1.1154, "step": 829 }, { "epoch": 1.6869918699186992, "grad_norm": 1.1667627841662978, "learning_rate": 1.9008872398242776e-05, "loss": 0.8454, "step": 830 }, { "epoch": 1.6890243902439024, "grad_norm": 1.3375605347860098, "learning_rate": 1.9006012905639282e-05, "loss": 1.15, "step": 831 }, { "epoch": 1.6910569105691056, "grad_norm": 1.266230994178795, "learning_rate": 1.9003149509740347e-05, "loss": 1.0467, "step": 832 }, { "epoch": 1.693089430894309, "grad_norm": 1.16295552599261, "learning_rate": 1.9000282211786988e-05, "loss": 0.9168, "step": 833 }, { "epoch": 1.6951219512195121, "grad_norm": 1.2054324420662588, "learning_rate": 1.8997411013021928e-05, "loss": 1.0672, "step": 834 }, { "epoch": 1.6971544715447155, "grad_norm": 1.089537596831206, "learning_rate": 1.899453591468957e-05, "loss": 0.9558, "step": 835 }, { "epoch": 1.6991869918699187, "grad_norm": 1.0786343708268638, "learning_rate": 1.899165691803601e-05, "loss": 0.7901, "step": 836 }, { "epoch": 1.701219512195122, "grad_norm": 1.25640549491399, "learning_rate": 1.898877402430903e-05, "loss": 0.8744, "step": 837 }, { "epoch": 1.703252032520325, "grad_norm": 1.1623533683893847, "learning_rate": 1.898588723475811e-05, "loss": 1.1085, "step": 838 }, { "epoch": 1.7052845528455285, "grad_norm": 1.2708478938357488, "learning_rate": 1.898299655063441e-05, "loss": 0.9496, "step": 839 }, { "epoch": 1.7073170731707317, "grad_norm": 1.217590177530208, "learning_rate": 1.8980101973190787e-05, "loss": 0.9255, "step": 840 }, { "epoch": 1.709349593495935, "grad_norm": 1.3249325208800697, "learning_rate": 1.8977203503681772e-05, "loss": 0.8972, "step": 841 }, { "epoch": 1.7113821138211383, "grad_norm": 1.1992754980306088, "learning_rate": 1.8974301143363598e-05, "loss": 1.026, "step": 842 }, { "epoch": 1.7134146341463414, "grad_norm": 1.1596368404486401, "learning_rate": 1.897139489349417e-05, "loss": 0.9308, "step": 843 }, { "epoch": 1.7154471544715446, "grad_norm": 1.1480723117731568, "learning_rate": 1.896848475533309e-05, "loss": 0.9204, "step": 844 }, { "epoch": 1.717479674796748, "grad_norm": 1.1162185562567155, "learning_rate": 1.8965570730141636e-05, "loss": 0.8969, "step": 845 }, { "epoch": 1.7195121951219512, "grad_norm": 1.2651606749157265, "learning_rate": 1.8962652819182778e-05, "loss": 0.9312, "step": 846 }, { "epoch": 1.7215447154471546, "grad_norm": 1.1972418469548767, "learning_rate": 1.895973102372117e-05, "loss": 0.9262, "step": 847 }, { "epoch": 1.7235772357723578, "grad_norm": 1.3182898873874023, "learning_rate": 1.8956805345023145e-05, "loss": 0.9502, "step": 848 }, { "epoch": 1.725609756097561, "grad_norm": 1.204716087437911, "learning_rate": 1.8953875784356724e-05, "loss": 0.9035, "step": 849 }, { "epoch": 1.7276422764227641, "grad_norm": 1.235540640621242, "learning_rate": 1.89509423429916e-05, "loss": 0.9373, "step": 850 }, { "epoch": 1.7296747967479673, "grad_norm": 1.267761937653081, "learning_rate": 1.8948005022199164e-05, "loss": 1.0488, "step": 851 }, { "epoch": 1.7317073170731707, "grad_norm": 1.122453111574252, "learning_rate": 1.894506382325248e-05, "loss": 0.8328, "step": 852 }, { "epoch": 1.7337398373983741, "grad_norm": 1.205726552234568, "learning_rate": 1.8942118747426283e-05, "loss": 0.8623, "step": 853 }, { "epoch": 1.7357723577235773, "grad_norm": 1.1656496181382452, "learning_rate": 1.8939169795997008e-05, "loss": 0.9254, "step": 854 }, { "epoch": 1.7378048780487805, "grad_norm": 1.2339628407028638, "learning_rate": 1.8936216970242757e-05, "loss": 0.9651, "step": 855 }, { "epoch": 1.7398373983739837, "grad_norm": 1.1929463946557908, "learning_rate": 1.8933260271443313e-05, "loss": 0.8411, "step": 856 }, { "epoch": 1.7418699186991868, "grad_norm": 1.2971515873686552, "learning_rate": 1.8930299700880138e-05, "loss": 0.7746, "step": 857 }, { "epoch": 1.7439024390243902, "grad_norm": 1.1701366239635334, "learning_rate": 1.8927335259836376e-05, "loss": 0.9913, "step": 858 }, { "epoch": 1.7459349593495936, "grad_norm": 1.3228782270857706, "learning_rate": 1.892436694959684e-05, "loss": 0.9191, "step": 859 }, { "epoch": 1.7479674796747968, "grad_norm": 1.3176263753142274, "learning_rate": 1.8921394771448032e-05, "loss": 1.0559, "step": 860 }, { "epoch": 1.75, "grad_norm": 1.1541033230574347, "learning_rate": 1.8918418726678118e-05, "loss": 0.9983, "step": 861 }, { "epoch": 1.7520325203252032, "grad_norm": 1.288924926533074, "learning_rate": 1.8915438816576945e-05, "loss": 1.0271, "step": 862 }, { "epoch": 1.7540650406504064, "grad_norm": 1.3753001058218326, "learning_rate": 1.8912455042436038e-05, "loss": 0.9792, "step": 863 }, { "epoch": 1.7560975609756098, "grad_norm": 1.1560448415244677, "learning_rate": 1.89094674055486e-05, "loss": 0.8342, "step": 864 }, { "epoch": 1.7581300813008132, "grad_norm": 1.2078938397962165, "learning_rate": 1.8906475907209493e-05, "loss": 1.0108, "step": 865 }, { "epoch": 1.7601626016260163, "grad_norm": 1.1252438644427005, "learning_rate": 1.8903480548715263e-05, "loss": 0.8591, "step": 866 }, { "epoch": 1.7621951219512195, "grad_norm": 1.251058760784179, "learning_rate": 1.8900481331364132e-05, "loss": 0.8629, "step": 867 }, { "epoch": 1.7642276422764227, "grad_norm": 1.2906408673823884, "learning_rate": 1.889747825645599e-05, "loss": 1.001, "step": 868 }, { "epoch": 1.7662601626016259, "grad_norm": 1.32749715145804, "learning_rate": 1.88944713252924e-05, "loss": 1.0028, "step": 869 }, { "epoch": 1.7682926829268293, "grad_norm": 1.2358771679713556, "learning_rate": 1.8891460539176597e-05, "loss": 0.8072, "step": 870 }, { "epoch": 1.7703252032520327, "grad_norm": 1.2533352659621375, "learning_rate": 1.888844589941348e-05, "loss": 0.8601, "step": 871 }, { "epoch": 1.7723577235772359, "grad_norm": 1.2248830442434901, "learning_rate": 1.8885427407309627e-05, "loss": 0.9429, "step": 872 }, { "epoch": 1.774390243902439, "grad_norm": 1.6511761611786544, "learning_rate": 1.888240506417328e-05, "loss": 1.1255, "step": 873 }, { "epoch": 1.7764227642276422, "grad_norm": 1.1222439614387971, "learning_rate": 1.887937887131436e-05, "loss": 0.768, "step": 874 }, { "epoch": 1.7784552845528454, "grad_norm": 1.3589295417362288, "learning_rate": 1.8876348830044443e-05, "loss": 1.0101, "step": 875 }, { "epoch": 1.7804878048780488, "grad_norm": 1.1886089320747426, "learning_rate": 1.887331494167678e-05, "loss": 0.9389, "step": 876 }, { "epoch": 1.782520325203252, "grad_norm": 1.4537617759022587, "learning_rate": 1.8870277207526286e-05, "loss": 1.0051, "step": 877 }, { "epoch": 1.7845528455284554, "grad_norm": 1.178042056376556, "learning_rate": 1.8867235628909553e-05, "loss": 0.8019, "step": 878 }, { "epoch": 1.7865853658536586, "grad_norm": 1.2181378405783834, "learning_rate": 1.8864190207144822e-05, "loss": 0.7901, "step": 879 }, { "epoch": 1.7886178861788617, "grad_norm": 1.2962742074334026, "learning_rate": 1.8861140943552014e-05, "loss": 1.0022, "step": 880 }, { "epoch": 1.790650406504065, "grad_norm": 1.216073286817628, "learning_rate": 1.8858087839452706e-05, "loss": 0.9979, "step": 881 }, { "epoch": 1.7926829268292683, "grad_norm": 1.161835411378759, "learning_rate": 1.8855030896170152e-05, "loss": 0.85, "step": 882 }, { "epoch": 1.7947154471544715, "grad_norm": 1.4231825703448275, "learning_rate": 1.8851970115029256e-05, "loss": 0.9809, "step": 883 }, { "epoch": 1.796747967479675, "grad_norm": 1.2659040496129021, "learning_rate": 1.884890549735659e-05, "loss": 0.9435, "step": 884 }, { "epoch": 1.798780487804878, "grad_norm": 1.2077096978766293, "learning_rate": 1.884583704448039e-05, "loss": 0.87, "step": 885 }, { "epoch": 1.8008130081300813, "grad_norm": 1.1418230026311411, "learning_rate": 1.8842764757730555e-05, "loss": 0.8865, "step": 886 }, { "epoch": 1.8028455284552845, "grad_norm": 1.3055427079657604, "learning_rate": 1.8839688638438645e-05, "loss": 0.9888, "step": 887 }, { "epoch": 1.8048780487804879, "grad_norm": 1.1564609472304461, "learning_rate": 1.8836608687937883e-05, "loss": 0.8439, "step": 888 }, { "epoch": 1.806910569105691, "grad_norm": 1.190699199490442, "learning_rate": 1.8833524907563146e-05, "loss": 0.987, "step": 889 }, { "epoch": 1.8089430894308944, "grad_norm": 1.2141644296718352, "learning_rate": 1.8830437298650976e-05, "loss": 0.7641, "step": 890 }, { "epoch": 1.8109756097560976, "grad_norm": 1.3477248714371663, "learning_rate": 1.882734586253957e-05, "loss": 1.1374, "step": 891 }, { "epoch": 1.8130081300813008, "grad_norm": 1.3162023632094837, "learning_rate": 1.8824250600568798e-05, "loss": 1.1128, "step": 892 }, { "epoch": 1.815040650406504, "grad_norm": 1.2071419400582744, "learning_rate": 1.8821151514080167e-05, "loss": 0.8885, "step": 893 }, { "epoch": 1.8170731707317072, "grad_norm": 1.0925865442608562, "learning_rate": 1.8818048604416856e-05, "loss": 0.8013, "step": 894 }, { "epoch": 1.8191056910569106, "grad_norm": 1.2043525208386396, "learning_rate": 1.8814941872923696e-05, "loss": 0.7962, "step": 895 }, { "epoch": 1.821138211382114, "grad_norm": 1.5543970982867878, "learning_rate": 1.8811831320947177e-05, "loss": 1.1399, "step": 896 }, { "epoch": 1.8231707317073171, "grad_norm": 1.028744716503267, "learning_rate": 1.8808716949835437e-05, "loss": 0.6664, "step": 897 }, { "epoch": 1.8252032520325203, "grad_norm": 1.2027572374589375, "learning_rate": 1.8805598760938282e-05, "loss": 1.0266, "step": 898 }, { "epoch": 1.8272357723577235, "grad_norm": 1.18795029591468, "learning_rate": 1.880247675560716e-05, "loss": 1.0048, "step": 899 }, { "epoch": 1.8292682926829267, "grad_norm": 1.238543428508024, "learning_rate": 1.879935093519519e-05, "loss": 0.9114, "step": 900 }, { "epoch": 1.83130081300813, "grad_norm": 1.1281186130013956, "learning_rate": 1.8796221301057125e-05, "loss": 0.8914, "step": 901 }, { "epoch": 1.8333333333333335, "grad_norm": 1.2124365716092875, "learning_rate": 1.8793087854549378e-05, "loss": 1.0503, "step": 902 }, { "epoch": 1.8353658536585367, "grad_norm": 1.369876334601961, "learning_rate": 1.878995059703002e-05, "loss": 0.8278, "step": 903 }, { "epoch": 1.8373983739837398, "grad_norm": 1.097431938616776, "learning_rate": 1.878680952985877e-05, "loss": 0.9011, "step": 904 }, { "epoch": 1.839430894308943, "grad_norm": 1.3573976043311473, "learning_rate": 1.8783664654396997e-05, "loss": 0.8267, "step": 905 }, { "epoch": 1.8414634146341462, "grad_norm": 1.2073290100046041, "learning_rate": 1.8780515972007717e-05, "loss": 1.0286, "step": 906 }, { "epoch": 1.8434959349593496, "grad_norm": 1.2109112325231328, "learning_rate": 1.8777363484055602e-05, "loss": 0.8503, "step": 907 }, { "epoch": 1.845528455284553, "grad_norm": 1.4113369174370944, "learning_rate": 1.8774207191906976e-05, "loss": 0.8396, "step": 908 }, { "epoch": 1.8475609756097562, "grad_norm": 1.4155911222912296, "learning_rate": 1.8771047096929808e-05, "loss": 0.891, "step": 909 }, { "epoch": 1.8495934959349594, "grad_norm": 1.3840437721804986, "learning_rate": 1.8767883200493705e-05, "loss": 0.9667, "step": 910 }, { "epoch": 1.8516260162601625, "grad_norm": 1.3710599970880568, "learning_rate": 1.8764715503969944e-05, "loss": 0.9553, "step": 911 }, { "epoch": 1.8536585365853657, "grad_norm": 1.367217760621146, "learning_rate": 1.8761544008731426e-05, "loss": 1.0331, "step": 912 }, { "epoch": 1.8556910569105691, "grad_norm": 1.421368729029175, "learning_rate": 1.8758368716152712e-05, "loss": 0.9663, "step": 913 }, { "epoch": 1.8577235772357723, "grad_norm": 1.3595165187132696, "learning_rate": 1.8755189627610004e-05, "loss": 1.0049, "step": 914 }, { "epoch": 1.8597560975609757, "grad_norm": 1.3250645020180385, "learning_rate": 1.8752006744481153e-05, "loss": 1.0246, "step": 915 }, { "epoch": 1.8617886178861789, "grad_norm": 1.501488258447364, "learning_rate": 1.874882006814565e-05, "loss": 0.8256, "step": 916 }, { "epoch": 1.863821138211382, "grad_norm": 1.244918421739128, "learning_rate": 1.8745629599984636e-05, "loss": 0.9235, "step": 917 }, { "epoch": 1.8658536585365852, "grad_norm": 1.3096669084113624, "learning_rate": 1.874243534138089e-05, "loss": 1.0135, "step": 918 }, { "epoch": 1.8678861788617886, "grad_norm": 1.4683245258157782, "learning_rate": 1.8739237293718832e-05, "loss": 1.0954, "step": 919 }, { "epoch": 1.8699186991869918, "grad_norm": 1.5629517489065625, "learning_rate": 1.8736035458384528e-05, "loss": 1.0264, "step": 920 }, { "epoch": 1.8719512195121952, "grad_norm": 1.462971281406613, "learning_rate": 1.873282983676569e-05, "loss": 0.9583, "step": 921 }, { "epoch": 1.8739837398373984, "grad_norm": 1.2636511558456152, "learning_rate": 1.8729620430251662e-05, "loss": 0.9908, "step": 922 }, { "epoch": 1.8760162601626016, "grad_norm": 1.2037687590595072, "learning_rate": 1.872640724023343e-05, "loss": 0.9068, "step": 923 }, { "epoch": 1.8780487804878048, "grad_norm": 1.3689577567922973, "learning_rate": 1.8723190268103634e-05, "loss": 0.8641, "step": 924 }, { "epoch": 1.8800813008130082, "grad_norm": 1.349054254567135, "learning_rate": 1.8719969515256527e-05, "loss": 1.0016, "step": 925 }, { "epoch": 1.8821138211382114, "grad_norm": 1.2709905307074487, "learning_rate": 1.8716744983088028e-05, "loss": 1.1029, "step": 926 }, { "epoch": 1.8841463414634148, "grad_norm": 1.5234712867642315, "learning_rate": 1.8713516672995676e-05, "loss": 1.0098, "step": 927 }, { "epoch": 1.886178861788618, "grad_norm": 1.2619139386111011, "learning_rate": 1.8710284586378645e-05, "loss": 0.9131, "step": 928 }, { "epoch": 1.8882113821138211, "grad_norm": 1.2249342832785213, "learning_rate": 1.8707048724637767e-05, "loss": 0.9839, "step": 929 }, { "epoch": 1.8902439024390243, "grad_norm": 1.3244770398912464, "learning_rate": 1.870380908917549e-05, "loss": 1.0702, "step": 930 }, { "epoch": 1.8922764227642277, "grad_norm": 1.167221172264963, "learning_rate": 1.8700565681395905e-05, "loss": 0.9473, "step": 931 }, { "epoch": 1.8943089430894309, "grad_norm": 1.4886052494653303, "learning_rate": 1.8697318502704734e-05, "loss": 0.9977, "step": 932 }, { "epoch": 1.8963414634146343, "grad_norm": 1.164852567130357, "learning_rate": 1.869406755450934e-05, "loss": 0.8249, "step": 933 }, { "epoch": 1.8983739837398375, "grad_norm": 1.1697028069439874, "learning_rate": 1.869081283821872e-05, "loss": 0.9318, "step": 934 }, { "epoch": 1.9004065040650406, "grad_norm": 1.2691774569484497, "learning_rate": 1.8687554355243488e-05, "loss": 0.9797, "step": 935 }, { "epoch": 1.9024390243902438, "grad_norm": 1.2304272007434736, "learning_rate": 1.8684292106995916e-05, "loss": 1.0467, "step": 936 }, { "epoch": 1.904471544715447, "grad_norm": 1.246059914802852, "learning_rate": 1.868102609488989e-05, "loss": 0.9941, "step": 937 }, { "epoch": 1.9065040650406504, "grad_norm": 1.2622800743052902, "learning_rate": 1.8677756320340927e-05, "loss": 1.0686, "step": 938 }, { "epoch": 1.9085365853658538, "grad_norm": 1.487458811253934, "learning_rate": 1.867448278476619e-05, "loss": 0.9309, "step": 939 }, { "epoch": 1.910569105691057, "grad_norm": 1.4934782656769383, "learning_rate": 1.8671205489584453e-05, "loss": 1.1807, "step": 940 }, { "epoch": 1.9126016260162602, "grad_norm": 1.2716512970058351, "learning_rate": 1.866792443621613e-05, "loss": 0.9894, "step": 941 }, { "epoch": 1.9146341463414633, "grad_norm": 1.3960049230928435, "learning_rate": 1.8664639626083265e-05, "loss": 0.9246, "step": 942 }, { "epoch": 1.9166666666666665, "grad_norm": 1.2927585147969496, "learning_rate": 1.866135106060953e-05, "loss": 0.7901, "step": 943 }, { "epoch": 1.91869918699187, "grad_norm": 1.3766022820320754, "learning_rate": 1.865805874122021e-05, "loss": 0.8386, "step": 944 }, { "epoch": 1.9207317073170733, "grad_norm": 1.1421918626334568, "learning_rate": 1.8654762669342242e-05, "loss": 0.8309, "step": 945 }, { "epoch": 1.9227642276422765, "grad_norm": 1.413057204483458, "learning_rate": 1.8651462846404173e-05, "loss": 0.8662, "step": 946 }, { "epoch": 1.9247967479674797, "grad_norm": 1.3632106535904092, "learning_rate": 1.8648159273836174e-05, "loss": 0.9517, "step": 947 }, { "epoch": 1.9268292682926829, "grad_norm": 1.2198872884141134, "learning_rate": 1.8644851953070045e-05, "loss": 0.942, "step": 948 }, { "epoch": 1.928861788617886, "grad_norm": 1.2810358830759332, "learning_rate": 1.864154088553922e-05, "loss": 1.1012, "step": 949 }, { "epoch": 1.9308943089430894, "grad_norm": 1.1135262591560668, "learning_rate": 1.8638226072678746e-05, "loss": 0.7727, "step": 950 }, { "epoch": 1.9329268292682928, "grad_norm": 1.1862761657620051, "learning_rate": 1.8634907515925294e-05, "loss": 1.0686, "step": 951 }, { "epoch": 1.934959349593496, "grad_norm": 1.3003263654085992, "learning_rate": 1.863158521671716e-05, "loss": 0.8975, "step": 952 }, { "epoch": 1.9369918699186992, "grad_norm": 1.2897713430980464, "learning_rate": 1.862825917649426e-05, "loss": 0.9594, "step": 953 }, { "epoch": 1.9390243902439024, "grad_norm": 1.3890051655147155, "learning_rate": 1.8624929396698138e-05, "loss": 1.0883, "step": 954 }, { "epoch": 1.9410569105691056, "grad_norm": 1.1674997337727862, "learning_rate": 1.8621595878771944e-05, "loss": 0.9489, "step": 955 }, { "epoch": 1.943089430894309, "grad_norm": 1.2726758343394586, "learning_rate": 1.8618258624160465e-05, "loss": 0.8871, "step": 956 }, { "epoch": 1.9451219512195121, "grad_norm": 1.2338091961216646, "learning_rate": 1.8614917634310103e-05, "loss": 0.935, "step": 957 }, { "epoch": 1.9471544715447155, "grad_norm": 1.1479037820616147, "learning_rate": 1.8611572910668866e-05, "loss": 0.8446, "step": 958 }, { "epoch": 1.9491869918699187, "grad_norm": 1.145557127793306, "learning_rate": 1.86082244546864e-05, "loss": 0.8884, "step": 959 }, { "epoch": 1.951219512195122, "grad_norm": 1.2015311755896472, "learning_rate": 1.8604872267813954e-05, "loss": 0.8401, "step": 960 }, { "epoch": 1.953252032520325, "grad_norm": 1.4013334779904247, "learning_rate": 1.86015163515044e-05, "loss": 1.0397, "step": 961 }, { "epoch": 1.9552845528455285, "grad_norm": 1.3043967366556026, "learning_rate": 1.859815670721223e-05, "loss": 0.8687, "step": 962 }, { "epoch": 1.9573170731707317, "grad_norm": 1.4315980705469857, "learning_rate": 1.859479333639354e-05, "loss": 1.2383, "step": 963 }, { "epoch": 1.959349593495935, "grad_norm": 1.2670740457190888, "learning_rate": 1.859142624050605e-05, "loss": 0.797, "step": 964 }, { "epoch": 1.9613821138211383, "grad_norm": 1.3093742824255525, "learning_rate": 1.85880554210091e-05, "loss": 0.7158, "step": 965 }, { "epoch": 1.9634146341463414, "grad_norm": 1.3137566962886087, "learning_rate": 1.8584680879363626e-05, "loss": 1.1175, "step": 966 }, { "epoch": 1.9654471544715446, "grad_norm": 1.2649798013739673, "learning_rate": 1.8581302617032196e-05, "loss": 0.9345, "step": 967 }, { "epoch": 1.967479674796748, "grad_norm": 1.4127971070560819, "learning_rate": 1.8577920635478976e-05, "loss": 1.2753, "step": 968 }, { "epoch": 1.9695121951219512, "grad_norm": 1.1592072848008426, "learning_rate": 1.8574534936169758e-05, "loss": 0.9858, "step": 969 }, { "epoch": 1.9715447154471546, "grad_norm": 1.2343932258200556, "learning_rate": 1.857114552057193e-05, "loss": 0.9302, "step": 970 }, { "epoch": 1.9735772357723578, "grad_norm": 1.2135576974009623, "learning_rate": 1.8567752390154506e-05, "loss": 0.8016, "step": 971 }, { "epoch": 1.975609756097561, "grad_norm": 1.3013469759158633, "learning_rate": 1.8564355546388094e-05, "loss": 0.9992, "step": 972 }, { "epoch": 1.9776422764227641, "grad_norm": 1.263806863011205, "learning_rate": 1.8560954990744926e-05, "loss": 0.9207, "step": 973 }, { "epoch": 1.9796747967479673, "grad_norm": 1.6262776634440501, "learning_rate": 1.8557550724698835e-05, "loss": 1.2201, "step": 974 }, { "epoch": 1.9817073170731707, "grad_norm": 1.274457739909222, "learning_rate": 1.8554142749725264e-05, "loss": 0.9491, "step": 975 }, { "epoch": 1.9837398373983741, "grad_norm": 1.3271028014617434, "learning_rate": 1.855073106730126e-05, "loss": 0.8712, "step": 976 }, { "epoch": 1.9857723577235773, "grad_norm": 1.3969502103495006, "learning_rate": 1.854731567890548e-05, "loss": 1.024, "step": 977 }, { "epoch": 1.9878048780487805, "grad_norm": 1.3112594476253963, "learning_rate": 1.85438965860182e-05, "loss": 1.1375, "step": 978 }, { "epoch": 1.9898373983739837, "grad_norm": 1.4624152761237306, "learning_rate": 1.854047379012127e-05, "loss": 1.0088, "step": 979 }, { "epoch": 1.9918699186991868, "grad_norm": 1.2498180813962336, "learning_rate": 1.8537047292698175e-05, "loss": 0.8426, "step": 980 }, { "epoch": 1.9939024390243902, "grad_norm": 1.2250031610393444, "learning_rate": 1.8533617095233993e-05, "loss": 0.8153, "step": 981 }, { "epoch": 1.9959349593495936, "grad_norm": 1.4039478518108948, "learning_rate": 1.85301831992154e-05, "loss": 0.889, "step": 982 }, { "epoch": 1.9979674796747968, "grad_norm": 1.381278945320266, "learning_rate": 1.8526745606130685e-05, "loss": 0.8871, "step": 983 }, { "epoch": 2.0, "grad_norm": 1.5193217742124172, "learning_rate": 1.852330431746973e-05, "loss": 0.9692, "step": 984 }, { "epoch": 2.0, "eval_loss": 0.9892057180404663, "eval_runtime": 132.4937, "eval_samples_per_second": 7.427, "eval_steps_per_second": 0.928, "step": 984 }, { "epoch": 2.002032520325203, "grad_norm": 1.2272851208911684, "learning_rate": 1.8519859334724036e-05, "loss": 0.9103, "step": 985 }, { "epoch": 2.0040650406504064, "grad_norm": 1.3625914730364137, "learning_rate": 1.8516410659386677e-05, "loss": 0.8232, "step": 986 }, { "epoch": 2.0060975609756095, "grad_norm": 1.2586543160468813, "learning_rate": 1.8512958292952354e-05, "loss": 1.0493, "step": 987 }, { "epoch": 2.008130081300813, "grad_norm": 1.109971107419663, "learning_rate": 1.8509502236917353e-05, "loss": 0.7991, "step": 988 }, { "epoch": 2.0101626016260163, "grad_norm": 1.3860666217741207, "learning_rate": 1.850604249277956e-05, "loss": 1.0413, "step": 989 }, { "epoch": 2.0121951219512195, "grad_norm": 1.2653109352924414, "learning_rate": 1.850257906203847e-05, "loss": 0.7125, "step": 990 }, { "epoch": 2.0142276422764227, "grad_norm": 1.290440747430027, "learning_rate": 1.8499111946195163e-05, "loss": 0.8034, "step": 991 }, { "epoch": 2.016260162601626, "grad_norm": 1.3959470604965913, "learning_rate": 1.8495641146752322e-05, "loss": 0.8645, "step": 992 }, { "epoch": 2.018292682926829, "grad_norm": 1.2140498615848396, "learning_rate": 1.8492166665214224e-05, "loss": 0.8532, "step": 993 }, { "epoch": 2.0203252032520327, "grad_norm": 1.2643381755959746, "learning_rate": 1.8488688503086747e-05, "loss": 0.9758, "step": 994 }, { "epoch": 2.022357723577236, "grad_norm": 1.2891453552147862, "learning_rate": 1.8485206661877363e-05, "loss": 1.0529, "step": 995 }, { "epoch": 2.024390243902439, "grad_norm": 1.35604222304764, "learning_rate": 1.848172114309513e-05, "loss": 0.8124, "step": 996 }, { "epoch": 2.0264227642276422, "grad_norm": 1.321579669639766, "learning_rate": 1.8478231948250715e-05, "loss": 0.9774, "step": 997 }, { "epoch": 2.0284552845528454, "grad_norm": 1.5141947077893128, "learning_rate": 1.847473907885636e-05, "loss": 0.8751, "step": 998 }, { "epoch": 2.0304878048780486, "grad_norm": 1.372709844860798, "learning_rate": 1.847124253642592e-05, "loss": 0.9516, "step": 999 }, { "epoch": 2.032520325203252, "grad_norm": 1.338603056921872, "learning_rate": 1.8467742322474822e-05, "loss": 1.0163, "step": 1000 }, { "epoch": 2.0345528455284554, "grad_norm": 1.1889521314538796, "learning_rate": 1.8464238438520096e-05, "loss": 0.7503, "step": 1001 }, { "epoch": 2.0365853658536586, "grad_norm": 1.391641196474358, "learning_rate": 1.8460730886080363e-05, "loss": 0.8099, "step": 1002 }, { "epoch": 2.0386178861788617, "grad_norm": 1.366138758790902, "learning_rate": 1.8457219666675826e-05, "loss": 1.0441, "step": 1003 }, { "epoch": 2.040650406504065, "grad_norm": 1.405572248721484, "learning_rate": 1.845370478182829e-05, "loss": 0.919, "step": 1004 }, { "epoch": 2.042682926829268, "grad_norm": 1.4509835305893626, "learning_rate": 1.8450186233061136e-05, "loss": 1.1183, "step": 1005 }, { "epoch": 2.0447154471544717, "grad_norm": 1.2976576238286457, "learning_rate": 1.8446664021899338e-05, "loss": 0.8405, "step": 1006 }, { "epoch": 2.046747967479675, "grad_norm": 1.3550760071082746, "learning_rate": 1.8443138149869453e-05, "loss": 0.8482, "step": 1007 }, { "epoch": 2.048780487804878, "grad_norm": 1.4364208138476842, "learning_rate": 1.8439608618499637e-05, "loss": 1.0704, "step": 1008 }, { "epoch": 2.0508130081300813, "grad_norm": 1.2663979704939454, "learning_rate": 1.843607542931962e-05, "loss": 0.7404, "step": 1009 }, { "epoch": 2.0528455284552845, "grad_norm": 1.7450114243036032, "learning_rate": 1.8432538583860723e-05, "loss": 0.9502, "step": 1010 }, { "epoch": 2.0548780487804876, "grad_norm": 1.5804454731953854, "learning_rate": 1.842899808365585e-05, "loss": 0.8464, "step": 1011 }, { "epoch": 2.0569105691056913, "grad_norm": 1.4066763837817862, "learning_rate": 1.842545393023949e-05, "loss": 0.7933, "step": 1012 }, { "epoch": 2.0589430894308944, "grad_norm": 1.3497813389395974, "learning_rate": 1.842190612514771e-05, "loss": 0.975, "step": 1013 }, { "epoch": 2.0609756097560976, "grad_norm": 1.411685080010567, "learning_rate": 1.8418354669918162e-05, "loss": 0.8922, "step": 1014 }, { "epoch": 2.063008130081301, "grad_norm": 1.3821615433409349, "learning_rate": 1.8414799566090092e-05, "loss": 0.8348, "step": 1015 }, { "epoch": 2.065040650406504, "grad_norm": 1.2733929640137822, "learning_rate": 1.841124081520431e-05, "loss": 0.9589, "step": 1016 }, { "epoch": 2.067073170731707, "grad_norm": 1.4897140221578018, "learning_rate": 1.8407678418803213e-05, "loss": 0.8424, "step": 1017 }, { "epoch": 2.069105691056911, "grad_norm": 1.433997880195951, "learning_rate": 1.8404112378430782e-05, "loss": 0.8711, "step": 1018 }, { "epoch": 2.071138211382114, "grad_norm": 1.4612020851783867, "learning_rate": 1.8400542695632573e-05, "loss": 1.1675, "step": 1019 }, { "epoch": 2.073170731707317, "grad_norm": 1.408820376055267, "learning_rate": 1.8396969371955724e-05, "loss": 0.9084, "step": 1020 }, { "epoch": 2.0752032520325203, "grad_norm": 1.5573646967279073, "learning_rate": 1.8393392408948947e-05, "loss": 0.8413, "step": 1021 }, { "epoch": 2.0772357723577235, "grad_norm": 1.369508902407916, "learning_rate": 1.8389811808162536e-05, "loss": 0.7823, "step": 1022 }, { "epoch": 2.0792682926829267, "grad_norm": 1.3695756273498223, "learning_rate": 1.8386227571148354e-05, "loss": 0.9004, "step": 1023 }, { "epoch": 2.08130081300813, "grad_norm": 1.3843441568867605, "learning_rate": 1.838263969945985e-05, "loss": 0.8687, "step": 1024 }, { "epoch": 2.0833333333333335, "grad_norm": 1.4341898197555696, "learning_rate": 1.8379048194652043e-05, "loss": 0.8205, "step": 1025 }, { "epoch": 2.0853658536585367, "grad_norm": 1.3889352730643716, "learning_rate": 1.837545305828152e-05, "loss": 0.7812, "step": 1026 }, { "epoch": 2.08739837398374, "grad_norm": 1.309418767127143, "learning_rate": 1.8371854291906456e-05, "loss": 0.8729, "step": 1027 }, { "epoch": 2.089430894308943, "grad_norm": 1.283750952296138, "learning_rate": 1.836825189708659e-05, "loss": 0.8761, "step": 1028 }, { "epoch": 2.091463414634146, "grad_norm": 1.4735005737923805, "learning_rate": 1.836464587538324e-05, "loss": 0.801, "step": 1029 }, { "epoch": 2.0934959349593494, "grad_norm": 1.4490216782744312, "learning_rate": 1.836103622835928e-05, "loss": 0.9165, "step": 1030 }, { "epoch": 2.095528455284553, "grad_norm": 1.4557761040880843, "learning_rate": 1.8357422957579175e-05, "loss": 1.0351, "step": 1031 }, { "epoch": 2.097560975609756, "grad_norm": 1.53158437468583, "learning_rate": 1.8353806064608953e-05, "loss": 0.7769, "step": 1032 }, { "epoch": 2.0995934959349594, "grad_norm": 1.4321661338566187, "learning_rate": 1.835018555101621e-05, "loss": 0.8527, "step": 1033 }, { "epoch": 2.1016260162601625, "grad_norm": 1.4188390927202965, "learning_rate": 1.8346561418370113e-05, "loss": 1.008, "step": 1034 }, { "epoch": 2.1036585365853657, "grad_norm": 1.399631473912468, "learning_rate": 1.8342933668241392e-05, "loss": 0.8177, "step": 1035 }, { "epoch": 2.105691056910569, "grad_norm": 1.4462044765470472, "learning_rate": 1.833930230220236e-05, "loss": 1.0317, "step": 1036 }, { "epoch": 2.1077235772357725, "grad_norm": 1.5441367423423746, "learning_rate": 1.833566732182688e-05, "loss": 0.6832, "step": 1037 }, { "epoch": 2.1097560975609757, "grad_norm": 1.287676990945096, "learning_rate": 1.833202872869039e-05, "loss": 0.8528, "step": 1038 }, { "epoch": 2.111788617886179, "grad_norm": 1.3202174622920304, "learning_rate": 1.8328386524369892e-05, "loss": 0.762, "step": 1039 }, { "epoch": 2.113821138211382, "grad_norm": 1.3396399548956057, "learning_rate": 1.8324740710443955e-05, "loss": 0.8592, "step": 1040 }, { "epoch": 2.1158536585365852, "grad_norm": 1.63431932748599, "learning_rate": 1.832109128849271e-05, "loss": 0.9012, "step": 1041 }, { "epoch": 2.1178861788617884, "grad_norm": 1.5462035165992207, "learning_rate": 1.8317438260097847e-05, "loss": 0.9632, "step": 1042 }, { "epoch": 2.119918699186992, "grad_norm": 1.3749212315357866, "learning_rate": 1.8313781626842635e-05, "loss": 0.8397, "step": 1043 }, { "epoch": 2.1219512195121952, "grad_norm": 1.4216561539475785, "learning_rate": 1.831012139031189e-05, "loss": 0.7578, "step": 1044 }, { "epoch": 2.1239837398373984, "grad_norm": 1.5903990531520613, "learning_rate": 1.8306457552092e-05, "loss": 0.8379, "step": 1045 }, { "epoch": 2.1260162601626016, "grad_norm": 1.5110066214901965, "learning_rate": 1.8302790113770897e-05, "loss": 0.9264, "step": 1046 }, { "epoch": 2.1280487804878048, "grad_norm": 1.1598815994762204, "learning_rate": 1.8299119076938097e-05, "loss": 0.7211, "step": 1047 }, { "epoch": 2.130081300813008, "grad_norm": 1.4963902680439172, "learning_rate": 1.829544444318466e-05, "loss": 0.9238, "step": 1048 }, { "epoch": 2.1321138211382116, "grad_norm": 1.600896788954843, "learning_rate": 1.8291766214103207e-05, "loss": 0.7758, "step": 1049 }, { "epoch": 2.1341463414634148, "grad_norm": 1.4659785298159118, "learning_rate": 1.828808439128792e-05, "loss": 0.8894, "step": 1050 }, { "epoch": 2.136178861788618, "grad_norm": 1.5575211960521285, "learning_rate": 1.8284398976334538e-05, "loss": 0.9364, "step": 1051 }, { "epoch": 2.138211382113821, "grad_norm": 1.3393180007105614, "learning_rate": 1.8280709970840352e-05, "loss": 0.7822, "step": 1052 }, { "epoch": 2.1402439024390243, "grad_norm": 1.4186981727866537, "learning_rate": 1.8277017376404224e-05, "loss": 0.8587, "step": 1053 }, { "epoch": 2.1422764227642275, "grad_norm": 1.6259229323574385, "learning_rate": 1.8273321194626552e-05, "loss": 0.8563, "step": 1054 }, { "epoch": 2.144308943089431, "grad_norm": 1.3355999414075383, "learning_rate": 1.8269621427109302e-05, "loss": 0.9679, "step": 1055 }, { "epoch": 2.1463414634146343, "grad_norm": 1.3218275895710416, "learning_rate": 1.8265918075455985e-05, "loss": 0.8892, "step": 1056 }, { "epoch": 2.1483739837398375, "grad_norm": 1.2883222477076337, "learning_rate": 1.8262211141271678e-05, "loss": 0.8907, "step": 1057 }, { "epoch": 2.1504065040650406, "grad_norm": 1.4484549106629436, "learning_rate": 1.8258500626163e-05, "loss": 0.9496, "step": 1058 }, { "epoch": 2.152439024390244, "grad_norm": 1.6644950726668037, "learning_rate": 1.8254786531738124e-05, "loss": 1.0737, "step": 1059 }, { "epoch": 2.154471544715447, "grad_norm": 1.5690003239780932, "learning_rate": 1.8251068859606777e-05, "loss": 0.8554, "step": 1060 }, { "epoch": 2.15650406504065, "grad_norm": 1.4376314413425977, "learning_rate": 1.824734761138023e-05, "loss": 0.907, "step": 1061 }, { "epoch": 2.158536585365854, "grad_norm": 1.518902295827743, "learning_rate": 1.8243622788671316e-05, "loss": 0.9541, "step": 1062 }, { "epoch": 2.160569105691057, "grad_norm": 1.6415810915479836, "learning_rate": 1.8239894393094403e-05, "loss": 0.9354, "step": 1063 }, { "epoch": 2.16260162601626, "grad_norm": 1.5014746244241026, "learning_rate": 1.823616242626542e-05, "loss": 0.7335, "step": 1064 }, { "epoch": 2.1646341463414633, "grad_norm": 1.5371582500953065, "learning_rate": 1.8232426889801843e-05, "loss": 0.9941, "step": 1065 }, { "epoch": 2.1666666666666665, "grad_norm": 1.496033036462389, "learning_rate": 1.822868778532268e-05, "loss": 0.7742, "step": 1066 }, { "epoch": 2.16869918699187, "grad_norm": 1.3260221785383963, "learning_rate": 1.82249451144485e-05, "loss": 0.9486, "step": 1067 }, { "epoch": 2.1707317073170733, "grad_norm": 1.5580813781598792, "learning_rate": 1.8221198878801415e-05, "loss": 1.0465, "step": 1068 }, { "epoch": 2.1727642276422765, "grad_norm": 1.4963662663951598, "learning_rate": 1.8217449080005078e-05, "loss": 0.9479, "step": 1069 }, { "epoch": 2.1747967479674797, "grad_norm": 1.3239865377876459, "learning_rate": 1.8213695719684696e-05, "loss": 0.7862, "step": 1070 }, { "epoch": 2.176829268292683, "grad_norm": 1.3389566015698402, "learning_rate": 1.8209938799467003e-05, "loss": 0.7852, "step": 1071 }, { "epoch": 2.178861788617886, "grad_norm": 1.5617743677581213, "learning_rate": 1.8206178320980295e-05, "loss": 0.9661, "step": 1072 }, { "epoch": 2.180894308943089, "grad_norm": 1.3140258329578287, "learning_rate": 1.8202414285854394e-05, "loss": 0.7373, "step": 1073 }, { "epoch": 2.182926829268293, "grad_norm": 1.6798743486427425, "learning_rate": 1.8198646695720674e-05, "loss": 0.9093, "step": 1074 }, { "epoch": 2.184959349593496, "grad_norm": 1.5258233898313864, "learning_rate": 1.819487555221204e-05, "loss": 0.9126, "step": 1075 }, { "epoch": 2.186991869918699, "grad_norm": 1.51163508433622, "learning_rate": 1.819110085696295e-05, "loss": 0.9059, "step": 1076 }, { "epoch": 2.1890243902439024, "grad_norm": 1.3235102620448917, "learning_rate": 1.8187322611609394e-05, "loss": 0.859, "step": 1077 }, { "epoch": 2.1910569105691056, "grad_norm": 1.5390808379280645, "learning_rate": 1.8183540817788897e-05, "loss": 0.8988, "step": 1078 }, { "epoch": 2.1930894308943087, "grad_norm": 1.4415198715276627, "learning_rate": 1.8179755477140525e-05, "loss": 0.8096, "step": 1079 }, { "epoch": 2.1951219512195124, "grad_norm": 1.7144943805449848, "learning_rate": 1.817596659130489e-05, "loss": 0.9373, "step": 1080 }, { "epoch": 2.1971544715447155, "grad_norm": 1.5653842771965603, "learning_rate": 1.8172174161924125e-05, "loss": 0.6989, "step": 1081 }, { "epoch": 2.1991869918699187, "grad_norm": 1.3982277511485066, "learning_rate": 1.8168378190641912e-05, "loss": 0.8357, "step": 1082 }, { "epoch": 2.201219512195122, "grad_norm": 1.6752651377197707, "learning_rate": 1.8164578679103457e-05, "loss": 0.8955, "step": 1083 }, { "epoch": 2.203252032520325, "grad_norm": 1.548541656474497, "learning_rate": 1.816077562895551e-05, "loss": 0.9483, "step": 1084 }, { "epoch": 2.2052845528455283, "grad_norm": 1.5406753515718419, "learning_rate": 1.8156969041846353e-05, "loss": 0.8821, "step": 1085 }, { "epoch": 2.207317073170732, "grad_norm": 1.4935133412604773, "learning_rate": 1.815315891942579e-05, "loss": 0.91, "step": 1086 }, { "epoch": 2.209349593495935, "grad_norm": 1.5346465960012916, "learning_rate": 1.8149345263345175e-05, "loss": 0.8851, "step": 1087 }, { "epoch": 2.2113821138211383, "grad_norm": 1.3289989840765335, "learning_rate": 1.814552807525738e-05, "loss": 0.9691, "step": 1088 }, { "epoch": 2.2134146341463414, "grad_norm": 1.7677187898888127, "learning_rate": 1.8141707356816812e-05, "loss": 1.0694, "step": 1089 }, { "epoch": 2.2154471544715446, "grad_norm": 1.3820573641569451, "learning_rate": 1.813788310967941e-05, "loss": 0.9286, "step": 1090 }, { "epoch": 2.217479674796748, "grad_norm": 1.3756515730019847, "learning_rate": 1.8134055335502642e-05, "loss": 0.6712, "step": 1091 }, { "epoch": 2.2195121951219514, "grad_norm": 1.6166600495989427, "learning_rate": 1.81302240359455e-05, "loss": 0.9681, "step": 1092 }, { "epoch": 2.2215447154471546, "grad_norm": 1.5666735201159092, "learning_rate": 1.812638921266851e-05, "loss": 0.9081, "step": 1093 }, { "epoch": 2.2235772357723578, "grad_norm": 1.6311173883344805, "learning_rate": 1.812255086733372e-05, "loss": 0.9403, "step": 1094 }, { "epoch": 2.225609756097561, "grad_norm": 1.4391006737066314, "learning_rate": 1.8118709001604705e-05, "loss": 0.8862, "step": 1095 }, { "epoch": 2.227642276422764, "grad_norm": 1.4450090601133254, "learning_rate": 1.8114863617146576e-05, "loss": 0.8178, "step": 1096 }, { "epoch": 2.2296747967479673, "grad_norm": 1.4709731090719822, "learning_rate": 1.8111014715625955e-05, "loss": 0.8836, "step": 1097 }, { "epoch": 2.231707317073171, "grad_norm": 1.445075514815924, "learning_rate": 1.8107162298710995e-05, "loss": 0.7309, "step": 1098 }, { "epoch": 2.233739837398374, "grad_norm": 1.4878026538672262, "learning_rate": 1.810330636807137e-05, "loss": 0.8519, "step": 1099 }, { "epoch": 2.2357723577235773, "grad_norm": 1.5595055484623166, "learning_rate": 1.8099446925378278e-05, "loss": 1.0935, "step": 1100 }, { "epoch": 2.2378048780487805, "grad_norm": 1.415283400167496, "learning_rate": 1.8095583972304448e-05, "loss": 0.9266, "step": 1101 }, { "epoch": 2.2398373983739837, "grad_norm": 1.5080970001266507, "learning_rate": 1.8091717510524113e-05, "loss": 1.1409, "step": 1102 }, { "epoch": 2.241869918699187, "grad_norm": 1.6481489022613627, "learning_rate": 1.808784754171304e-05, "loss": 0.8497, "step": 1103 }, { "epoch": 2.2439024390243905, "grad_norm": 1.5035828643073685, "learning_rate": 1.8083974067548506e-05, "loss": 1.1053, "step": 1104 }, { "epoch": 2.2459349593495936, "grad_norm": 1.9887848389442575, "learning_rate": 1.8080097089709326e-05, "loss": 0.8639, "step": 1105 }, { "epoch": 2.247967479674797, "grad_norm": 1.4198531789747608, "learning_rate": 1.807621660987581e-05, "loss": 0.7662, "step": 1106 }, { "epoch": 2.25, "grad_norm": 1.5455668147231352, "learning_rate": 1.8072332629729796e-05, "loss": 0.8719, "step": 1107 }, { "epoch": 2.252032520325203, "grad_norm": 1.6945907132680562, "learning_rate": 1.806844515095465e-05, "loss": 1.1957, "step": 1108 }, { "epoch": 2.2540650406504064, "grad_norm": 1.4343154680781143, "learning_rate": 1.806455417523523e-05, "loss": 0.8749, "step": 1109 }, { "epoch": 2.2560975609756095, "grad_norm": 1.5475170892320265, "learning_rate": 1.8060659704257937e-05, "loss": 0.9515, "step": 1110 }, { "epoch": 2.258130081300813, "grad_norm": 1.4002624309570026, "learning_rate": 1.8056761739710664e-05, "loss": 0.6861, "step": 1111 }, { "epoch": 2.2601626016260163, "grad_norm": 1.6180121453767682, "learning_rate": 1.8052860283282832e-05, "loss": 0.8775, "step": 1112 }, { "epoch": 2.2621951219512195, "grad_norm": 1.4282136340815432, "learning_rate": 1.8048955336665367e-05, "loss": 0.7873, "step": 1113 }, { "epoch": 2.2642276422764227, "grad_norm": 1.5464103436045251, "learning_rate": 1.8045046901550715e-05, "loss": 0.8238, "step": 1114 }, { "epoch": 2.266260162601626, "grad_norm": 1.293682275908345, "learning_rate": 1.8041134979632833e-05, "loss": 0.8781, "step": 1115 }, { "epoch": 2.2682926829268295, "grad_norm": 1.3109657387551548, "learning_rate": 1.8037219572607177e-05, "loss": 0.7125, "step": 1116 }, { "epoch": 2.2703252032520327, "grad_norm": 1.6851800791084077, "learning_rate": 1.8033300682170736e-05, "loss": 1.0388, "step": 1117 }, { "epoch": 2.272357723577236, "grad_norm": 1.35239043187063, "learning_rate": 1.8029378310021987e-05, "loss": 0.6827, "step": 1118 }, { "epoch": 2.274390243902439, "grad_norm": 1.6485737337243598, "learning_rate": 1.802545245786093e-05, "loss": 0.7251, "step": 1119 }, { "epoch": 2.2764227642276422, "grad_norm": 1.491010991502711, "learning_rate": 1.8021523127389066e-05, "loss": 0.8979, "step": 1120 }, { "epoch": 2.2784552845528454, "grad_norm": 1.6512128176579937, "learning_rate": 1.8017590320309403e-05, "loss": 0.7751, "step": 1121 }, { "epoch": 2.2804878048780486, "grad_norm": 1.556379533879269, "learning_rate": 1.8013654038326463e-05, "loss": 0.8908, "step": 1122 }, { "epoch": 2.282520325203252, "grad_norm": 1.5856135430400258, "learning_rate": 1.8009714283146268e-05, "loss": 0.8887, "step": 1123 }, { "epoch": 2.2845528455284554, "grad_norm": 1.38548442318059, "learning_rate": 1.800577105647635e-05, "loss": 0.8269, "step": 1124 }, { "epoch": 2.2865853658536586, "grad_norm": 1.695461332960446, "learning_rate": 1.8001824360025737e-05, "loss": 0.7527, "step": 1125 }, { "epoch": 2.2886178861788617, "grad_norm": 1.3547128107039184, "learning_rate": 1.7997874195504974e-05, "loss": 0.8764, "step": 1126 }, { "epoch": 2.290650406504065, "grad_norm": 1.6305104931637568, "learning_rate": 1.7993920564626093e-05, "loss": 0.8739, "step": 1127 }, { "epoch": 2.292682926829268, "grad_norm": 1.5442840613998718, "learning_rate": 1.7989963469102643e-05, "loss": 0.7395, "step": 1128 }, { "epoch": 2.2947154471544717, "grad_norm": 1.5863321873883192, "learning_rate": 1.7986002910649664e-05, "loss": 0.7963, "step": 1129 }, { "epoch": 2.296747967479675, "grad_norm": 1.8026311009905842, "learning_rate": 1.7982038890983707e-05, "loss": 0.8836, "step": 1130 }, { "epoch": 2.298780487804878, "grad_norm": 1.6972719034934376, "learning_rate": 1.797807141182281e-05, "loss": 0.7924, "step": 1131 }, { "epoch": 2.3008130081300813, "grad_norm": 1.7895535306487804, "learning_rate": 1.797410047488653e-05, "loss": 0.7891, "step": 1132 }, { "epoch": 2.3028455284552845, "grad_norm": 1.6767951547333353, "learning_rate": 1.7970126081895896e-05, "loss": 0.9443, "step": 1133 }, { "epoch": 2.3048780487804876, "grad_norm": 1.6416151750708727, "learning_rate": 1.796614823457346e-05, "loss": 1.0295, "step": 1134 }, { "epoch": 2.306910569105691, "grad_norm": 1.6363383976212602, "learning_rate": 1.7962166934643255e-05, "loss": 1.0332, "step": 1135 }, { "epoch": 2.3089430894308944, "grad_norm": 1.6943108088231362, "learning_rate": 1.7958182183830816e-05, "loss": 0.7822, "step": 1136 }, { "epoch": 2.3109756097560976, "grad_norm": 1.6431664099847627, "learning_rate": 1.7954193983863177e-05, "loss": 0.7827, "step": 1137 }, { "epoch": 2.313008130081301, "grad_norm": 1.6381841824018082, "learning_rate": 1.795020233646886e-05, "loss": 0.8925, "step": 1138 }, { "epoch": 2.315040650406504, "grad_norm": 1.3331236652831466, "learning_rate": 1.7946207243377885e-05, "loss": 0.7546, "step": 1139 }, { "epoch": 2.317073170731707, "grad_norm": 1.550759273435591, "learning_rate": 1.794220870632177e-05, "loss": 0.9287, "step": 1140 }, { "epoch": 2.319105691056911, "grad_norm": 1.838264184482661, "learning_rate": 1.793820672703352e-05, "loss": 0.8843, "step": 1141 }, { "epoch": 2.321138211382114, "grad_norm": 1.6090227182675636, "learning_rate": 1.7934201307247623e-05, "loss": 0.947, "step": 1142 }, { "epoch": 2.323170731707317, "grad_norm": 1.624335407209434, "learning_rate": 1.7930192448700076e-05, "loss": 0.8355, "step": 1143 }, { "epoch": 2.3252032520325203, "grad_norm": 1.4602784338696755, "learning_rate": 1.7926180153128358e-05, "loss": 0.7701, "step": 1144 }, { "epoch": 2.3272357723577235, "grad_norm": 1.5361151467565832, "learning_rate": 1.7922164422271434e-05, "loss": 0.8659, "step": 1145 }, { "epoch": 2.3292682926829267, "grad_norm": 1.412465256349486, "learning_rate": 1.7918145257869765e-05, "loss": 1.0211, "step": 1146 }, { "epoch": 2.33130081300813, "grad_norm": 1.6525745633254052, "learning_rate": 1.7914122661665298e-05, "loss": 0.819, "step": 1147 }, { "epoch": 2.3333333333333335, "grad_norm": 1.4730361653060309, "learning_rate": 1.791009663540146e-05, "loss": 0.8424, "step": 1148 }, { "epoch": 2.3353658536585367, "grad_norm": 1.584942922288531, "learning_rate": 1.7906067180823175e-05, "loss": 1.0165, "step": 1149 }, { "epoch": 2.33739837398374, "grad_norm": 1.5711888572087893, "learning_rate": 1.7902034299676847e-05, "loss": 0.8592, "step": 1150 }, { "epoch": 2.339430894308943, "grad_norm": 1.569514200013143, "learning_rate": 1.789799799371037e-05, "loss": 0.7387, "step": 1151 }, { "epoch": 2.341463414634146, "grad_norm": 1.451858993581849, "learning_rate": 1.789395826467312e-05, "loss": 0.7021, "step": 1152 }, { "epoch": 2.34349593495935, "grad_norm": 1.545858198159394, "learning_rate": 1.7889915114315947e-05, "loss": 0.8303, "step": 1153 }, { "epoch": 2.345528455284553, "grad_norm": 1.6354849951285422, "learning_rate": 1.78858685443912e-05, "loss": 0.8226, "step": 1154 }, { "epoch": 2.347560975609756, "grad_norm": 1.6107025143864508, "learning_rate": 1.78818185566527e-05, "loss": 0.8796, "step": 1155 }, { "epoch": 2.3495934959349594, "grad_norm": 1.559203486687833, "learning_rate": 1.7877765152855757e-05, "loss": 0.7863, "step": 1156 }, { "epoch": 2.3516260162601625, "grad_norm": 1.600910421001587, "learning_rate": 1.7873708334757148e-05, "loss": 0.7314, "step": 1157 }, { "epoch": 2.3536585365853657, "grad_norm": 1.6953977511308527, "learning_rate": 1.7869648104115142e-05, "loss": 1.016, "step": 1158 }, { "epoch": 2.355691056910569, "grad_norm": 1.5268543919438529, "learning_rate": 1.7865584462689484e-05, "loss": 0.8668, "step": 1159 }, { "epoch": 2.3577235772357725, "grad_norm": 1.8615824251551398, "learning_rate": 1.78615174122414e-05, "loss": 0.739, "step": 1160 }, { "epoch": 2.3597560975609757, "grad_norm": 1.4693962804133567, "learning_rate": 1.7857446954533584e-05, "loss": 0.7486, "step": 1161 }, { "epoch": 2.361788617886179, "grad_norm": 1.7314295061109426, "learning_rate": 1.7853373091330213e-05, "loss": 0.8004, "step": 1162 }, { "epoch": 2.363821138211382, "grad_norm": 1.5095586523010103, "learning_rate": 1.7849295824396946e-05, "loss": 0.7302, "step": 1163 }, { "epoch": 2.3658536585365852, "grad_norm": 1.6497902013223014, "learning_rate": 1.78452151555009e-05, "loss": 0.841, "step": 1164 }, { "epoch": 2.3678861788617884, "grad_norm": 1.9152378176089753, "learning_rate": 1.784113108641069e-05, "loss": 0.8348, "step": 1165 }, { "epoch": 2.369918699186992, "grad_norm": 1.5662626769514134, "learning_rate": 1.7837043618896386e-05, "loss": 0.8588, "step": 1166 }, { "epoch": 2.3719512195121952, "grad_norm": 1.710901478397318, "learning_rate": 1.7832952754729536e-05, "loss": 1.0463, "step": 1167 }, { "epoch": 2.3739837398373984, "grad_norm": 1.6057946047215839, "learning_rate": 1.7828858495683162e-05, "loss": 0.6816, "step": 1168 }, { "epoch": 2.3760162601626016, "grad_norm": 1.6729199987735808, "learning_rate": 1.782476084353176e-05, "loss": 0.9275, "step": 1169 }, { "epoch": 2.3780487804878048, "grad_norm": 1.7525144231510865, "learning_rate": 1.7820659800051284e-05, "loss": 0.9054, "step": 1170 }, { "epoch": 2.380081300813008, "grad_norm": 1.72505898372264, "learning_rate": 1.7816555367019178e-05, "loss": 0.9147, "step": 1171 }, { "epoch": 2.3821138211382116, "grad_norm": 1.5132556769502743, "learning_rate": 1.781244754621434e-05, "loss": 0.9028, "step": 1172 }, { "epoch": 2.3841463414634148, "grad_norm": 1.4480658940151023, "learning_rate": 1.780833633941714e-05, "loss": 0.7407, "step": 1173 }, { "epoch": 2.386178861788618, "grad_norm": 1.7631431896578582, "learning_rate": 1.7804221748409414e-05, "loss": 0.7374, "step": 1174 }, { "epoch": 2.388211382113821, "grad_norm": 1.722128439297568, "learning_rate": 1.7800103774974466e-05, "loss": 0.8456, "step": 1175 }, { "epoch": 2.3902439024390243, "grad_norm": 1.7891985165538487, "learning_rate": 1.779598242089707e-05, "loss": 0.9528, "step": 1176 }, { "epoch": 2.3922764227642275, "grad_norm": 1.6428240291007574, "learning_rate": 1.7791857687963464e-05, "loss": 0.8584, "step": 1177 }, { "epoch": 2.394308943089431, "grad_norm": 1.7411703030771697, "learning_rate": 1.7787729577961343e-05, "loss": 0.8504, "step": 1178 }, { "epoch": 2.3963414634146343, "grad_norm": 1.619334467230389, "learning_rate": 1.778359809267987e-05, "loss": 0.8378, "step": 1179 }, { "epoch": 2.3983739837398375, "grad_norm": 1.818319672622193, "learning_rate": 1.7779463233909677e-05, "loss": 0.982, "step": 1180 }, { "epoch": 2.4004065040650406, "grad_norm": 1.4354954413641383, "learning_rate": 1.7775325003442848e-05, "loss": 0.8734, "step": 1181 }, { "epoch": 2.402439024390244, "grad_norm": 1.7755252267441228, "learning_rate": 1.7771183403072937e-05, "loss": 0.7892, "step": 1182 }, { "epoch": 2.404471544715447, "grad_norm": 1.5927888121546465, "learning_rate": 1.776703843459495e-05, "loss": 1.0684, "step": 1183 }, { "epoch": 2.40650406504065, "grad_norm": 1.5532805999715613, "learning_rate": 1.7762890099805362e-05, "loss": 0.7672, "step": 1184 }, { "epoch": 2.408536585365854, "grad_norm": 1.4680098015500684, "learning_rate": 1.77587384005021e-05, "loss": 0.723, "step": 1185 }, { "epoch": 2.410569105691057, "grad_norm": 1.873402121868627, "learning_rate": 1.7754583338484555e-05, "loss": 0.8894, "step": 1186 }, { "epoch": 2.41260162601626, "grad_norm": 1.6123846132601654, "learning_rate": 1.7750424915553573e-05, "loss": 0.6689, "step": 1187 }, { "epoch": 2.4146341463414633, "grad_norm": 1.6562375991335065, "learning_rate": 1.774626313351145e-05, "loss": 0.7485, "step": 1188 }, { "epoch": 2.4166666666666665, "grad_norm": 1.6934791940601217, "learning_rate": 1.7742097994161946e-05, "loss": 0.7833, "step": 1189 }, { "epoch": 2.41869918699187, "grad_norm": 1.6446508045599708, "learning_rate": 1.7737929499310274e-05, "loss": 0.8733, "step": 1190 }, { "epoch": 2.4207317073170733, "grad_norm": 1.4381758858131648, "learning_rate": 1.7733757650763103e-05, "loss": 0.7277, "step": 1191 }, { "epoch": 2.4227642276422765, "grad_norm": 1.7075652651607143, "learning_rate": 1.7729582450328547e-05, "loss": 0.9904, "step": 1192 }, { "epoch": 2.4247967479674797, "grad_norm": 1.5209927454219099, "learning_rate": 1.7725403899816188e-05, "loss": 0.8095, "step": 1193 }, { "epoch": 2.426829268292683, "grad_norm": 1.5441315275249952, "learning_rate": 1.772122200103705e-05, "loss": 0.7997, "step": 1194 }, { "epoch": 2.428861788617886, "grad_norm": 1.598577649307059, "learning_rate": 1.7717036755803604e-05, "loss": 0.8742, "step": 1195 }, { "epoch": 2.430894308943089, "grad_norm": 1.5482112212858397, "learning_rate": 1.771284816592978e-05, "loss": 0.859, "step": 1196 }, { "epoch": 2.432926829268293, "grad_norm": 1.9913499834032267, "learning_rate": 1.770865623323095e-05, "loss": 0.9318, "step": 1197 }, { "epoch": 2.434959349593496, "grad_norm": 1.6340647126849897, "learning_rate": 1.7704460959523947e-05, "loss": 0.8655, "step": 1198 }, { "epoch": 2.436991869918699, "grad_norm": 1.6553459832411719, "learning_rate": 1.770026234662704e-05, "loss": 0.9649, "step": 1199 }, { "epoch": 2.4390243902439024, "grad_norm": 1.4976547605221389, "learning_rate": 1.7696060396359956e-05, "loss": 0.7887, "step": 1200 }, { "epoch": 2.4410569105691056, "grad_norm": 1.771200692140032, "learning_rate": 1.769185511054385e-05, "loss": 0.9615, "step": 1201 }, { "epoch": 2.443089430894309, "grad_norm": 1.7012423681662103, "learning_rate": 1.7687646491001346e-05, "loss": 0.8144, "step": 1202 }, { "epoch": 2.4451219512195124, "grad_norm": 1.634963544991747, "learning_rate": 1.768343453955649e-05, "loss": 0.818, "step": 1203 }, { "epoch": 2.4471544715447155, "grad_norm": 1.5443667225086775, "learning_rate": 1.7679219258034798e-05, "loss": 0.8377, "step": 1204 }, { "epoch": 2.4491869918699187, "grad_norm": 1.6863870407566612, "learning_rate": 1.76750006482632e-05, "loss": 0.8042, "step": 1205 }, { "epoch": 2.451219512195122, "grad_norm": 1.6427326076105084, "learning_rate": 1.7670778712070092e-05, "loss": 0.7305, "step": 1206 }, { "epoch": 2.453252032520325, "grad_norm": 1.6949726284754236, "learning_rate": 1.7666553451285303e-05, "loss": 0.8066, "step": 1207 }, { "epoch": 2.4552845528455283, "grad_norm": 1.6008018468738423, "learning_rate": 1.7662324867740102e-05, "loss": 0.9322, "step": 1208 }, { "epoch": 2.457317073170732, "grad_norm": 1.7571296700413508, "learning_rate": 1.7658092963267197e-05, "loss": 0.9406, "step": 1209 }, { "epoch": 2.459349593495935, "grad_norm": 1.5901819916778053, "learning_rate": 1.765385773970074e-05, "loss": 0.937, "step": 1210 }, { "epoch": 2.4613821138211383, "grad_norm": 1.6009540283247536, "learning_rate": 1.7649619198876316e-05, "loss": 0.8077, "step": 1211 }, { "epoch": 2.4634146341463414, "grad_norm": 1.5166731277917334, "learning_rate": 1.7645377342630956e-05, "loss": 0.7994, "step": 1212 }, { "epoch": 2.4654471544715446, "grad_norm": 1.757392164216317, "learning_rate": 1.7641132172803117e-05, "loss": 0.8608, "step": 1213 }, { "epoch": 2.467479674796748, "grad_norm": 1.5804123636527811, "learning_rate": 1.76368836912327e-05, "loss": 0.8269, "step": 1214 }, { "epoch": 2.4695121951219514, "grad_norm": 1.722370331328912, "learning_rate": 1.763263189976104e-05, "loss": 0.8052, "step": 1215 }, { "epoch": 2.4715447154471546, "grad_norm": 1.7684700870797183, "learning_rate": 1.76283768002309e-05, "loss": 0.9892, "step": 1216 }, { "epoch": 2.4735772357723578, "grad_norm": 1.910324154162393, "learning_rate": 1.7624118394486495e-05, "loss": 0.9814, "step": 1217 }, { "epoch": 2.475609756097561, "grad_norm": 1.8145430827922908, "learning_rate": 1.761985668437345e-05, "loss": 0.9456, "step": 1218 }, { "epoch": 2.477642276422764, "grad_norm": 1.5276568812080416, "learning_rate": 1.761559167173883e-05, "loss": 0.7539, "step": 1219 }, { "epoch": 2.4796747967479673, "grad_norm": 1.8368804241482377, "learning_rate": 1.7611323358431145e-05, "loss": 0.8228, "step": 1220 }, { "epoch": 2.4817073170731705, "grad_norm": 1.707384096709947, "learning_rate": 1.7607051746300315e-05, "loss": 0.8555, "step": 1221 }, { "epoch": 2.483739837398374, "grad_norm": 1.6359055132760354, "learning_rate": 1.7602776837197703e-05, "loss": 0.9963, "step": 1222 }, { "epoch": 2.4857723577235773, "grad_norm": 1.8256399895696942, "learning_rate": 1.75984986329761e-05, "loss": 0.9287, "step": 1223 }, { "epoch": 2.4878048780487805, "grad_norm": 1.626479023348185, "learning_rate": 1.759421713548971e-05, "loss": 0.7937, "step": 1224 }, { "epoch": 2.4898373983739837, "grad_norm": 1.6686046370699212, "learning_rate": 1.7589932346594193e-05, "loss": 0.7708, "step": 1225 }, { "epoch": 2.491869918699187, "grad_norm": 1.6880966535368118, "learning_rate": 1.758564426814661e-05, "loss": 0.7866, "step": 1226 }, { "epoch": 2.4939024390243905, "grad_norm": 1.7068373873738372, "learning_rate": 1.7581352902005456e-05, "loss": 0.8439, "step": 1227 }, { "epoch": 2.4959349593495936, "grad_norm": 1.758660868243928, "learning_rate": 1.757705825003065e-05, "loss": 0.8816, "step": 1228 }, { "epoch": 2.497967479674797, "grad_norm": 1.814859483150012, "learning_rate": 1.7572760314083547e-05, "loss": 0.9281, "step": 1229 }, { "epoch": 2.5, "grad_norm": 1.8068104726931369, "learning_rate": 1.756845909602691e-05, "loss": 0.7673, "step": 1230 }, { "epoch": 2.502032520325203, "grad_norm": 1.4551922275586961, "learning_rate": 1.756415459772492e-05, "loss": 0.8579, "step": 1231 }, { "epoch": 2.5040650406504064, "grad_norm": 1.408117976274147, "learning_rate": 1.7559846821043205e-05, "loss": 0.7408, "step": 1232 }, { "epoch": 2.5060975609756095, "grad_norm": 1.5088420676250853, "learning_rate": 1.7555535767848788e-05, "loss": 0.8495, "step": 1233 }, { "epoch": 2.508130081300813, "grad_norm": 1.763095457025744, "learning_rate": 1.7551221440010126e-05, "loss": 0.8715, "step": 1234 }, { "epoch": 2.5101626016260163, "grad_norm": 1.4966171453963752, "learning_rate": 1.754690383939709e-05, "loss": 0.7237, "step": 1235 }, { "epoch": 2.5121951219512195, "grad_norm": 1.7872760179609648, "learning_rate": 1.754258296788097e-05, "loss": 1.0186, "step": 1236 }, { "epoch": 2.5142276422764227, "grad_norm": 1.7049575448185132, "learning_rate": 1.7538258827334477e-05, "loss": 0.7706, "step": 1237 }, { "epoch": 2.516260162601626, "grad_norm": 1.8257373571314754, "learning_rate": 1.7533931419631736e-05, "loss": 1.0181, "step": 1238 }, { "epoch": 2.5182926829268295, "grad_norm": 1.759427279128996, "learning_rate": 1.752960074664829e-05, "loss": 0.7691, "step": 1239 }, { "epoch": 2.5203252032520327, "grad_norm": 1.945427964091183, "learning_rate": 1.7525266810261096e-05, "loss": 0.8766, "step": 1240 }, { "epoch": 2.522357723577236, "grad_norm": 1.6128529716372713, "learning_rate": 1.7520929612348516e-05, "loss": 0.841, "step": 1241 }, { "epoch": 2.524390243902439, "grad_norm": 1.6957194972687302, "learning_rate": 1.7516589154790346e-05, "loss": 0.977, "step": 1242 }, { "epoch": 2.5264227642276422, "grad_norm": 1.7660769378845431, "learning_rate": 1.751224543946778e-05, "loss": 0.9366, "step": 1243 }, { "epoch": 2.5284552845528454, "grad_norm": 1.6459288304773898, "learning_rate": 1.7507898468263422e-05, "loss": 0.9183, "step": 1244 }, { "epoch": 2.5304878048780486, "grad_norm": 1.7469029934302338, "learning_rate": 1.75035482430613e-05, "loss": 0.888, "step": 1245 }, { "epoch": 2.5325203252032518, "grad_norm": 1.8652803546759347, "learning_rate": 1.749919476574684e-05, "loss": 0.886, "step": 1246 }, { "epoch": 2.5345528455284554, "grad_norm": 1.7517951914369267, "learning_rate": 1.7494838038206886e-05, "loss": 0.9052, "step": 1247 }, { "epoch": 2.5365853658536586, "grad_norm": 1.773635480916089, "learning_rate": 1.7490478062329686e-05, "loss": 0.8484, "step": 1248 }, { "epoch": 2.5386178861788617, "grad_norm": 1.5881356914740683, "learning_rate": 1.748611484000489e-05, "loss": 0.894, "step": 1249 }, { "epoch": 2.540650406504065, "grad_norm": 1.5122469154804405, "learning_rate": 1.7481748373123574e-05, "loss": 0.9552, "step": 1250 }, { "epoch": 2.5426829268292686, "grad_norm": 1.7032231812976946, "learning_rate": 1.74773786635782e-05, "loss": 0.7304, "step": 1251 }, { "epoch": 2.5447154471544717, "grad_norm": 1.722943708413078, "learning_rate": 1.7473005713262644e-05, "loss": 0.9188, "step": 1252 }, { "epoch": 2.546747967479675, "grad_norm": 1.5262556175517126, "learning_rate": 1.7468629524072187e-05, "loss": 0.912, "step": 1253 }, { "epoch": 2.548780487804878, "grad_norm": 1.7314327788501667, "learning_rate": 1.7464250097903515e-05, "loss": 0.9698, "step": 1254 }, { "epoch": 2.5508130081300813, "grad_norm": 1.6842292811010664, "learning_rate": 1.7459867436654715e-05, "loss": 0.7972, "step": 1255 }, { "epoch": 2.5528455284552845, "grad_norm": 1.7957650193020307, "learning_rate": 1.7455481542225272e-05, "loss": 0.8785, "step": 1256 }, { "epoch": 2.5548780487804876, "grad_norm": 1.8144477248471902, "learning_rate": 1.7451092416516076e-05, "loss": 0.768, "step": 1257 }, { "epoch": 2.556910569105691, "grad_norm": 1.8418899143561107, "learning_rate": 1.744670006142942e-05, "loss": 0.8754, "step": 1258 }, { "epoch": 2.5589430894308944, "grad_norm": 1.6573510996652663, "learning_rate": 1.7442304478868998e-05, "loss": 0.7245, "step": 1259 }, { "epoch": 2.5609756097560976, "grad_norm": 1.7980842126809347, "learning_rate": 1.7437905670739893e-05, "loss": 0.935, "step": 1260 }, { "epoch": 2.563008130081301, "grad_norm": 2.038864628905934, "learning_rate": 1.7433503638948597e-05, "loss": 0.8392, "step": 1261 }, { "epoch": 2.565040650406504, "grad_norm": 1.7132563437509656, "learning_rate": 1.742909838540299e-05, "loss": 0.8181, "step": 1262 }, { "epoch": 2.567073170731707, "grad_norm": 1.7049722954729376, "learning_rate": 1.7424689912012354e-05, "loss": 0.8684, "step": 1263 }, { "epoch": 2.569105691056911, "grad_norm": 2.0069559919494333, "learning_rate": 1.7420278220687366e-05, "loss": 0.9314, "step": 1264 }, { "epoch": 2.571138211382114, "grad_norm": 1.4997694331153393, "learning_rate": 1.7415863313340095e-05, "loss": 0.954, "step": 1265 }, { "epoch": 2.573170731707317, "grad_norm": 2.0136648676360323, "learning_rate": 1.741144519188401e-05, "loss": 0.8097, "step": 1266 }, { "epoch": 2.5752032520325203, "grad_norm": 1.4873759734798757, "learning_rate": 1.7407023858233963e-05, "loss": 0.8164, "step": 1267 }, { "epoch": 2.5772357723577235, "grad_norm": 1.6918428711716433, "learning_rate": 1.7402599314306207e-05, "loss": 0.9555, "step": 1268 }, { "epoch": 2.5792682926829267, "grad_norm": 1.494894915580852, "learning_rate": 1.7398171562018388e-05, "loss": 0.8402, "step": 1269 }, { "epoch": 2.58130081300813, "grad_norm": 1.6828175702986834, "learning_rate": 1.739374060328953e-05, "loss": 0.754, "step": 1270 }, { "epoch": 2.5833333333333335, "grad_norm": 2.1182633261226638, "learning_rate": 1.7389306440040053e-05, "loss": 0.9518, "step": 1271 }, { "epoch": 2.5853658536585367, "grad_norm": 1.8156346747418595, "learning_rate": 1.7384869074191777e-05, "loss": 1.0413, "step": 1272 }, { "epoch": 2.58739837398374, "grad_norm": 1.6266516081637155, "learning_rate": 1.7380428507667893e-05, "loss": 0.7475, "step": 1273 }, { "epoch": 2.589430894308943, "grad_norm": 1.5833552532802282, "learning_rate": 1.7375984742392988e-05, "loss": 0.9688, "step": 1274 }, { "epoch": 2.591463414634146, "grad_norm": 1.9988642235113938, "learning_rate": 1.737153778029304e-05, "loss": 0.949, "step": 1275 }, { "epoch": 2.59349593495935, "grad_norm": 1.5163736978584956, "learning_rate": 1.7367087623295394e-05, "loss": 0.6489, "step": 1276 }, { "epoch": 2.595528455284553, "grad_norm": 1.875544027991632, "learning_rate": 1.7362634273328806e-05, "loss": 0.7641, "step": 1277 }, { "epoch": 2.597560975609756, "grad_norm": 1.6405864097646592, "learning_rate": 1.735817773232339e-05, "loss": 0.7821, "step": 1278 }, { "epoch": 2.5995934959349594, "grad_norm": 1.7410440369460078, "learning_rate": 1.7353718002210663e-05, "loss": 0.9441, "step": 1279 }, { "epoch": 2.6016260162601625, "grad_norm": 1.7476866767740231, "learning_rate": 1.7349255084923517e-05, "loss": 0.8302, "step": 1280 }, { "epoch": 2.6036585365853657, "grad_norm": 1.9715615453231792, "learning_rate": 1.734478898239622e-05, "loss": 0.8039, "step": 1281 }, { "epoch": 2.605691056910569, "grad_norm": 1.542833288304705, "learning_rate": 1.7340319696564428e-05, "loss": 0.9041, "step": 1282 }, { "epoch": 2.607723577235772, "grad_norm": 1.6450164955075839, "learning_rate": 1.733584722936517e-05, "loss": 0.8616, "step": 1283 }, { "epoch": 2.6097560975609757, "grad_norm": 1.7440128974385725, "learning_rate": 1.7331371582736864e-05, "loss": 0.6932, "step": 1284 }, { "epoch": 2.611788617886179, "grad_norm": 1.6222136318582496, "learning_rate": 1.7326892758619296e-05, "loss": 0.7968, "step": 1285 }, { "epoch": 2.613821138211382, "grad_norm": 1.8826727535865664, "learning_rate": 1.7322410758953633e-05, "loss": 0.8285, "step": 1286 }, { "epoch": 2.6158536585365852, "grad_norm": 1.4923791782163989, "learning_rate": 1.731792558568242e-05, "loss": 0.7657, "step": 1287 }, { "epoch": 2.617886178861789, "grad_norm": 1.7353896433940201, "learning_rate": 1.731343724074957e-05, "loss": 0.9245, "step": 1288 }, { "epoch": 2.619918699186992, "grad_norm": 1.6763897707368398, "learning_rate": 1.7308945726100382e-05, "loss": 0.9172, "step": 1289 }, { "epoch": 2.6219512195121952, "grad_norm": 1.8385223902509937, "learning_rate": 1.7304451043681522e-05, "loss": 0.8985, "step": 1290 }, { "epoch": 2.6239837398373984, "grad_norm": 1.9582054192505376, "learning_rate": 1.729995319544103e-05, "loss": 0.8951, "step": 1291 }, { "epoch": 2.6260162601626016, "grad_norm": 1.7921605506169902, "learning_rate": 1.7295452183328317e-05, "loss": 0.8609, "step": 1292 }, { "epoch": 2.6280487804878048, "grad_norm": 1.853026385558461, "learning_rate": 1.7290948009294166e-05, "loss": 0.7592, "step": 1293 }, { "epoch": 2.630081300813008, "grad_norm": 1.392968668991951, "learning_rate": 1.7286440675290733e-05, "loss": 0.6756, "step": 1294 }, { "epoch": 2.632113821138211, "grad_norm": 1.9391085847872789, "learning_rate": 1.728193018327154e-05, "loss": 0.8946, "step": 1295 }, { "epoch": 2.6341463414634148, "grad_norm": 1.534993785689433, "learning_rate": 1.7277416535191478e-05, "loss": 0.7807, "step": 1296 }, { "epoch": 2.636178861788618, "grad_norm": 1.7446118050003, "learning_rate": 1.7272899733006808e-05, "loss": 0.8324, "step": 1297 }, { "epoch": 2.638211382113821, "grad_norm": 1.8351233662447173, "learning_rate": 1.7268379778675154e-05, "loss": 0.7191, "step": 1298 }, { "epoch": 2.6402439024390243, "grad_norm": 1.719298058395915, "learning_rate": 1.7263856674155517e-05, "loss": 0.7517, "step": 1299 }, { "epoch": 2.642276422764228, "grad_norm": 1.8440479967488312, "learning_rate": 1.7259330421408247e-05, "loss": 1.0633, "step": 1300 }, { "epoch": 2.644308943089431, "grad_norm": 1.8845061843387985, "learning_rate": 1.7254801022395072e-05, "loss": 0.848, "step": 1301 }, { "epoch": 2.6463414634146343, "grad_norm": 1.7615723163275017, "learning_rate": 1.7250268479079076e-05, "loss": 0.9243, "step": 1302 }, { "epoch": 2.6483739837398375, "grad_norm": 1.8597337727825098, "learning_rate": 1.724573279342471e-05, "loss": 0.9775, "step": 1303 }, { "epoch": 2.6504065040650406, "grad_norm": 1.8180629514040183, "learning_rate": 1.7241193967397784e-05, "loss": 0.9274, "step": 1304 }, { "epoch": 2.652439024390244, "grad_norm": 1.7016548458451421, "learning_rate": 1.7236652002965468e-05, "loss": 0.9783, "step": 1305 }, { "epoch": 2.654471544715447, "grad_norm": 1.642721839130805, "learning_rate": 1.72321069020963e-05, "loss": 0.7701, "step": 1306 }, { "epoch": 2.65650406504065, "grad_norm": 1.9066476136629082, "learning_rate": 1.7227558666760168e-05, "loss": 0.951, "step": 1307 }, { "epoch": 2.658536585365854, "grad_norm": 1.5353867508060055, "learning_rate": 1.7223007298928322e-05, "loss": 0.7368, "step": 1308 }, { "epoch": 2.660569105691057, "grad_norm": 1.873736348212854, "learning_rate": 1.7218452800573373e-05, "loss": 0.862, "step": 1309 }, { "epoch": 2.66260162601626, "grad_norm": 2.021154702652314, "learning_rate": 1.7213895173669284e-05, "loss": 0.8723, "step": 1310 }, { "epoch": 2.6646341463414633, "grad_norm": 1.7605560666067273, "learning_rate": 1.7209334420191376e-05, "loss": 0.8803, "step": 1311 }, { "epoch": 2.6666666666666665, "grad_norm": 1.8301672706122947, "learning_rate": 1.7204770542116326e-05, "loss": 0.8696, "step": 1312 }, { "epoch": 2.66869918699187, "grad_norm": 1.8214868694186008, "learning_rate": 1.720020354142216e-05, "loss": 0.8882, "step": 1313 }, { "epoch": 2.6707317073170733, "grad_norm": 1.592977532335379, "learning_rate": 1.7195633420088266e-05, "loss": 0.7299, "step": 1314 }, { "epoch": 2.6727642276422765, "grad_norm": 1.700644683504626, "learning_rate": 1.7191060180095375e-05, "loss": 0.8771, "step": 1315 }, { "epoch": 2.6747967479674797, "grad_norm": 1.7845185541282307, "learning_rate": 1.7186483823425582e-05, "loss": 0.826, "step": 1316 }, { "epoch": 2.676829268292683, "grad_norm": 1.4755662833194054, "learning_rate": 1.718190435206232e-05, "loss": 0.8592, "step": 1317 }, { "epoch": 2.678861788617886, "grad_norm": 1.8666113474684811, "learning_rate": 1.7177321767990377e-05, "loss": 0.8941, "step": 1318 }, { "epoch": 2.680894308943089, "grad_norm": 1.7845121151776588, "learning_rate": 1.7172736073195896e-05, "loss": 0.8563, "step": 1319 }, { "epoch": 2.682926829268293, "grad_norm": 2.104604540284695, "learning_rate": 1.7168147269666357e-05, "loss": 0.7676, "step": 1320 }, { "epoch": 2.684959349593496, "grad_norm": 1.8209973922954932, "learning_rate": 1.7163555359390594e-05, "loss": 0.8932, "step": 1321 }, { "epoch": 2.686991869918699, "grad_norm": 1.753269481883007, "learning_rate": 1.7158960344358785e-05, "loss": 0.9255, "step": 1322 }, { "epoch": 2.6890243902439024, "grad_norm": 2.0966700620635397, "learning_rate": 1.7154362226562464e-05, "loss": 0.8317, "step": 1323 }, { "epoch": 2.6910569105691056, "grad_norm": 1.906948259273836, "learning_rate": 1.714976100799449e-05, "loss": 0.7805, "step": 1324 }, { "epoch": 2.693089430894309, "grad_norm": 1.9350033493299499, "learning_rate": 1.7145156690649082e-05, "loss": 0.9793, "step": 1325 }, { "epoch": 2.6951219512195124, "grad_norm": 1.7831751670080205, "learning_rate": 1.7140549276521797e-05, "loss": 0.8019, "step": 1326 }, { "epoch": 2.6971544715447155, "grad_norm": 1.657758424193191, "learning_rate": 1.713593876760953e-05, "loss": 0.7891, "step": 1327 }, { "epoch": 2.6991869918699187, "grad_norm": 1.9466069984759953, "learning_rate": 1.713132516591053e-05, "loss": 0.8942, "step": 1328 }, { "epoch": 2.701219512195122, "grad_norm": 1.9066388370154628, "learning_rate": 1.7126708473424368e-05, "loss": 0.8651, "step": 1329 }, { "epoch": 2.703252032520325, "grad_norm": 1.8344074347664554, "learning_rate": 1.7122088692151974e-05, "loss": 0.7607, "step": 1330 }, { "epoch": 2.7052845528455283, "grad_norm": 1.8084720674287964, "learning_rate": 1.7117465824095598e-05, "loss": 0.7505, "step": 1331 }, { "epoch": 2.7073170731707314, "grad_norm": 1.6732558559698407, "learning_rate": 1.7112839871258838e-05, "loss": 0.591, "step": 1332 }, { "epoch": 2.709349593495935, "grad_norm": 1.6882050593075189, "learning_rate": 1.7108210835646636e-05, "loss": 0.9236, "step": 1333 }, { "epoch": 2.7113821138211383, "grad_norm": 2.001770362959608, "learning_rate": 1.7103578719265254e-05, "loss": 0.9124, "step": 1334 }, { "epoch": 2.7134146341463414, "grad_norm": 1.8405928049236713, "learning_rate": 1.70989435241223e-05, "loss": 0.7706, "step": 1335 }, { "epoch": 2.7154471544715446, "grad_norm": 1.7787533305758294, "learning_rate": 1.7094305252226713e-05, "loss": 0.9539, "step": 1336 }, { "epoch": 2.7174796747967482, "grad_norm": 1.7067207470713832, "learning_rate": 1.7089663905588767e-05, "loss": 0.9756, "step": 1337 }, { "epoch": 2.7195121951219514, "grad_norm": 1.7507511333413337, "learning_rate": 1.7085019486220068e-05, "loss": 0.9275, "step": 1338 }, { "epoch": 2.7215447154471546, "grad_norm": 2.050668739769618, "learning_rate": 1.7080371996133552e-05, "loss": 0.9895, "step": 1339 }, { "epoch": 2.7235772357723578, "grad_norm": 2.0514440583108122, "learning_rate": 1.7075721437343488e-05, "loss": 0.8284, "step": 1340 }, { "epoch": 2.725609756097561, "grad_norm": 1.7171869531673913, "learning_rate": 1.7071067811865477e-05, "loss": 0.7786, "step": 1341 }, { "epoch": 2.727642276422764, "grad_norm": 1.9373951843422426, "learning_rate": 1.7066411121716442e-05, "loss": 0.8458, "step": 1342 }, { "epoch": 2.7296747967479673, "grad_norm": 1.911677514884066, "learning_rate": 1.7061751368914642e-05, "loss": 0.8478, "step": 1343 }, { "epoch": 2.7317073170731705, "grad_norm": 1.8453077643690732, "learning_rate": 1.705708855547966e-05, "loss": 0.9711, "step": 1344 }, { "epoch": 2.733739837398374, "grad_norm": 1.7704566410976135, "learning_rate": 1.7052422683432402e-05, "loss": 0.9006, "step": 1345 }, { "epoch": 2.7357723577235773, "grad_norm": 2.149702593130116, "learning_rate": 1.7047753754795106e-05, "loss": 0.8355, "step": 1346 }, { "epoch": 2.7378048780487805, "grad_norm": 1.76860846410808, "learning_rate": 1.7043081771591335e-05, "loss": 0.9329, "step": 1347 }, { "epoch": 2.7398373983739837, "grad_norm": 1.7195960845238272, "learning_rate": 1.7038406735845967e-05, "loss": 1.0556, "step": 1348 }, { "epoch": 2.741869918699187, "grad_norm": 1.8505273562742584, "learning_rate": 1.7033728649585213e-05, "loss": 0.9061, "step": 1349 }, { "epoch": 2.7439024390243905, "grad_norm": 1.7953076993203385, "learning_rate": 1.70290475148366e-05, "loss": 0.9275, "step": 1350 }, { "epoch": 2.7459349593495936, "grad_norm": 1.5874154984649096, "learning_rate": 1.702436333362898e-05, "loss": 0.7776, "step": 1351 }, { "epoch": 2.747967479674797, "grad_norm": 2.075676612511462, "learning_rate": 1.7019676107992523e-05, "loss": 0.8787, "step": 1352 }, { "epoch": 2.75, "grad_norm": 1.9273732230073715, "learning_rate": 1.701498583995872e-05, "loss": 0.9313, "step": 1353 }, { "epoch": 2.752032520325203, "grad_norm": 1.4820215649825497, "learning_rate": 1.7010292531560378e-05, "loss": 0.8982, "step": 1354 }, { "epoch": 2.7540650406504064, "grad_norm": 1.7828046513616154, "learning_rate": 1.7005596184831623e-05, "loss": 0.8437, "step": 1355 }, { "epoch": 2.7560975609756095, "grad_norm": 2.034093524681604, "learning_rate": 1.70008968018079e-05, "loss": 0.9119, "step": 1356 }, { "epoch": 2.758130081300813, "grad_norm": 1.8863310300270357, "learning_rate": 1.6996194384525967e-05, "loss": 0.9748, "step": 1357 }, { "epoch": 2.7601626016260163, "grad_norm": 1.7094695771215123, "learning_rate": 1.69914889350239e-05, "loss": 0.8647, "step": 1358 }, { "epoch": 2.7621951219512195, "grad_norm": 2.0368310125518874, "learning_rate": 1.698678045534109e-05, "loss": 0.8612, "step": 1359 }, { "epoch": 2.7642276422764227, "grad_norm": 1.6092281049821302, "learning_rate": 1.6982068947518235e-05, "loss": 0.852, "step": 1360 }, { "epoch": 2.766260162601626, "grad_norm": 1.8046705727451329, "learning_rate": 1.6977354413597344e-05, "loss": 0.8254, "step": 1361 }, { "epoch": 2.7682926829268295, "grad_norm": 1.6157247210763062, "learning_rate": 1.6972636855621752e-05, "loss": 0.8917, "step": 1362 }, { "epoch": 2.7703252032520327, "grad_norm": 1.6931912754589287, "learning_rate": 1.6967916275636093e-05, "loss": 0.817, "step": 1363 }, { "epoch": 2.772357723577236, "grad_norm": 1.7934091923565116, "learning_rate": 1.6963192675686312e-05, "loss": 0.778, "step": 1364 }, { "epoch": 2.774390243902439, "grad_norm": 1.891067854208905, "learning_rate": 1.6958466057819666e-05, "loss": 0.9352, "step": 1365 }, { "epoch": 2.7764227642276422, "grad_norm": 2.0280466843697984, "learning_rate": 1.6953736424084716e-05, "loss": 0.9129, "step": 1366 }, { "epoch": 2.7784552845528454, "grad_norm": 1.500922374648234, "learning_rate": 1.694900377653133e-05, "loss": 0.7221, "step": 1367 }, { "epoch": 2.7804878048780486, "grad_norm": 1.6690013630994494, "learning_rate": 1.694426811721069e-05, "loss": 0.7238, "step": 1368 }, { "epoch": 2.7825203252032518, "grad_norm": 1.9077694256037703, "learning_rate": 1.6939529448175275e-05, "loss": 0.9348, "step": 1369 }, { "epoch": 2.7845528455284554, "grad_norm": 1.5404414796266144, "learning_rate": 1.6934787771478872e-05, "loss": 0.9244, "step": 1370 }, { "epoch": 2.7865853658536586, "grad_norm": 1.722773064809447, "learning_rate": 1.693004308917657e-05, "loss": 0.8018, "step": 1371 }, { "epoch": 2.7886178861788617, "grad_norm": 1.9631261982617638, "learning_rate": 1.6925295403324758e-05, "loss": 0.9527, "step": 1372 }, { "epoch": 2.790650406504065, "grad_norm": 1.8996447330913864, "learning_rate": 1.692054471598114e-05, "loss": 0.961, "step": 1373 }, { "epoch": 2.7926829268292686, "grad_norm": 1.564129028432555, "learning_rate": 1.69157910292047e-05, "loss": 0.8705, "step": 1374 }, { "epoch": 2.7947154471544717, "grad_norm": 1.618397738048459, "learning_rate": 1.6911034345055738e-05, "loss": 0.7521, "step": 1375 }, { "epoch": 2.796747967479675, "grad_norm": 1.7298170977310696, "learning_rate": 1.6906274665595854e-05, "loss": 0.8672, "step": 1376 }, { "epoch": 2.798780487804878, "grad_norm": 2.001843749340993, "learning_rate": 1.690151199288793e-05, "loss": 0.9731, "step": 1377 }, { "epoch": 2.8008130081300813, "grad_norm": 1.800710058608638, "learning_rate": 1.689674632899616e-05, "loss": 0.8625, "step": 1378 }, { "epoch": 2.8028455284552845, "grad_norm": 1.7201023675967981, "learning_rate": 1.6891977675986032e-05, "loss": 0.8053, "step": 1379 }, { "epoch": 2.8048780487804876, "grad_norm": 1.668433178730035, "learning_rate": 1.688720603592432e-05, "loss": 0.7715, "step": 1380 }, { "epoch": 2.806910569105691, "grad_norm": 1.6558031337600236, "learning_rate": 1.6882431410879112e-05, "loss": 0.7958, "step": 1381 }, { "epoch": 2.8089430894308944, "grad_norm": 1.8376890819247187, "learning_rate": 1.687765380291977e-05, "loss": 0.8786, "step": 1382 }, { "epoch": 2.8109756097560976, "grad_norm": 1.7881631632066124, "learning_rate": 1.687287321411696e-05, "loss": 0.8361, "step": 1383 }, { "epoch": 2.813008130081301, "grad_norm": 1.6353689741614341, "learning_rate": 1.6868089646542632e-05, "loss": 0.7077, "step": 1384 }, { "epoch": 2.815040650406504, "grad_norm": 2.1129044741516494, "learning_rate": 1.686330310227004e-05, "loss": 0.8464, "step": 1385 }, { "epoch": 2.817073170731707, "grad_norm": 1.7567810353000046, "learning_rate": 1.6858513583373712e-05, "loss": 1.048, "step": 1386 }, { "epoch": 2.819105691056911, "grad_norm": 1.7164614822063808, "learning_rate": 1.6853721091929477e-05, "loss": 0.75, "step": 1387 }, { "epoch": 2.821138211382114, "grad_norm": 1.805055194916085, "learning_rate": 1.6848925630014445e-05, "loss": 1.0499, "step": 1388 }, { "epoch": 2.823170731707317, "grad_norm": 1.8855851913499044, "learning_rate": 1.6844127199707022e-05, "loss": 0.6766, "step": 1389 }, { "epoch": 2.8252032520325203, "grad_norm": 1.7909196165122865, "learning_rate": 1.683932580308689e-05, "loss": 0.9691, "step": 1390 }, { "epoch": 2.8272357723577235, "grad_norm": 1.7783824161302582, "learning_rate": 1.6834521442235025e-05, "loss": 0.8517, "step": 1391 }, { "epoch": 2.8292682926829267, "grad_norm": 1.7972772425471666, "learning_rate": 1.6829714119233688e-05, "loss": 0.813, "step": 1392 }, { "epoch": 2.83130081300813, "grad_norm": 1.6512428161586807, "learning_rate": 1.682490383616641e-05, "loss": 0.9627, "step": 1393 }, { "epoch": 2.8333333333333335, "grad_norm": 1.7994599262461783, "learning_rate": 1.6820090595118025e-05, "loss": 0.7981, "step": 1394 }, { "epoch": 2.8353658536585367, "grad_norm": 1.7702725564831598, "learning_rate": 1.681527439817464e-05, "loss": 0.7847, "step": 1395 }, { "epoch": 2.83739837398374, "grad_norm": 1.8848325291069585, "learning_rate": 1.6810455247423634e-05, "loss": 0.9457, "step": 1396 }, { "epoch": 2.839430894308943, "grad_norm": 1.8538079167916017, "learning_rate": 1.6805633144953684e-05, "loss": 0.8244, "step": 1397 }, { "epoch": 2.841463414634146, "grad_norm": 2.196409324618555, "learning_rate": 1.680080809285473e-05, "loss": 0.8664, "step": 1398 }, { "epoch": 2.84349593495935, "grad_norm": 1.854524571378033, "learning_rate": 1.6795980093218004e-05, "loss": 0.8419, "step": 1399 }, { "epoch": 2.845528455284553, "grad_norm": 1.810301113755245, "learning_rate": 1.6791149148136003e-05, "loss": 0.931, "step": 1400 }, { "epoch": 2.847560975609756, "grad_norm": 1.7353381464805708, "learning_rate": 1.678631525970251e-05, "loss": 0.9167, "step": 1401 }, { "epoch": 2.8495934959349594, "grad_norm": 1.6213443126961555, "learning_rate": 1.6781478430012578e-05, "loss": 0.9104, "step": 1402 }, { "epoch": 2.8516260162601625, "grad_norm": 1.6859480061748093, "learning_rate": 1.677663866116254e-05, "loss": 0.7928, "step": 1403 }, { "epoch": 2.8536585365853657, "grad_norm": 2.026549171980612, "learning_rate": 1.677179595525e-05, "loss": 0.8942, "step": 1404 }, { "epoch": 2.855691056910569, "grad_norm": 1.800049557992789, "learning_rate": 1.6766950314373833e-05, "loss": 0.763, "step": 1405 }, { "epoch": 2.857723577235772, "grad_norm": 1.9887096841655885, "learning_rate": 1.6762101740634184e-05, "loss": 0.9628, "step": 1406 }, { "epoch": 2.8597560975609757, "grad_norm": 1.809409127943564, "learning_rate": 1.675725023613248e-05, "loss": 0.8935, "step": 1407 }, { "epoch": 2.861788617886179, "grad_norm": 1.6125151735312377, "learning_rate": 1.675239580297141e-05, "loss": 0.6885, "step": 1408 }, { "epoch": 2.863821138211382, "grad_norm": 1.6331147433814754, "learning_rate": 1.6747538443254926e-05, "loss": 0.8136, "step": 1409 }, { "epoch": 2.8658536585365852, "grad_norm": 1.8312171230932655, "learning_rate": 1.674267815908827e-05, "loss": 0.8586, "step": 1410 }, { "epoch": 2.867886178861789, "grad_norm": 1.8668173873751923, "learning_rate": 1.673781495257792e-05, "loss": 0.7168, "step": 1411 }, { "epoch": 2.869918699186992, "grad_norm": 2.112739939069473, "learning_rate": 1.6732948825831657e-05, "loss": 0.83, "step": 1412 }, { "epoch": 2.8719512195121952, "grad_norm": 1.636571807839052, "learning_rate": 1.6728079780958495e-05, "loss": 0.8114, "step": 1413 }, { "epoch": 2.8739837398373984, "grad_norm": 1.6738037270073411, "learning_rate": 1.6723207820068726e-05, "loss": 0.6946, "step": 1414 }, { "epoch": 2.8760162601626016, "grad_norm": 1.5899055451119637, "learning_rate": 1.671833294527392e-05, "loss": 0.8236, "step": 1415 }, { "epoch": 2.8780487804878048, "grad_norm": 1.734939360310716, "learning_rate": 1.671345515868688e-05, "loss": 0.7895, "step": 1416 }, { "epoch": 2.880081300813008, "grad_norm": 2.1727012716835867, "learning_rate": 1.6708574462421693e-05, "loss": 0.9727, "step": 1417 }, { "epoch": 2.882113821138211, "grad_norm": 1.8538979361506167, "learning_rate": 1.6703690858593704e-05, "loss": 0.8431, "step": 1418 }, { "epoch": 2.8841463414634148, "grad_norm": 1.974986533641898, "learning_rate": 1.6698804349319516e-05, "loss": 0.9871, "step": 1419 }, { "epoch": 2.886178861788618, "grad_norm": 1.6113072832933046, "learning_rate": 1.6693914936716983e-05, "loss": 0.8927, "step": 1420 }, { "epoch": 2.888211382113821, "grad_norm": 2.0946765026646847, "learning_rate": 1.6689022622905233e-05, "loss": 0.7648, "step": 1421 }, { "epoch": 2.8902439024390243, "grad_norm": 1.8021074169688696, "learning_rate": 1.6684127410004644e-05, "loss": 0.9348, "step": 1422 }, { "epoch": 2.892276422764228, "grad_norm": 2.3444561989764656, "learning_rate": 1.6679229300136842e-05, "loss": 0.8348, "step": 1423 }, { "epoch": 2.894308943089431, "grad_norm": 1.8121563010555026, "learning_rate": 1.6674328295424723e-05, "loss": 0.8993, "step": 1424 }, { "epoch": 2.8963414634146343, "grad_norm": 1.6392674570997934, "learning_rate": 1.666942439799243e-05, "loss": 0.8625, "step": 1425 }, { "epoch": 2.8983739837398375, "grad_norm": 1.9442303047299183, "learning_rate": 1.6664517609965355e-05, "loss": 0.7231, "step": 1426 }, { "epoch": 2.9004065040650406, "grad_norm": 1.7813062511741349, "learning_rate": 1.665960793347016e-05, "loss": 1.0486, "step": 1427 }, { "epoch": 2.902439024390244, "grad_norm": 1.6421008484340496, "learning_rate": 1.6654695370634738e-05, "loss": 0.7843, "step": 1428 }, { "epoch": 2.904471544715447, "grad_norm": 1.9429033828090623, "learning_rate": 1.6649779923588246e-05, "loss": 0.9129, "step": 1429 }, { "epoch": 2.90650406504065, "grad_norm": 1.576398216801647, "learning_rate": 1.6644861594461086e-05, "loss": 0.7536, "step": 1430 }, { "epoch": 2.908536585365854, "grad_norm": 1.804253756643881, "learning_rate": 1.6639940385384913e-05, "loss": 0.8429, "step": 1431 }, { "epoch": 2.910569105691057, "grad_norm": 1.649803680155989, "learning_rate": 1.6635016298492628e-05, "loss": 0.8908, "step": 1432 }, { "epoch": 2.91260162601626, "grad_norm": 2.1707027818488203, "learning_rate": 1.6630089335918374e-05, "loss": 0.8798, "step": 1433 }, { "epoch": 2.9146341463414633, "grad_norm": 1.9549674742783472, "learning_rate": 1.6625159499797553e-05, "loss": 0.8427, "step": 1434 }, { "epoch": 2.9166666666666665, "grad_norm": 1.7932386053663498, "learning_rate": 1.6620226792266804e-05, "loss": 0.9562, "step": 1435 }, { "epoch": 2.91869918699187, "grad_norm": 1.8608019501718633, "learning_rate": 1.6615291215464005e-05, "loss": 0.9816, "step": 1436 }, { "epoch": 2.9207317073170733, "grad_norm": 1.940327829301988, "learning_rate": 1.661035277152829e-05, "loss": 0.7494, "step": 1437 }, { "epoch": 2.9227642276422765, "grad_norm": 1.9941683627513382, "learning_rate": 1.6605411462600023e-05, "loss": 0.8071, "step": 1438 }, { "epoch": 2.9247967479674797, "grad_norm": 1.616059096380054, "learning_rate": 1.660046729082083e-05, "loss": 0.7864, "step": 1439 }, { "epoch": 2.926829268292683, "grad_norm": 1.6649089898910636, "learning_rate": 1.6595520258333545e-05, "loss": 0.8465, "step": 1440 }, { "epoch": 2.928861788617886, "grad_norm": 1.907511331295219, "learning_rate": 1.6590570367282276e-05, "loss": 0.8268, "step": 1441 }, { "epoch": 2.930894308943089, "grad_norm": 1.6291508148606038, "learning_rate": 1.658561761981235e-05, "loss": 0.6541, "step": 1442 }, { "epoch": 2.932926829268293, "grad_norm": 2.3037002387679126, "learning_rate": 1.6580662018070338e-05, "loss": 0.9726, "step": 1443 }, { "epoch": 2.934959349593496, "grad_norm": 1.8484599288584054, "learning_rate": 1.657570356420404e-05, "loss": 0.8592, "step": 1444 }, { "epoch": 2.936991869918699, "grad_norm": 1.795162306730542, "learning_rate": 1.657074226036251e-05, "loss": 0.9606, "step": 1445 }, { "epoch": 2.9390243902439024, "grad_norm": 1.882446994726848, "learning_rate": 1.656577810869602e-05, "loss": 0.69, "step": 1446 }, { "epoch": 2.9410569105691056, "grad_norm": 1.9082123850719444, "learning_rate": 1.6560811111356087e-05, "loss": 0.9575, "step": 1447 }, { "epoch": 2.943089430894309, "grad_norm": 1.689387902516079, "learning_rate": 1.6555841270495456e-05, "loss": 0.6781, "step": 1448 }, { "epoch": 2.9451219512195124, "grad_norm": 1.8637932508610762, "learning_rate": 1.65508685882681e-05, "loss": 0.7407, "step": 1449 }, { "epoch": 2.9471544715447155, "grad_norm": 1.7678264323920982, "learning_rate": 1.654589306682923e-05, "loss": 0.84, "step": 1450 }, { "epoch": 2.9491869918699187, "grad_norm": 2.0686968883839283, "learning_rate": 1.654091470833529e-05, "loss": 0.6978, "step": 1451 }, { "epoch": 2.951219512195122, "grad_norm": 2.084376428096746, "learning_rate": 1.6535933514943955e-05, "loss": 0.8883, "step": 1452 }, { "epoch": 2.953252032520325, "grad_norm": 1.7415982488930928, "learning_rate": 1.6530949488814113e-05, "loss": 0.6296, "step": 1453 }, { "epoch": 2.9552845528455283, "grad_norm": 1.9907128161597905, "learning_rate": 1.6525962632105892e-05, "loss": 0.9462, "step": 1454 }, { "epoch": 2.9573170731707314, "grad_norm": 1.8814380867506704, "learning_rate": 1.652097294698065e-05, "loss": 0.813, "step": 1455 }, { "epoch": 2.959349593495935, "grad_norm": 1.9012589065257868, "learning_rate": 1.6515980435600965e-05, "loss": 0.7322, "step": 1456 }, { "epoch": 2.9613821138211383, "grad_norm": 1.9899614495310793, "learning_rate": 1.6510985100130642e-05, "loss": 0.8745, "step": 1457 }, { "epoch": 2.9634146341463414, "grad_norm": 1.661956963921339, "learning_rate": 1.6505986942734703e-05, "loss": 0.6304, "step": 1458 }, { "epoch": 2.9654471544715446, "grad_norm": 2.0647744351276254, "learning_rate": 1.650098596557941e-05, "loss": 0.925, "step": 1459 }, { "epoch": 2.9674796747967482, "grad_norm": 1.6578840962761594, "learning_rate": 1.6495982170832224e-05, "loss": 0.7996, "step": 1460 }, { "epoch": 2.9695121951219514, "grad_norm": 1.79258139735451, "learning_rate": 1.6490975560661845e-05, "loss": 0.882, "step": 1461 }, { "epoch": 2.9715447154471546, "grad_norm": 1.7380910014783246, "learning_rate": 1.648596613723819e-05, "loss": 0.816, "step": 1462 }, { "epoch": 2.9735772357723578, "grad_norm": 2.0184630721055656, "learning_rate": 1.648095390273239e-05, "loss": 0.7594, "step": 1463 }, { "epoch": 2.975609756097561, "grad_norm": 1.7727265612449783, "learning_rate": 1.6475938859316795e-05, "loss": 0.7924, "step": 1464 }, { "epoch": 2.977642276422764, "grad_norm": 2.473729104895325, "learning_rate": 1.647092100916498e-05, "loss": 0.9644, "step": 1465 }, { "epoch": 2.9796747967479673, "grad_norm": 1.8549931426116513, "learning_rate": 1.6465900354451726e-05, "loss": 0.8109, "step": 1466 }, { "epoch": 2.9817073170731705, "grad_norm": 1.7743414807134803, "learning_rate": 1.646087689735304e-05, "loss": 0.9245, "step": 1467 }, { "epoch": 2.983739837398374, "grad_norm": 2.140473719569873, "learning_rate": 1.6455850640046134e-05, "loss": 0.9394, "step": 1468 }, { "epoch": 2.9857723577235773, "grad_norm": 1.8929876538162425, "learning_rate": 1.6450821584709438e-05, "loss": 0.831, "step": 1469 }, { "epoch": 2.9878048780487805, "grad_norm": 1.7418783751557012, "learning_rate": 1.6445789733522595e-05, "loss": 1.0545, "step": 1470 }, { "epoch": 2.9898373983739837, "grad_norm": 1.8421088482846497, "learning_rate": 1.6440755088666457e-05, "loss": 0.6863, "step": 1471 }, { "epoch": 2.991869918699187, "grad_norm": 2.146840760559921, "learning_rate": 1.6435717652323097e-05, "loss": 1.0051, "step": 1472 }, { "epoch": 2.9939024390243905, "grad_norm": 1.9365018056705057, "learning_rate": 1.6430677426675787e-05, "loss": 0.8572, "step": 1473 }, { "epoch": 2.9959349593495936, "grad_norm": 2.129255121646729, "learning_rate": 1.6425634413909003e-05, "loss": 0.8586, "step": 1474 }, { "epoch": 2.997967479674797, "grad_norm": 2.045285673966025, "learning_rate": 1.6420588616208443e-05, "loss": 0.7447, "step": 1475 }, { "epoch": 3.0, "grad_norm": 2.070430351801844, "learning_rate": 1.6415540035761008e-05, "loss": 0.9241, "step": 1476 }, { "epoch": 3.0, "eval_loss": 0.943085789680481, "eval_runtime": 132.2356, "eval_samples_per_second": 7.441, "eval_steps_per_second": 0.93, "step": 1476 }, { "epoch": 3.002032520325203, "grad_norm": 1.8788473386449596, "learning_rate": 1.6410488674754796e-05, "loss": 0.6711, "step": 1477 }, { "epoch": 3.0040650406504064, "grad_norm": 2.2593611706049956, "learning_rate": 1.6405434535379124e-05, "loss": 0.7026, "step": 1478 }, { "epoch": 3.0060975609756095, "grad_norm": 1.8810592893967435, "learning_rate": 1.6400377619824504e-05, "loss": 0.6005, "step": 1479 }, { "epoch": 3.008130081300813, "grad_norm": 1.8123611795773409, "learning_rate": 1.639531793028265e-05, "loss": 0.5962, "step": 1480 }, { "epoch": 3.0101626016260163, "grad_norm": 1.859388860210283, "learning_rate": 1.6390255468946484e-05, "loss": 0.7079, "step": 1481 }, { "epoch": 3.0121951219512195, "grad_norm": 1.481474362032175, "learning_rate": 1.6385190238010132e-05, "loss": 0.7869, "step": 1482 }, { "epoch": 3.0142276422764227, "grad_norm": 1.937739650229458, "learning_rate": 1.6380122239668907e-05, "loss": 0.7234, "step": 1483 }, { "epoch": 3.016260162601626, "grad_norm": 1.8516146419508954, "learning_rate": 1.637505147611934e-05, "loss": 0.7833, "step": 1484 }, { "epoch": 3.018292682926829, "grad_norm": 1.8926876098661851, "learning_rate": 1.6369977949559138e-05, "loss": 0.6887, "step": 1485 }, { "epoch": 3.0203252032520327, "grad_norm": 1.8160051362924408, "learning_rate": 1.6364901662187227e-05, "loss": 0.6424, "step": 1486 }, { "epoch": 3.022357723577236, "grad_norm": 2.0239785792897593, "learning_rate": 1.6359822616203715e-05, "loss": 0.7426, "step": 1487 }, { "epoch": 3.024390243902439, "grad_norm": 2.1871191384792557, "learning_rate": 1.6354740813809917e-05, "loss": 0.8102, "step": 1488 }, { "epoch": 3.0264227642276422, "grad_norm": 1.9113549315334324, "learning_rate": 1.6349656257208327e-05, "loss": 0.6863, "step": 1489 }, { "epoch": 3.0284552845528454, "grad_norm": 2.164877232757629, "learning_rate": 1.6344568948602652e-05, "loss": 0.746, "step": 1490 }, { "epoch": 3.0304878048780486, "grad_norm": 1.754086819503603, "learning_rate": 1.6339478890197778e-05, "loss": 0.763, "step": 1491 }, { "epoch": 3.032520325203252, "grad_norm": 2.3495137912236177, "learning_rate": 1.6334386084199787e-05, "loss": 0.8732, "step": 1492 }, { "epoch": 3.0345528455284554, "grad_norm": 2.0403769635161875, "learning_rate": 1.632929053281595e-05, "loss": 0.7155, "step": 1493 }, { "epoch": 3.0365853658536586, "grad_norm": 1.9836833857632195, "learning_rate": 1.6324192238254735e-05, "loss": 0.6717, "step": 1494 }, { "epoch": 3.0386178861788617, "grad_norm": 1.9716191233795388, "learning_rate": 1.631909120272579e-05, "loss": 0.7073, "step": 1495 }, { "epoch": 3.040650406504065, "grad_norm": 2.038358083556779, "learning_rate": 1.631398742843995e-05, "loss": 0.6419, "step": 1496 }, { "epoch": 3.042682926829268, "grad_norm": 2.595445041200764, "learning_rate": 1.6308880917609248e-05, "loss": 0.9239, "step": 1497 }, { "epoch": 3.0447154471544717, "grad_norm": 2.091725963646514, "learning_rate": 1.6303771672446896e-05, "loss": 0.7336, "step": 1498 }, { "epoch": 3.046747967479675, "grad_norm": 2.1311915553901923, "learning_rate": 1.6298659695167287e-05, "loss": 0.831, "step": 1499 }, { "epoch": 3.048780487804878, "grad_norm": 2.006284715844999, "learning_rate": 1.629354498798601e-05, "loss": 0.7642, "step": 1500 }, { "epoch": 3.0508130081300813, "grad_norm": 2.076683349142068, "learning_rate": 1.628842755311982e-05, "loss": 0.6291, "step": 1501 }, { "epoch": 3.0528455284552845, "grad_norm": 2.2442979029425345, "learning_rate": 1.6283307392786674e-05, "loss": 0.7509, "step": 1502 }, { "epoch": 3.0548780487804876, "grad_norm": 2.16185485411131, "learning_rate": 1.6278184509205693e-05, "loss": 0.8757, "step": 1503 }, { "epoch": 3.0569105691056913, "grad_norm": 2.3134334965667076, "learning_rate": 1.627305890459719e-05, "loss": 0.6486, "step": 1504 }, { "epoch": 3.0589430894308944, "grad_norm": 1.9931083015087687, "learning_rate": 1.6267930581182654e-05, "loss": 0.7532, "step": 1505 }, { "epoch": 3.0609756097560976, "grad_norm": 2.0131083652209956, "learning_rate": 1.626279954118475e-05, "loss": 0.8438, "step": 1506 }, { "epoch": 3.063008130081301, "grad_norm": 2.3951095560000297, "learning_rate": 1.6257665786827315e-05, "loss": 0.8444, "step": 1507 }, { "epoch": 3.065040650406504, "grad_norm": 1.9451474419452472, "learning_rate": 1.625252932033538e-05, "loss": 0.8547, "step": 1508 }, { "epoch": 3.067073170731707, "grad_norm": 2.169931656767241, "learning_rate": 1.624739014393513e-05, "loss": 0.6325, "step": 1509 }, { "epoch": 3.069105691056911, "grad_norm": 2.242776050141578, "learning_rate": 1.6242248259853943e-05, "loss": 0.8329, "step": 1510 }, { "epoch": 3.071138211382114, "grad_norm": 2.232347743375029, "learning_rate": 1.623710367032036e-05, "loss": 0.7349, "step": 1511 }, { "epoch": 3.073170731707317, "grad_norm": 2.2600120388965323, "learning_rate": 1.6231956377564095e-05, "loss": 0.6987, "step": 1512 }, { "epoch": 3.0752032520325203, "grad_norm": 2.1885272191122267, "learning_rate": 1.622680638381604e-05, "loss": 0.7281, "step": 1513 }, { "epoch": 3.0772357723577235, "grad_norm": 2.0042663818317523, "learning_rate": 1.6221653691308253e-05, "loss": 0.8348, "step": 1514 }, { "epoch": 3.0792682926829267, "grad_norm": 2.3946522029468276, "learning_rate": 1.621649830227396e-05, "loss": 0.7646, "step": 1515 }, { "epoch": 3.08130081300813, "grad_norm": 2.245941054042891, "learning_rate": 1.621134021894756e-05, "loss": 0.7152, "step": 1516 }, { "epoch": 3.0833333333333335, "grad_norm": 2.4919649942581286, "learning_rate": 1.6206179443564615e-05, "loss": 0.753, "step": 1517 }, { "epoch": 3.0853658536585367, "grad_norm": 2.069602506083772, "learning_rate": 1.6201015978361852e-05, "loss": 0.8215, "step": 1518 }, { "epoch": 3.08739837398374, "grad_norm": 2.092251708568793, "learning_rate": 1.6195849825577178e-05, "loss": 0.8825, "step": 1519 }, { "epoch": 3.089430894308943, "grad_norm": 1.984886862909424, "learning_rate": 1.619068098744965e-05, "loss": 0.8446, "step": 1520 }, { "epoch": 3.091463414634146, "grad_norm": 2.1810724032247806, "learning_rate": 1.6185509466219495e-05, "loss": 0.733, "step": 1521 }, { "epoch": 3.0934959349593494, "grad_norm": 2.071935542183464, "learning_rate": 1.61803352641281e-05, "loss": 0.7828, "step": 1522 }, { "epoch": 3.095528455284553, "grad_norm": 2.202691761683876, "learning_rate": 1.6175158383418014e-05, "loss": 0.7344, "step": 1523 }, { "epoch": 3.097560975609756, "grad_norm": 2.2394383182745448, "learning_rate": 1.6169978826332955e-05, "loss": 0.8029, "step": 1524 }, { "epoch": 3.0995934959349594, "grad_norm": 2.5575797490445327, "learning_rate": 1.616479659511779e-05, "loss": 0.7911, "step": 1525 }, { "epoch": 3.1016260162601625, "grad_norm": 2.0947821867608005, "learning_rate": 1.615961169201855e-05, "loss": 0.8751, "step": 1526 }, { "epoch": 3.1036585365853657, "grad_norm": 2.134832876627905, "learning_rate": 1.6154424119282417e-05, "loss": 0.7665, "step": 1527 }, { "epoch": 3.105691056910569, "grad_norm": 2.0256575791208333, "learning_rate": 1.6149233879157747e-05, "loss": 0.735, "step": 1528 }, { "epoch": 3.1077235772357725, "grad_norm": 2.2499478699196906, "learning_rate": 1.614404097389404e-05, "loss": 0.7749, "step": 1529 }, { "epoch": 3.1097560975609757, "grad_norm": 2.124546361386739, "learning_rate": 1.6138845405741946e-05, "loss": 0.7693, "step": 1530 }, { "epoch": 3.111788617886179, "grad_norm": 2.3562685694635257, "learning_rate": 1.6133647176953283e-05, "loss": 0.8193, "step": 1531 }, { "epoch": 3.113821138211382, "grad_norm": 2.1866552968635746, "learning_rate": 1.6128446289781012e-05, "loss": 0.8193, "step": 1532 }, { "epoch": 3.1158536585365852, "grad_norm": 2.3137415586792263, "learning_rate": 1.612324274647925e-05, "loss": 0.8737, "step": 1533 }, { "epoch": 3.1178861788617884, "grad_norm": 1.7533141089005766, "learning_rate": 1.611803654930326e-05, "loss": 0.8403, "step": 1534 }, { "epoch": 3.119918699186992, "grad_norm": 2.150710329290202, "learning_rate": 1.6112827700509468e-05, "loss": 0.6993, "step": 1535 }, { "epoch": 3.1219512195121952, "grad_norm": 2.3375733230776197, "learning_rate": 1.610761620235543e-05, "loss": 0.7861, "step": 1536 }, { "epoch": 3.1239837398373984, "grad_norm": 2.024823628361143, "learning_rate": 1.6102402057099872e-05, "loss": 0.7812, "step": 1537 }, { "epoch": 3.1260162601626016, "grad_norm": 2.2458531257938934, "learning_rate": 1.609718526700265e-05, "loss": 0.7352, "step": 1538 }, { "epoch": 3.1280487804878048, "grad_norm": 1.8578404905832935, "learning_rate": 1.6091965834324775e-05, "loss": 0.7442, "step": 1539 }, { "epoch": 3.130081300813008, "grad_norm": 1.8152049190890092, "learning_rate": 1.60867437613284e-05, "loss": 0.5988, "step": 1540 }, { "epoch": 3.1321138211382116, "grad_norm": 2.2476084954165225, "learning_rate": 1.6081519050276817e-05, "loss": 0.6822, "step": 1541 }, { "epoch": 3.1341463414634148, "grad_norm": 1.861350623636714, "learning_rate": 1.6076291703434482e-05, "loss": 0.6752, "step": 1542 }, { "epoch": 3.136178861788618, "grad_norm": 2.533780471637948, "learning_rate": 1.6071061723066964e-05, "loss": 0.8114, "step": 1543 }, { "epoch": 3.138211382113821, "grad_norm": 2.013758280491651, "learning_rate": 1.6065829111441e-05, "loss": 0.7613, "step": 1544 }, { "epoch": 3.1402439024390243, "grad_norm": 2.4172485483740735, "learning_rate": 1.606059387082445e-05, "loss": 0.7057, "step": 1545 }, { "epoch": 3.1422764227642275, "grad_norm": 2.134981214548729, "learning_rate": 1.6055356003486318e-05, "loss": 0.9011, "step": 1546 }, { "epoch": 3.144308943089431, "grad_norm": 2.1066459377859252, "learning_rate": 1.605011551169676e-05, "loss": 0.6445, "step": 1547 }, { "epoch": 3.1463414634146343, "grad_norm": 2.526676221339888, "learning_rate": 1.6044872397727037e-05, "loss": 0.7788, "step": 1548 }, { "epoch": 3.1483739837398375, "grad_norm": 2.020853092369005, "learning_rate": 1.6039626663849583e-05, "loss": 0.7594, "step": 1549 }, { "epoch": 3.1504065040650406, "grad_norm": 2.08968425863876, "learning_rate": 1.6034378312337938e-05, "loss": 0.919, "step": 1550 }, { "epoch": 3.152439024390244, "grad_norm": 1.785198997758361, "learning_rate": 1.6029127345466805e-05, "loss": 0.7187, "step": 1551 }, { "epoch": 3.154471544715447, "grad_norm": 1.917623804693256, "learning_rate": 1.6023873765511993e-05, "loss": 0.7064, "step": 1552 }, { "epoch": 3.15650406504065, "grad_norm": 1.9904353907046126, "learning_rate": 1.601861757475046e-05, "loss": 0.7853, "step": 1553 }, { "epoch": 3.158536585365854, "grad_norm": 1.7949962747194081, "learning_rate": 1.6013358775460293e-05, "loss": 0.75, "step": 1554 }, { "epoch": 3.160569105691057, "grad_norm": 1.7296204649367068, "learning_rate": 1.6008097369920706e-05, "loss": 0.8154, "step": 1555 }, { "epoch": 3.16260162601626, "grad_norm": 1.8729547601979513, "learning_rate": 1.6002833360412044e-05, "loss": 0.627, "step": 1556 }, { "epoch": 3.1646341463414633, "grad_norm": 1.867824430557775, "learning_rate": 1.5997566749215778e-05, "loss": 0.6144, "step": 1557 }, { "epoch": 3.1666666666666665, "grad_norm": 2.113167222829037, "learning_rate": 1.5992297538614517e-05, "loss": 0.8069, "step": 1558 }, { "epoch": 3.16869918699187, "grad_norm": 1.9478386747711205, "learning_rate": 1.5987025730891984e-05, "loss": 0.6763, "step": 1559 }, { "epoch": 3.1707317073170733, "grad_norm": 1.9092382570141868, "learning_rate": 1.5981751328333036e-05, "loss": 0.6952, "step": 1560 }, { "epoch": 3.1727642276422765, "grad_norm": 2.1483820738985187, "learning_rate": 1.5976474333223647e-05, "loss": 0.8381, "step": 1561 }, { "epoch": 3.1747967479674797, "grad_norm": 2.2163812209686418, "learning_rate": 1.597119474785092e-05, "loss": 0.8616, "step": 1562 }, { "epoch": 3.176829268292683, "grad_norm": 2.0440370498447593, "learning_rate": 1.5965912574503083e-05, "loss": 0.7149, "step": 1563 }, { "epoch": 3.178861788617886, "grad_norm": 2.263724717708822, "learning_rate": 1.5960627815469486e-05, "loss": 0.6828, "step": 1564 }, { "epoch": 3.180894308943089, "grad_norm": 1.9378109491248132, "learning_rate": 1.5955340473040585e-05, "loss": 0.7271, "step": 1565 }, { "epoch": 3.182926829268293, "grad_norm": 1.8434620311373608, "learning_rate": 1.5950050549507974e-05, "loss": 0.8125, "step": 1566 }, { "epoch": 3.184959349593496, "grad_norm": 1.9795132291100914, "learning_rate": 1.5944758047164355e-05, "loss": 0.791, "step": 1567 }, { "epoch": 3.186991869918699, "grad_norm": 2.1643560620770437, "learning_rate": 1.5939462968303554e-05, "loss": 0.8053, "step": 1568 }, { "epoch": 3.1890243902439024, "grad_norm": 2.3063614371859127, "learning_rate": 1.593416531522051e-05, "loss": 0.773, "step": 1569 }, { "epoch": 3.1910569105691056, "grad_norm": 2.1419248065963994, "learning_rate": 1.592886509021128e-05, "loss": 0.747, "step": 1570 }, { "epoch": 3.1930894308943087, "grad_norm": 2.0954480276088296, "learning_rate": 1.5923562295573032e-05, "loss": 0.5913, "step": 1571 }, { "epoch": 3.1951219512195124, "grad_norm": 2.251502748918036, "learning_rate": 1.5918256933604047e-05, "loss": 0.7549, "step": 1572 }, { "epoch": 3.1971544715447155, "grad_norm": 2.035217947258066, "learning_rate": 1.5912949006603732e-05, "loss": 0.7838, "step": 1573 }, { "epoch": 3.1991869918699187, "grad_norm": 2.526002600157204, "learning_rate": 1.590763851687258e-05, "loss": 0.7082, "step": 1574 }, { "epoch": 3.201219512195122, "grad_norm": 2.7649661400686973, "learning_rate": 1.5902325466712223e-05, "loss": 0.841, "step": 1575 }, { "epoch": 3.203252032520325, "grad_norm": 2.638199803983525, "learning_rate": 1.589700985842538e-05, "loss": 0.8677, "step": 1576 }, { "epoch": 3.2052845528455283, "grad_norm": 2.269272927430384, "learning_rate": 1.58916916943159e-05, "loss": 0.7714, "step": 1577 }, { "epoch": 3.207317073170732, "grad_norm": 2.3597977955925495, "learning_rate": 1.5886370976688716e-05, "loss": 0.7693, "step": 1578 }, { "epoch": 3.209349593495935, "grad_norm": 2.6970433251088295, "learning_rate": 1.5881047707849887e-05, "loss": 0.8855, "step": 1579 }, { "epoch": 3.2113821138211383, "grad_norm": 2.1267801105585526, "learning_rate": 1.5875721890106574e-05, "loss": 0.6791, "step": 1580 }, { "epoch": 3.2134146341463414, "grad_norm": 2.2630195184631208, "learning_rate": 1.5870393525767025e-05, "loss": 0.9316, "step": 1581 }, { "epoch": 3.2154471544715446, "grad_norm": 2.2656488000847075, "learning_rate": 1.586506261714062e-05, "loss": 0.7898, "step": 1582 }, { "epoch": 3.217479674796748, "grad_norm": 2.3110974779992612, "learning_rate": 1.5859729166537824e-05, "loss": 0.7503, "step": 1583 }, { "epoch": 3.2195121951219514, "grad_norm": 2.180693580599105, "learning_rate": 1.5854393176270205e-05, "loss": 0.6463, "step": 1584 }, { "epoch": 3.2215447154471546, "grad_norm": 2.350361676266963, "learning_rate": 1.5849054648650436e-05, "loss": 0.8737, "step": 1585 }, { "epoch": 3.2235772357723578, "grad_norm": 1.8527271999149693, "learning_rate": 1.5843713585992285e-05, "loss": 0.6493, "step": 1586 }, { "epoch": 3.225609756097561, "grad_norm": 2.043057472771518, "learning_rate": 1.5838369990610628e-05, "loss": 0.8638, "step": 1587 }, { "epoch": 3.227642276422764, "grad_norm": 2.7147708457837805, "learning_rate": 1.5833023864821427e-05, "loss": 0.9631, "step": 1588 }, { "epoch": 3.2296747967479673, "grad_norm": 2.072133213405817, "learning_rate": 1.5827675210941747e-05, "loss": 0.7238, "step": 1589 }, { "epoch": 3.231707317073171, "grad_norm": 2.165257036934692, "learning_rate": 1.5822324031289748e-05, "loss": 0.803, "step": 1590 }, { "epoch": 3.233739837398374, "grad_norm": 2.5161814472778583, "learning_rate": 1.5816970328184683e-05, "loss": 0.7412, "step": 1591 }, { "epoch": 3.2357723577235773, "grad_norm": 2.2214194649477106, "learning_rate": 1.5811614103946905e-05, "loss": 0.7843, "step": 1592 }, { "epoch": 3.2378048780487805, "grad_norm": 2.373407880503463, "learning_rate": 1.580625536089785e-05, "loss": 0.815, "step": 1593 }, { "epoch": 3.2398373983739837, "grad_norm": 2.7088414269969934, "learning_rate": 1.5800894101360052e-05, "loss": 0.8972, "step": 1594 }, { "epoch": 3.241869918699187, "grad_norm": 2.3567769587721044, "learning_rate": 1.5795530327657132e-05, "loss": 0.822, "step": 1595 }, { "epoch": 3.2439024390243905, "grad_norm": 2.3423575037534055, "learning_rate": 1.5790164042113805e-05, "loss": 0.6806, "step": 1596 }, { "epoch": 3.2459349593495936, "grad_norm": 2.277124449506264, "learning_rate": 1.5784795247055876e-05, "loss": 0.8688, "step": 1597 }, { "epoch": 3.247967479674797, "grad_norm": 1.967086168431996, "learning_rate": 1.577942394481023e-05, "loss": 0.7081, "step": 1598 }, { "epoch": 3.25, "grad_norm": 2.0798585973083803, "learning_rate": 1.577405013770484e-05, "loss": 0.6665, "step": 1599 }, { "epoch": 3.252032520325203, "grad_norm": 2.6276434009906735, "learning_rate": 1.576867382806877e-05, "loss": 0.7424, "step": 1600 }, { "epoch": 3.2540650406504064, "grad_norm": 2.2441442800701057, "learning_rate": 1.5763295018232172e-05, "loss": 0.756, "step": 1601 }, { "epoch": 3.2560975609756095, "grad_norm": 2.5783940773457457, "learning_rate": 1.5757913710526265e-05, "loss": 0.7413, "step": 1602 }, { "epoch": 3.258130081300813, "grad_norm": 2.1163482531064703, "learning_rate": 1.5752529907283368e-05, "loss": 0.7346, "step": 1603 }, { "epoch": 3.2601626016260163, "grad_norm": 2.1151370386984634, "learning_rate": 1.5747143610836873e-05, "loss": 0.7423, "step": 1604 }, { "epoch": 3.2621951219512195, "grad_norm": 2.280962607587702, "learning_rate": 1.5741754823521254e-05, "loss": 0.8608, "step": 1605 }, { "epoch": 3.2642276422764227, "grad_norm": 1.910689043871177, "learning_rate": 1.5736363547672065e-05, "loss": 0.7316, "step": 1606 }, { "epoch": 3.266260162601626, "grad_norm": 2.3733311813242413, "learning_rate": 1.5730969785625943e-05, "loss": 0.7618, "step": 1607 }, { "epoch": 3.2682926829268295, "grad_norm": 2.2278731236025813, "learning_rate": 1.5725573539720592e-05, "loss": 0.6828, "step": 1608 }, { "epoch": 3.2703252032520327, "grad_norm": 2.245334211343595, "learning_rate": 1.5720174812294796e-05, "loss": 0.8858, "step": 1609 }, { "epoch": 3.272357723577236, "grad_norm": 2.205070010880492, "learning_rate": 1.5714773605688426e-05, "loss": 0.6639, "step": 1610 }, { "epoch": 3.274390243902439, "grad_norm": 2.1771845045649147, "learning_rate": 1.5709369922242413e-05, "loss": 0.8939, "step": 1611 }, { "epoch": 3.2764227642276422, "grad_norm": 2.1989706962459876, "learning_rate": 1.570396376429877e-05, "loss": 0.7551, "step": 1612 }, { "epoch": 3.2784552845528454, "grad_norm": 2.6473609921863948, "learning_rate": 1.5698555134200576e-05, "loss": 0.6771, "step": 1613 }, { "epoch": 3.2804878048780486, "grad_norm": 2.0112820905040185, "learning_rate": 1.5693144034291984e-05, "loss": 0.7702, "step": 1614 }, { "epoch": 3.282520325203252, "grad_norm": 2.344208057228964, "learning_rate": 1.5687730466918226e-05, "loss": 0.6598, "step": 1615 }, { "epoch": 3.2845528455284554, "grad_norm": 2.6066116886263924, "learning_rate": 1.5682314434425593e-05, "loss": 0.84, "step": 1616 }, { "epoch": 3.2865853658536586, "grad_norm": 2.072780856248526, "learning_rate": 1.567689593916144e-05, "loss": 0.7875, "step": 1617 }, { "epoch": 3.2886178861788617, "grad_norm": 2.671402719780134, "learning_rate": 1.5671474983474203e-05, "loss": 0.94, "step": 1618 }, { "epoch": 3.290650406504065, "grad_norm": 2.172263529808615, "learning_rate": 1.566605156971338e-05, "loss": 0.744, "step": 1619 }, { "epoch": 3.292682926829268, "grad_norm": 2.195020004212121, "learning_rate": 1.5660625700229526e-05, "loss": 0.8118, "step": 1620 }, { "epoch": 3.2947154471544717, "grad_norm": 2.439090269549893, "learning_rate": 1.5655197377374273e-05, "loss": 0.7721, "step": 1621 }, { "epoch": 3.296747967479675, "grad_norm": 2.475968366210411, "learning_rate": 1.564976660350031e-05, "loss": 0.6887, "step": 1622 }, { "epoch": 3.298780487804878, "grad_norm": 2.2375528107095453, "learning_rate": 1.564433338096138e-05, "loss": 0.9012, "step": 1623 }, { "epoch": 3.3008130081300813, "grad_norm": 2.219352461741057, "learning_rate": 1.5638897712112303e-05, "loss": 0.8076, "step": 1624 }, { "epoch": 3.3028455284552845, "grad_norm": 2.4892844099190343, "learning_rate": 1.5633459599308953e-05, "loss": 0.7033, "step": 1625 }, { "epoch": 3.3048780487804876, "grad_norm": 2.18601702380822, "learning_rate": 1.5628019044908252e-05, "loss": 0.6564, "step": 1626 }, { "epoch": 3.306910569105691, "grad_norm": 2.3826938196914655, "learning_rate": 1.5622576051268207e-05, "loss": 0.7087, "step": 1627 }, { "epoch": 3.3089430894308944, "grad_norm": 2.501068395423087, "learning_rate": 1.561713062074785e-05, "loss": 0.8124, "step": 1628 }, { "epoch": 3.3109756097560976, "grad_norm": 2.0629583083891947, "learning_rate": 1.5611682755707294e-05, "loss": 0.6993, "step": 1629 }, { "epoch": 3.313008130081301, "grad_norm": 1.9452939749910267, "learning_rate": 1.560623245850769e-05, "loss": 0.587, "step": 1630 }, { "epoch": 3.315040650406504, "grad_norm": 2.316721337616592, "learning_rate": 1.560077973151126e-05, "loss": 0.6361, "step": 1631 }, { "epoch": 3.317073170731707, "grad_norm": 2.0185206987879076, "learning_rate": 1.5595324577081265e-05, "loss": 0.7023, "step": 1632 }, { "epoch": 3.319105691056911, "grad_norm": 2.0186915245228807, "learning_rate": 1.558986699758202e-05, "loss": 0.8334, "step": 1633 }, { "epoch": 3.321138211382114, "grad_norm": 2.303065518921551, "learning_rate": 1.55844069953789e-05, "loss": 0.8791, "step": 1634 }, { "epoch": 3.323170731707317, "grad_norm": 2.4943431659013484, "learning_rate": 1.5578944572838324e-05, "loss": 0.842, "step": 1635 }, { "epoch": 3.3252032520325203, "grad_norm": 2.3475972216470185, "learning_rate": 1.5573479732327758e-05, "loss": 0.8366, "step": 1636 }, { "epoch": 3.3272357723577235, "grad_norm": 2.0720930162464914, "learning_rate": 1.5568012476215716e-05, "loss": 0.8918, "step": 1637 }, { "epoch": 3.3292682926829267, "grad_norm": 2.333827535681539, "learning_rate": 1.5562542806871765e-05, "loss": 0.827, "step": 1638 }, { "epoch": 3.33130081300813, "grad_norm": 2.150227172588249, "learning_rate": 1.555707072666651e-05, "loss": 0.8494, "step": 1639 }, { "epoch": 3.3333333333333335, "grad_norm": 2.3562395159232805, "learning_rate": 1.555159623797161e-05, "loss": 0.6991, "step": 1640 }, { "epoch": 3.3353658536585367, "grad_norm": 2.1832754825400276, "learning_rate": 1.5546119343159758e-05, "loss": 0.8301, "step": 1641 }, { "epoch": 3.33739837398374, "grad_norm": 2.4512422404891803, "learning_rate": 1.55406400446047e-05, "loss": 0.8545, "step": 1642 }, { "epoch": 3.339430894308943, "grad_norm": 2.0525938503084844, "learning_rate": 1.5535158344681215e-05, "loss": 0.6053, "step": 1643 }, { "epoch": 3.341463414634146, "grad_norm": 2.1717911904433915, "learning_rate": 1.552967424576512e-05, "loss": 0.7803, "step": 1644 }, { "epoch": 3.34349593495935, "grad_norm": 2.128954906241172, "learning_rate": 1.552418775023329e-05, "loss": 0.9302, "step": 1645 }, { "epoch": 3.345528455284553, "grad_norm": 2.314520862385919, "learning_rate": 1.5518698860463622e-05, "loss": 0.9062, "step": 1646 }, { "epoch": 3.347560975609756, "grad_norm": 2.4558864503502518, "learning_rate": 1.551320757883505e-05, "loss": 0.8707, "step": 1647 }, { "epoch": 3.3495934959349594, "grad_norm": 2.377138259930465, "learning_rate": 1.5507713907727557e-05, "loss": 0.8185, "step": 1648 }, { "epoch": 3.3516260162601625, "grad_norm": 2.2222799628341963, "learning_rate": 1.550221784952215e-05, "loss": 0.633, "step": 1649 }, { "epoch": 3.3536585365853657, "grad_norm": 2.0830111854780418, "learning_rate": 1.549671940660088e-05, "loss": 0.6593, "step": 1650 }, { "epoch": 3.355691056910569, "grad_norm": 2.172230854849265, "learning_rate": 1.549121858134682e-05, "loss": 0.7102, "step": 1651 }, { "epoch": 3.3577235772357725, "grad_norm": 2.177237281630638, "learning_rate": 1.5485715376144087e-05, "loss": 0.845, "step": 1652 }, { "epoch": 3.3597560975609757, "grad_norm": 2.084870915691989, "learning_rate": 1.5480209793377827e-05, "loss": 0.619, "step": 1653 }, { "epoch": 3.361788617886179, "grad_norm": 2.4328810292275103, "learning_rate": 1.5474701835434212e-05, "loss": 0.8579, "step": 1654 }, { "epoch": 3.363821138211382, "grad_norm": 2.1490119717803244, "learning_rate": 1.546919150470044e-05, "loss": 0.8233, "step": 1655 }, { "epoch": 3.3658536585365852, "grad_norm": 2.1154552440963434, "learning_rate": 1.5463678803564753e-05, "loss": 0.6795, "step": 1656 }, { "epoch": 3.3678861788617884, "grad_norm": 2.328996208658754, "learning_rate": 1.5458163734416403e-05, "loss": 0.7751, "step": 1657 }, { "epoch": 3.369918699186992, "grad_norm": 2.2017465868844703, "learning_rate": 1.545264629964568e-05, "loss": 0.7444, "step": 1658 }, { "epoch": 3.3719512195121952, "grad_norm": 2.486386949723798, "learning_rate": 1.5447126501643887e-05, "loss": 0.8152, "step": 1659 }, { "epoch": 3.3739837398373984, "grad_norm": 2.3224881606455643, "learning_rate": 1.5441604342803374e-05, "loss": 0.7339, "step": 1660 }, { "epoch": 3.3760162601626016, "grad_norm": 2.1056793177366675, "learning_rate": 1.5436079825517486e-05, "loss": 0.7854, "step": 1661 }, { "epoch": 3.3780487804878048, "grad_norm": 2.1162393251580793, "learning_rate": 1.5430552952180612e-05, "loss": 0.6754, "step": 1662 }, { "epoch": 3.380081300813008, "grad_norm": 1.8956705365987307, "learning_rate": 1.5425023725188144e-05, "loss": 0.6234, "step": 1663 }, { "epoch": 3.3821138211382116, "grad_norm": 2.00251666877645, "learning_rate": 1.5419492146936518e-05, "loss": 0.6941, "step": 1664 }, { "epoch": 3.3841463414634148, "grad_norm": 2.0145565934836416, "learning_rate": 1.5413958219823163e-05, "loss": 0.5742, "step": 1665 }, { "epoch": 3.386178861788618, "grad_norm": 2.189377943781308, "learning_rate": 1.540842194624654e-05, "loss": 0.7712, "step": 1666 }, { "epoch": 3.388211382113821, "grad_norm": 1.8973646922108174, "learning_rate": 1.5402883328606132e-05, "loss": 0.7374, "step": 1667 }, { "epoch": 3.3902439024390243, "grad_norm": 2.3165165333371056, "learning_rate": 1.5397342369302425e-05, "loss": 0.6606, "step": 1668 }, { "epoch": 3.3922764227642275, "grad_norm": 2.2513972098096455, "learning_rate": 1.5391799070736925e-05, "loss": 0.7561, "step": 1669 }, { "epoch": 3.394308943089431, "grad_norm": 1.906454624705497, "learning_rate": 1.5386253435312156e-05, "loss": 0.6247, "step": 1670 }, { "epoch": 3.3963414634146343, "grad_norm": 1.8480116033826013, "learning_rate": 1.538070546543165e-05, "loss": 0.5999, "step": 1671 }, { "epoch": 3.3983739837398375, "grad_norm": 2.0091996068249425, "learning_rate": 1.5375155163499953e-05, "loss": 0.6818, "step": 1672 }, { "epoch": 3.4004065040650406, "grad_norm": 2.382449559229901, "learning_rate": 1.5369602531922615e-05, "loss": 0.8757, "step": 1673 }, { "epoch": 3.402439024390244, "grad_norm": 2.2279724575281166, "learning_rate": 1.5364047573106215e-05, "loss": 0.6136, "step": 1674 }, { "epoch": 3.404471544715447, "grad_norm": 2.609690516425482, "learning_rate": 1.5358490289458315e-05, "loss": 0.6449, "step": 1675 }, { "epoch": 3.40650406504065, "grad_norm": 2.652147627447481, "learning_rate": 1.5352930683387502e-05, "loss": 0.6299, "step": 1676 }, { "epoch": 3.408536585365854, "grad_norm": 2.5054359524240897, "learning_rate": 1.5347368757303366e-05, "loss": 0.9364, "step": 1677 }, { "epoch": 3.410569105691057, "grad_norm": 2.3704796619659514, "learning_rate": 1.5341804513616497e-05, "loss": 0.7129, "step": 1678 }, { "epoch": 3.41260162601626, "grad_norm": 1.7982115394040779, "learning_rate": 1.533623795473849e-05, "loss": 0.6088, "step": 1679 }, { "epoch": 3.4146341463414633, "grad_norm": 2.370875491738983, "learning_rate": 1.5330669083081956e-05, "loss": 0.8323, "step": 1680 }, { "epoch": 3.4166666666666665, "grad_norm": 2.177961148267198, "learning_rate": 1.53250979010605e-05, "loss": 1.0088, "step": 1681 }, { "epoch": 3.41869918699187, "grad_norm": 2.45626422788114, "learning_rate": 1.531952441108872e-05, "loss": 0.835, "step": 1682 }, { "epoch": 3.4207317073170733, "grad_norm": 2.397091193497994, "learning_rate": 1.5313948615582225e-05, "loss": 0.7225, "step": 1683 }, { "epoch": 3.4227642276422765, "grad_norm": 2.3409836037789806, "learning_rate": 1.5308370516957617e-05, "loss": 0.5725, "step": 1684 }, { "epoch": 3.4247967479674797, "grad_norm": 2.246687762090141, "learning_rate": 1.53027901176325e-05, "loss": 0.5938, "step": 1685 }, { "epoch": 3.426829268292683, "grad_norm": 2.1640171195675313, "learning_rate": 1.5297207420025484e-05, "loss": 0.8625, "step": 1686 }, { "epoch": 3.428861788617886, "grad_norm": 2.5898628653896503, "learning_rate": 1.5291622426556158e-05, "loss": 0.7517, "step": 1687 }, { "epoch": 3.430894308943089, "grad_norm": 1.8945188911843114, "learning_rate": 1.528603513964511e-05, "loss": 0.8206, "step": 1688 }, { "epoch": 3.432926829268293, "grad_norm": 2.402531034058431, "learning_rate": 1.528044556171393e-05, "loss": 0.7873, "step": 1689 }, { "epoch": 3.434959349593496, "grad_norm": 2.181933249847049, "learning_rate": 1.52748536951852e-05, "loss": 0.9133, "step": 1690 }, { "epoch": 3.436991869918699, "grad_norm": 2.086420642038962, "learning_rate": 1.5269259542482486e-05, "loss": 0.708, "step": 1691 }, { "epoch": 3.4390243902439024, "grad_norm": 2.1369644052706196, "learning_rate": 1.5263663106030347e-05, "loss": 0.7143, "step": 1692 }, { "epoch": 3.4410569105691056, "grad_norm": 2.2667834435610463, "learning_rate": 1.5258064388254342e-05, "loss": 0.8638, "step": 1693 }, { "epoch": 3.443089430894309, "grad_norm": 2.293288526692144, "learning_rate": 1.5252463391581004e-05, "loss": 0.7317, "step": 1694 }, { "epoch": 3.4451219512195124, "grad_norm": 2.269461086087901, "learning_rate": 1.5246860118437863e-05, "loss": 0.8373, "step": 1695 }, { "epoch": 3.4471544715447155, "grad_norm": 2.172209050828749, "learning_rate": 1.5241254571253433e-05, "loss": 0.7113, "step": 1696 }, { "epoch": 3.4491869918699187, "grad_norm": 1.8643274554020834, "learning_rate": 1.5235646752457214e-05, "loss": 0.646, "step": 1697 }, { "epoch": 3.451219512195122, "grad_norm": 1.7691749000659098, "learning_rate": 1.523003666447969e-05, "loss": 0.5408, "step": 1698 }, { "epoch": 3.453252032520325, "grad_norm": 2.764085977593175, "learning_rate": 1.5224424309752327e-05, "loss": 0.9322, "step": 1699 }, { "epoch": 3.4552845528455283, "grad_norm": 2.200248903487396, "learning_rate": 1.5218809690707583e-05, "loss": 0.8345, "step": 1700 }, { "epoch": 3.457317073170732, "grad_norm": 3.1995569038303153, "learning_rate": 1.5213192809778878e-05, "loss": 0.8277, "step": 1701 }, { "epoch": 3.459349593495935, "grad_norm": 2.4425519546794248, "learning_rate": 1.5207573669400636e-05, "loss": 0.7624, "step": 1702 }, { "epoch": 3.4613821138211383, "grad_norm": 2.2983722233658748, "learning_rate": 1.5201952272008236e-05, "loss": 0.7498, "step": 1703 }, { "epoch": 3.4634146341463414, "grad_norm": 2.320630399294237, "learning_rate": 1.5196328620038059e-05, "loss": 0.8424, "step": 1704 }, { "epoch": 3.4654471544715446, "grad_norm": 2.200878218752855, "learning_rate": 1.5190702715927443e-05, "loss": 0.7542, "step": 1705 }, { "epoch": 3.467479674796748, "grad_norm": 2.206868679011799, "learning_rate": 1.5185074562114715e-05, "loss": 0.6138, "step": 1706 }, { "epoch": 3.4695121951219514, "grad_norm": 1.9471071523977526, "learning_rate": 1.5179444161039172e-05, "loss": 0.6865, "step": 1707 }, { "epoch": 3.4715447154471546, "grad_norm": 2.424686241067536, "learning_rate": 1.5173811515141083e-05, "loss": 0.7131, "step": 1708 }, { "epoch": 3.4735772357723578, "grad_norm": 2.3023910464597206, "learning_rate": 1.51681766268617e-05, "loss": 0.8361, "step": 1709 }, { "epoch": 3.475609756097561, "grad_norm": 2.1907244360971707, "learning_rate": 1.5162539498643237e-05, "loss": 0.6992, "step": 1710 }, { "epoch": 3.477642276422764, "grad_norm": 2.376677571295473, "learning_rate": 1.5156900132928874e-05, "loss": 0.7543, "step": 1711 }, { "epoch": 3.4796747967479673, "grad_norm": 2.133142138735676, "learning_rate": 1.5151258532162771e-05, "loss": 0.8636, "step": 1712 }, { "epoch": 3.4817073170731705, "grad_norm": 2.783051341389404, "learning_rate": 1.5145614698790062e-05, "loss": 0.9374, "step": 1713 }, { "epoch": 3.483739837398374, "grad_norm": 2.198593881822009, "learning_rate": 1.5139968635256833e-05, "loss": 0.6561, "step": 1714 }, { "epoch": 3.4857723577235773, "grad_norm": 2.2601482974789575, "learning_rate": 1.5134320344010144e-05, "loss": 0.6729, "step": 1715 }, { "epoch": 3.4878048780487805, "grad_norm": 2.4816466357659728, "learning_rate": 1.5128669827498024e-05, "loss": 0.9606, "step": 1716 }, { "epoch": 3.4898373983739837, "grad_norm": 2.6398884089375594, "learning_rate": 1.5123017088169457e-05, "loss": 0.799, "step": 1717 }, { "epoch": 3.491869918699187, "grad_norm": 2.6181345352281773, "learning_rate": 1.5117362128474406e-05, "loss": 0.7626, "step": 1718 }, { "epoch": 3.4939024390243905, "grad_norm": 2.6035519155784934, "learning_rate": 1.5111704950863776e-05, "loss": 0.8132, "step": 1719 }, { "epoch": 3.4959349593495936, "grad_norm": 2.235540172273131, "learning_rate": 1.5106045557789453e-05, "loss": 0.7606, "step": 1720 }, { "epoch": 3.497967479674797, "grad_norm": 2.2092960760275546, "learning_rate": 1.510038395170427e-05, "loss": 0.6867, "step": 1721 }, { "epoch": 3.5, "grad_norm": 2.715269374921054, "learning_rate": 1.5094720135062022e-05, "loss": 0.8975, "step": 1722 }, { "epoch": 3.502032520325203, "grad_norm": 2.521341153151502, "learning_rate": 1.5089054110317464e-05, "loss": 0.9533, "step": 1723 }, { "epoch": 3.5040650406504064, "grad_norm": 2.144944601383175, "learning_rate": 1.5083385879926309e-05, "loss": 0.7888, "step": 1724 }, { "epoch": 3.5060975609756095, "grad_norm": 2.6258119326676463, "learning_rate": 1.5077715446345227e-05, "loss": 0.8267, "step": 1725 }, { "epoch": 3.508130081300813, "grad_norm": 2.2828317444796147, "learning_rate": 1.5072042812031835e-05, "loss": 0.6542, "step": 1726 }, { "epoch": 3.5101626016260163, "grad_norm": 2.5590393774085083, "learning_rate": 1.5066367979444709e-05, "loss": 0.7517, "step": 1727 }, { "epoch": 3.5121951219512195, "grad_norm": 2.117215414672995, "learning_rate": 1.5060690951043385e-05, "loss": 0.7488, "step": 1728 }, { "epoch": 3.5142276422764227, "grad_norm": 2.424675833253423, "learning_rate": 1.5055011729288331e-05, "loss": 0.8592, "step": 1729 }, { "epoch": 3.516260162601626, "grad_norm": 2.1717064496082785, "learning_rate": 1.5049330316640989e-05, "loss": 0.7398, "step": 1730 }, { "epoch": 3.5182926829268295, "grad_norm": 2.4873753052880123, "learning_rate": 1.504364671556374e-05, "loss": 0.8296, "step": 1731 }, { "epoch": 3.5203252032520327, "grad_norm": 2.1866274703587347, "learning_rate": 1.5037960928519902e-05, "loss": 0.6548, "step": 1732 }, { "epoch": 3.522357723577236, "grad_norm": 2.426492522506623, "learning_rate": 1.5032272957973764e-05, "loss": 0.6965, "step": 1733 }, { "epoch": 3.524390243902439, "grad_norm": 2.436552487324315, "learning_rate": 1.5026582806390541e-05, "loss": 0.6583, "step": 1734 }, { "epoch": 3.5264227642276422, "grad_norm": 2.5472186714987326, "learning_rate": 1.5020890476236405e-05, "loss": 0.7933, "step": 1735 }, { "epoch": 3.5284552845528454, "grad_norm": 2.3326012757739583, "learning_rate": 1.501519596997847e-05, "loss": 0.8639, "step": 1736 }, { "epoch": 3.5304878048780486, "grad_norm": 2.6075626073909506, "learning_rate": 1.5009499290084786e-05, "loss": 0.8119, "step": 1737 }, { "epoch": 3.5325203252032518, "grad_norm": 2.6857488555114952, "learning_rate": 1.5003800439024355e-05, "loss": 0.6526, "step": 1738 }, { "epoch": 3.5345528455284554, "grad_norm": 2.814667695866756, "learning_rate": 1.4998099419267111e-05, "loss": 0.7682, "step": 1739 }, { "epoch": 3.5365853658536586, "grad_norm": 2.249527590379016, "learning_rate": 1.499239623328394e-05, "loss": 0.5127, "step": 1740 }, { "epoch": 3.5386178861788617, "grad_norm": 2.6981258394032994, "learning_rate": 1.4986690883546653e-05, "loss": 0.7712, "step": 1741 }, { "epoch": 3.540650406504065, "grad_norm": 2.049075498055457, "learning_rate": 1.498098337252801e-05, "loss": 0.8141, "step": 1742 }, { "epoch": 3.5426829268292686, "grad_norm": 2.6646435503454433, "learning_rate": 1.4975273702701697e-05, "loss": 0.8896, "step": 1743 }, { "epoch": 3.5447154471544717, "grad_norm": 2.3161303260026185, "learning_rate": 1.4969561876542348e-05, "loss": 0.7735, "step": 1744 }, { "epoch": 3.546747967479675, "grad_norm": 2.4119256475919375, "learning_rate": 1.4963847896525524e-05, "loss": 0.7879, "step": 1745 }, { "epoch": 3.548780487804878, "grad_norm": 2.286844111711807, "learning_rate": 1.4958131765127712e-05, "loss": 0.784, "step": 1746 }, { "epoch": 3.5508130081300813, "grad_norm": 2.4201570256320415, "learning_rate": 1.4952413484826356e-05, "loss": 0.6819, "step": 1747 }, { "epoch": 3.5528455284552845, "grad_norm": 2.156654292840832, "learning_rate": 1.4946693058099802e-05, "loss": 0.8258, "step": 1748 }, { "epoch": 3.5548780487804876, "grad_norm": 2.331858401590426, "learning_rate": 1.4940970487427345e-05, "loss": 0.9134, "step": 1749 }, { "epoch": 3.556910569105691, "grad_norm": 2.333570878160564, "learning_rate": 1.4935245775289203e-05, "loss": 0.7213, "step": 1750 }, { "epoch": 3.5589430894308944, "grad_norm": 2.245634121777129, "learning_rate": 1.4929518924166521e-05, "loss": 0.7783, "step": 1751 }, { "epoch": 3.5609756097560976, "grad_norm": 2.284754015993947, "learning_rate": 1.4923789936541378e-05, "loss": 0.6694, "step": 1752 }, { "epoch": 3.563008130081301, "grad_norm": 2.247129487791626, "learning_rate": 1.4918058814896772e-05, "loss": 0.794, "step": 1753 }, { "epoch": 3.565040650406504, "grad_norm": 2.2977164297618313, "learning_rate": 1.4912325561716629e-05, "loss": 0.6541, "step": 1754 }, { "epoch": 3.567073170731707, "grad_norm": 2.425456457216071, "learning_rate": 1.4906590179485795e-05, "loss": 0.6871, "step": 1755 }, { "epoch": 3.569105691056911, "grad_norm": 2.3721552992010735, "learning_rate": 1.4900852670690044e-05, "loss": 0.7728, "step": 1756 }, { "epoch": 3.571138211382114, "grad_norm": 2.7527881913888907, "learning_rate": 1.4895113037816067e-05, "loss": 0.7221, "step": 1757 }, { "epoch": 3.573170731707317, "grad_norm": 2.4555979384393796, "learning_rate": 1.4889371283351482e-05, "loss": 0.7981, "step": 1758 }, { "epoch": 3.5752032520325203, "grad_norm": 2.2164584883165364, "learning_rate": 1.4883627409784819e-05, "loss": 0.7125, "step": 1759 }, { "epoch": 3.5772357723577235, "grad_norm": 2.504197997959473, "learning_rate": 1.487788141960553e-05, "loss": 0.8813, "step": 1760 }, { "epoch": 3.5792682926829267, "grad_norm": 2.315713296194338, "learning_rate": 1.4872133315303984e-05, "loss": 0.7688, "step": 1761 }, { "epoch": 3.58130081300813, "grad_norm": 2.5290731939822484, "learning_rate": 1.4866383099371468e-05, "loss": 0.797, "step": 1762 }, { "epoch": 3.5833333333333335, "grad_norm": 2.116110684497016, "learning_rate": 1.4860630774300186e-05, "loss": 0.6285, "step": 1763 }, { "epoch": 3.5853658536585367, "grad_norm": 2.5872639712813106, "learning_rate": 1.4854876342583246e-05, "loss": 0.8603, "step": 1764 }, { "epoch": 3.58739837398374, "grad_norm": 1.994128833896292, "learning_rate": 1.484911980671468e-05, "loss": 0.6286, "step": 1765 }, { "epoch": 3.589430894308943, "grad_norm": 2.3049809757050617, "learning_rate": 1.4843361169189425e-05, "loss": 0.7161, "step": 1766 }, { "epoch": 3.591463414634146, "grad_norm": 2.1296861697387452, "learning_rate": 1.4837600432503332e-05, "loss": 0.6037, "step": 1767 }, { "epoch": 3.59349593495935, "grad_norm": 2.5748759091161637, "learning_rate": 1.4831837599153165e-05, "loss": 0.7055, "step": 1768 }, { "epoch": 3.595528455284553, "grad_norm": 2.70832066067895, "learning_rate": 1.4826072671636592e-05, "loss": 0.8185, "step": 1769 }, { "epoch": 3.597560975609756, "grad_norm": 2.486850679538958, "learning_rate": 1.482030565245219e-05, "loss": 0.8361, "step": 1770 }, { "epoch": 3.5995934959349594, "grad_norm": 2.310244868720956, "learning_rate": 1.4814536544099439e-05, "loss": 0.762, "step": 1771 }, { "epoch": 3.6016260162601625, "grad_norm": 2.406762299554278, "learning_rate": 1.4808765349078729e-05, "loss": 0.7272, "step": 1772 }, { "epoch": 3.6036585365853657, "grad_norm": 2.439252470778541, "learning_rate": 1.4802992069891355e-05, "loss": 0.9479, "step": 1773 }, { "epoch": 3.605691056910569, "grad_norm": 2.1665002818351273, "learning_rate": 1.4797216709039515e-05, "loss": 0.646, "step": 1774 }, { "epoch": 3.607723577235772, "grad_norm": 2.715041569112013, "learning_rate": 1.4791439269026304e-05, "loss": 0.8866, "step": 1775 }, { "epoch": 3.6097560975609757, "grad_norm": 2.05226479931043, "learning_rate": 1.4785659752355724e-05, "loss": 0.7893, "step": 1776 }, { "epoch": 3.611788617886179, "grad_norm": 2.442993739288783, "learning_rate": 1.4779878161532675e-05, "loss": 0.7576, "step": 1777 }, { "epoch": 3.613821138211382, "grad_norm": 2.8434946216352066, "learning_rate": 1.4774094499062954e-05, "loss": 0.7898, "step": 1778 }, { "epoch": 3.6158536585365852, "grad_norm": 2.5546883278075683, "learning_rate": 1.476830876745326e-05, "loss": 0.6853, "step": 1779 }, { "epoch": 3.617886178861789, "grad_norm": 2.751841763863824, "learning_rate": 1.4762520969211186e-05, "loss": 0.6846, "step": 1780 }, { "epoch": 3.619918699186992, "grad_norm": 2.614302930383474, "learning_rate": 1.475673110684522e-05, "loss": 0.6873, "step": 1781 }, { "epoch": 3.6219512195121952, "grad_norm": 2.428819129660783, "learning_rate": 1.475093918286475e-05, "loss": 0.7591, "step": 1782 }, { "epoch": 3.6239837398373984, "grad_norm": 2.8251434569762734, "learning_rate": 1.4745145199780046e-05, "loss": 0.6428, "step": 1783 }, { "epoch": 3.6260162601626016, "grad_norm": 2.76900049236495, "learning_rate": 1.4739349160102285e-05, "loss": 0.7587, "step": 1784 }, { "epoch": 3.6280487804878048, "grad_norm": 2.6902054732348657, "learning_rate": 1.4733551066343525e-05, "loss": 0.8427, "step": 1785 }, { "epoch": 3.630081300813008, "grad_norm": 2.7801358875120883, "learning_rate": 1.4727750921016716e-05, "loss": 0.9173, "step": 1786 }, { "epoch": 3.632113821138211, "grad_norm": 2.36805829258717, "learning_rate": 1.4721948726635701e-05, "loss": 0.7762, "step": 1787 }, { "epoch": 3.6341463414634148, "grad_norm": 2.2493938192072838, "learning_rate": 1.4716144485715209e-05, "loss": 0.7014, "step": 1788 }, { "epoch": 3.636178861788618, "grad_norm": 3.1139492062941767, "learning_rate": 1.4710338200770853e-05, "loss": 0.9099, "step": 1789 }, { "epoch": 3.638211382113821, "grad_norm": 2.318539866578785, "learning_rate": 1.4704529874319135e-05, "loss": 0.718, "step": 1790 }, { "epoch": 3.6402439024390243, "grad_norm": 2.29688119482607, "learning_rate": 1.4698719508877448e-05, "loss": 0.7109, "step": 1791 }, { "epoch": 3.642276422764228, "grad_norm": 2.3824081603315257, "learning_rate": 1.4692907106964051e-05, "loss": 0.8201, "step": 1792 }, { "epoch": 3.644308943089431, "grad_norm": 2.5278038170479094, "learning_rate": 1.4687092671098104e-05, "loss": 0.8965, "step": 1793 }, { "epoch": 3.6463414634146343, "grad_norm": 2.562027634801068, "learning_rate": 1.468127620379964e-05, "loss": 0.7687, "step": 1794 }, { "epoch": 3.6483739837398375, "grad_norm": 2.2131666559884855, "learning_rate": 1.4675457707589572e-05, "loss": 0.652, "step": 1795 }, { "epoch": 3.6504065040650406, "grad_norm": 2.6716475485894744, "learning_rate": 1.4669637184989696e-05, "loss": 0.8936, "step": 1796 }, { "epoch": 3.652439024390244, "grad_norm": 2.2746921366429667, "learning_rate": 1.4663814638522684e-05, "loss": 0.6898, "step": 1797 }, { "epoch": 3.654471544715447, "grad_norm": 2.471097384398505, "learning_rate": 1.4657990070712088e-05, "loss": 0.6465, "step": 1798 }, { "epoch": 3.65650406504065, "grad_norm": 2.550204605915045, "learning_rate": 1.4652163484082325e-05, "loss": 0.7311, "step": 1799 }, { "epoch": 3.658536585365854, "grad_norm": 2.422781200582373, "learning_rate": 1.4646334881158704e-05, "loss": 0.6802, "step": 1800 }, { "epoch": 3.660569105691057, "grad_norm": 2.278361412883645, "learning_rate": 1.4640504264467397e-05, "loss": 0.7088, "step": 1801 }, { "epoch": 3.66260162601626, "grad_norm": 2.334420382834965, "learning_rate": 1.463467163653545e-05, "loss": 0.6594, "step": 1802 }, { "epoch": 3.6646341463414633, "grad_norm": 2.563602664412063, "learning_rate": 1.4628836999890783e-05, "loss": 0.7297, "step": 1803 }, { "epoch": 3.6666666666666665, "grad_norm": 2.7349788640045487, "learning_rate": 1.4623000357062184e-05, "loss": 0.7915, "step": 1804 }, { "epoch": 3.66869918699187, "grad_norm": 2.4572010664375243, "learning_rate": 1.4617161710579315e-05, "loss": 0.8307, "step": 1805 }, { "epoch": 3.6707317073170733, "grad_norm": 2.4332047614036982, "learning_rate": 1.46113210629727e-05, "loss": 0.7754, "step": 1806 }, { "epoch": 3.6727642276422765, "grad_norm": 2.3093640794932146, "learning_rate": 1.4605478416773733e-05, "loss": 0.7722, "step": 1807 }, { "epoch": 3.6747967479674797, "grad_norm": 2.261769700504792, "learning_rate": 1.459963377451468e-05, "loss": 0.7117, "step": 1808 }, { "epoch": 3.676829268292683, "grad_norm": 2.4051299467193266, "learning_rate": 1.4593787138728664e-05, "loss": 0.7615, "step": 1809 }, { "epoch": 3.678861788617886, "grad_norm": 2.2201002174789832, "learning_rate": 1.4587938511949673e-05, "loss": 0.7609, "step": 1810 }, { "epoch": 3.680894308943089, "grad_norm": 2.635751806109238, "learning_rate": 1.4582087896712563e-05, "loss": 0.7253, "step": 1811 }, { "epoch": 3.682926829268293, "grad_norm": 2.6396641978156907, "learning_rate": 1.457623529555305e-05, "loss": 0.6807, "step": 1812 }, { "epoch": 3.684959349593496, "grad_norm": 2.300333388649156, "learning_rate": 1.4570380711007705e-05, "loss": 0.6022, "step": 1813 }, { "epoch": 3.686991869918699, "grad_norm": 2.216035611014914, "learning_rate": 1.4564524145613962e-05, "loss": 0.7795, "step": 1814 }, { "epoch": 3.6890243902439024, "grad_norm": 2.3051832588478796, "learning_rate": 1.4558665601910124e-05, "loss": 0.665, "step": 1815 }, { "epoch": 3.6910569105691056, "grad_norm": 2.5748968320115524, "learning_rate": 1.4552805082435333e-05, "loss": 0.6807, "step": 1816 }, { "epoch": 3.693089430894309, "grad_norm": 2.516454623273677, "learning_rate": 1.4546942589729601e-05, "loss": 0.7347, "step": 1817 }, { "epoch": 3.6951219512195124, "grad_norm": 2.246067156663959, "learning_rate": 1.4541078126333785e-05, "loss": 0.7572, "step": 1818 }, { "epoch": 3.6971544715447155, "grad_norm": 2.4909127854903095, "learning_rate": 1.453521169478961e-05, "loss": 0.7859, "step": 1819 }, { "epoch": 3.6991869918699187, "grad_norm": 2.6202166602965775, "learning_rate": 1.4529343297639638e-05, "loss": 0.7215, "step": 1820 }, { "epoch": 3.701219512195122, "grad_norm": 2.6899892138161356, "learning_rate": 1.4523472937427295e-05, "loss": 0.8692, "step": 1821 }, { "epoch": 3.703252032520325, "grad_norm": 2.507737710748079, "learning_rate": 1.4517600616696852e-05, "loss": 0.6968, "step": 1822 }, { "epoch": 3.7052845528455283, "grad_norm": 2.296827483820878, "learning_rate": 1.4511726337993432e-05, "loss": 0.8793, "step": 1823 }, { "epoch": 3.7073170731707314, "grad_norm": 2.5400904771211152, "learning_rate": 1.4505850103863007e-05, "loss": 0.6979, "step": 1824 }, { "epoch": 3.709349593495935, "grad_norm": 2.1748690799644583, "learning_rate": 1.4499971916852388e-05, "loss": 0.7486, "step": 1825 }, { "epoch": 3.7113821138211383, "grad_norm": 2.1947934757720193, "learning_rate": 1.4494091779509249e-05, "loss": 0.6486, "step": 1826 }, { "epoch": 3.7134146341463414, "grad_norm": 2.3935791906016006, "learning_rate": 1.4488209694382093e-05, "loss": 0.7145, "step": 1827 }, { "epoch": 3.7154471544715446, "grad_norm": 2.8344252006694766, "learning_rate": 1.448232566402028e-05, "loss": 0.8467, "step": 1828 }, { "epoch": 3.7174796747967482, "grad_norm": 2.809522115432216, "learning_rate": 1.4476439690974004e-05, "loss": 0.6822, "step": 1829 }, { "epoch": 3.7195121951219514, "grad_norm": 2.582130140154344, "learning_rate": 1.4470551777794303e-05, "loss": 0.6092, "step": 1830 }, { "epoch": 3.7215447154471546, "grad_norm": 2.301808573579162, "learning_rate": 1.4464661927033062e-05, "loss": 0.7635, "step": 1831 }, { "epoch": 3.7235772357723578, "grad_norm": 2.9462707826953647, "learning_rate": 1.4458770141242992e-05, "loss": 0.9468, "step": 1832 }, { "epoch": 3.725609756097561, "grad_norm": 2.093886009088198, "learning_rate": 1.445287642297766e-05, "loss": 0.7809, "step": 1833 }, { "epoch": 3.727642276422764, "grad_norm": 2.9260414500345595, "learning_rate": 1.4446980774791456e-05, "loss": 0.8244, "step": 1834 }, { "epoch": 3.7296747967479673, "grad_norm": 2.422839952795823, "learning_rate": 1.444108319923962e-05, "loss": 0.9659, "step": 1835 }, { "epoch": 3.7317073170731705, "grad_norm": 2.5053479364391786, "learning_rate": 1.4435183698878212e-05, "loss": 0.7152, "step": 1836 }, { "epoch": 3.733739837398374, "grad_norm": 2.245233664933188, "learning_rate": 1.4429282276264136e-05, "loss": 0.7045, "step": 1837 }, { "epoch": 3.7357723577235773, "grad_norm": 2.2940608175090174, "learning_rate": 1.4423378933955133e-05, "loss": 0.8606, "step": 1838 }, { "epoch": 3.7378048780487805, "grad_norm": 2.4721594166564875, "learning_rate": 1.4417473674509761e-05, "loss": 0.8528, "step": 1839 }, { "epoch": 3.7398373983739837, "grad_norm": 2.07526781733691, "learning_rate": 1.4411566500487425e-05, "loss": 0.6935, "step": 1840 }, { "epoch": 3.741869918699187, "grad_norm": 2.5910267178353115, "learning_rate": 1.4405657414448348e-05, "loss": 0.9298, "step": 1841 }, { "epoch": 3.7439024390243905, "grad_norm": 2.562516098429953, "learning_rate": 1.4399746418953591e-05, "loss": 0.6771, "step": 1842 }, { "epoch": 3.7459349593495936, "grad_norm": 2.6075365291550026, "learning_rate": 1.4393833516565036e-05, "loss": 0.7483, "step": 1843 }, { "epoch": 3.747967479674797, "grad_norm": 2.1051194201555705, "learning_rate": 1.4387918709845395e-05, "loss": 0.7358, "step": 1844 }, { "epoch": 3.75, "grad_norm": 2.1990916248179686, "learning_rate": 1.4382002001358201e-05, "loss": 0.8396, "step": 1845 }, { "epoch": 3.752032520325203, "grad_norm": 2.3580608784264814, "learning_rate": 1.4376083393667818e-05, "loss": 0.8762, "step": 1846 }, { "epoch": 3.7540650406504064, "grad_norm": 2.7499837745155227, "learning_rate": 1.4370162889339425e-05, "loss": 0.8387, "step": 1847 }, { "epoch": 3.7560975609756095, "grad_norm": 2.4968495057892577, "learning_rate": 1.4364240490939032e-05, "loss": 0.9099, "step": 1848 }, { "epoch": 3.758130081300813, "grad_norm": 1.8755241656618975, "learning_rate": 1.4358316201033461e-05, "loss": 0.6756, "step": 1849 }, { "epoch": 3.7601626016260163, "grad_norm": 2.38792778478541, "learning_rate": 1.4352390022190363e-05, "loss": 0.7481, "step": 1850 }, { "epoch": 3.7621951219512195, "grad_norm": 2.1951324648986206, "learning_rate": 1.4346461956978199e-05, "loss": 0.7198, "step": 1851 }, { "epoch": 3.7642276422764227, "grad_norm": 2.3907920174766355, "learning_rate": 1.4340532007966252e-05, "loss": 0.6943, "step": 1852 }, { "epoch": 3.766260162601626, "grad_norm": 2.647047782043043, "learning_rate": 1.433460017772462e-05, "loss": 0.7424, "step": 1853 }, { "epoch": 3.7682926829268295, "grad_norm": 2.2833500260363904, "learning_rate": 1.4328666468824218e-05, "loss": 0.789, "step": 1854 }, { "epoch": 3.7703252032520327, "grad_norm": 2.435015919030787, "learning_rate": 1.4322730883836778e-05, "loss": 0.7648, "step": 1855 }, { "epoch": 3.772357723577236, "grad_norm": 2.7428967246188543, "learning_rate": 1.4316793425334836e-05, "loss": 0.8076, "step": 1856 }, { "epoch": 3.774390243902439, "grad_norm": 2.3484720528327565, "learning_rate": 1.431085409589175e-05, "loss": 0.7781, "step": 1857 }, { "epoch": 3.7764227642276422, "grad_norm": 2.4896317588317243, "learning_rate": 1.4304912898081677e-05, "loss": 0.7475, "step": 1858 }, { "epoch": 3.7784552845528454, "grad_norm": 2.460143364046719, "learning_rate": 1.4298969834479599e-05, "loss": 0.8125, "step": 1859 }, { "epoch": 3.7804878048780486, "grad_norm": 2.612074977242424, "learning_rate": 1.4293024907661295e-05, "loss": 0.894, "step": 1860 }, { "epoch": 3.7825203252032518, "grad_norm": 2.4661965053149104, "learning_rate": 1.4287078120203358e-05, "loss": 0.7966, "step": 1861 }, { "epoch": 3.7845528455284554, "grad_norm": 2.2511969543433583, "learning_rate": 1.4281129474683182e-05, "loss": 0.7569, "step": 1862 }, { "epoch": 3.7865853658536586, "grad_norm": 2.751292255856745, "learning_rate": 1.4275178973678969e-05, "loss": 0.9537, "step": 1863 }, { "epoch": 3.7886178861788617, "grad_norm": 2.508709732670518, "learning_rate": 1.4269226619769727e-05, "loss": 0.7267, "step": 1864 }, { "epoch": 3.790650406504065, "grad_norm": 2.473584158041306, "learning_rate": 1.4263272415535263e-05, "loss": 0.8876, "step": 1865 }, { "epoch": 3.7926829268292686, "grad_norm": 2.0506842762220816, "learning_rate": 1.4257316363556192e-05, "loss": 0.5663, "step": 1866 }, { "epoch": 3.7947154471544717, "grad_norm": 2.195658022000722, "learning_rate": 1.4251358466413924e-05, "loss": 0.7275, "step": 1867 }, { "epoch": 3.796747967479675, "grad_norm": 2.936644566845942, "learning_rate": 1.424539872669067e-05, "loss": 0.7185, "step": 1868 }, { "epoch": 3.798780487804878, "grad_norm": 2.5715365820442915, "learning_rate": 1.4239437146969443e-05, "loss": 0.7273, "step": 1869 }, { "epoch": 3.8008130081300813, "grad_norm": 2.7904819977425412, "learning_rate": 1.4233473729834046e-05, "loss": 0.7388, "step": 1870 }, { "epoch": 3.8028455284552845, "grad_norm": 2.928111105251171, "learning_rate": 1.422750847786909e-05, "loss": 0.78, "step": 1871 }, { "epoch": 3.8048780487804876, "grad_norm": 2.5501358803894894, "learning_rate": 1.4221541393659966e-05, "loss": 0.8601, "step": 1872 }, { "epoch": 3.806910569105691, "grad_norm": 2.599807371804817, "learning_rate": 1.4215572479792877e-05, "loss": 0.8466, "step": 1873 }, { "epoch": 3.8089430894308944, "grad_norm": 2.6304983813964413, "learning_rate": 1.4209601738854802e-05, "loss": 0.8735, "step": 1874 }, { "epoch": 3.8109756097560976, "grad_norm": 3.1478403649752287, "learning_rate": 1.420362917343352e-05, "loss": 0.8017, "step": 1875 }, { "epoch": 3.813008130081301, "grad_norm": 3.33940243405905, "learning_rate": 1.4197654786117604e-05, "loss": 0.9354, "step": 1876 }, { "epoch": 3.815040650406504, "grad_norm": 2.715985362521496, "learning_rate": 1.4191678579496413e-05, "loss": 0.7184, "step": 1877 }, { "epoch": 3.817073170731707, "grad_norm": 2.736911696942446, "learning_rate": 1.4185700556160094e-05, "loss": 0.8107, "step": 1878 }, { "epoch": 3.819105691056911, "grad_norm": 2.4494301244040777, "learning_rate": 1.4179720718699577e-05, "loss": 0.5888, "step": 1879 }, { "epoch": 3.821138211382114, "grad_norm": 2.4136080998004465, "learning_rate": 1.4173739069706586e-05, "loss": 0.6583, "step": 1880 }, { "epoch": 3.823170731707317, "grad_norm": 2.6830935263536455, "learning_rate": 1.4167755611773628e-05, "loss": 0.7252, "step": 1881 }, { "epoch": 3.8252032520325203, "grad_norm": 2.4031792354285235, "learning_rate": 1.4161770347493991e-05, "loss": 0.7299, "step": 1882 }, { "epoch": 3.8272357723577235, "grad_norm": 2.6262170242003227, "learning_rate": 1.415578327946175e-05, "loss": 0.65, "step": 1883 }, { "epoch": 3.8292682926829267, "grad_norm": 2.5594085021248745, "learning_rate": 1.414979441027176e-05, "loss": 0.6477, "step": 1884 }, { "epoch": 3.83130081300813, "grad_norm": 2.737286242699734, "learning_rate": 1.4143803742519657e-05, "loss": 0.7738, "step": 1885 }, { "epoch": 3.8333333333333335, "grad_norm": 2.8731692028634805, "learning_rate": 1.4137811278801849e-05, "loss": 0.816, "step": 1886 }, { "epoch": 3.8353658536585367, "grad_norm": 2.580954170190974, "learning_rate": 1.4131817021715538e-05, "loss": 0.6095, "step": 1887 }, { "epoch": 3.83739837398374, "grad_norm": 2.8858952595590543, "learning_rate": 1.4125820973858693e-05, "loss": 0.7184, "step": 1888 }, { "epoch": 3.839430894308943, "grad_norm": 2.5076065296939904, "learning_rate": 1.4119823137830058e-05, "loss": 0.7023, "step": 1889 }, { "epoch": 3.841463414634146, "grad_norm": 2.3637011340726737, "learning_rate": 1.4113823516229154e-05, "loss": 0.6486, "step": 1890 }, { "epoch": 3.84349593495935, "grad_norm": 3.096546558144839, "learning_rate": 1.410782211165628e-05, "loss": 0.7897, "step": 1891 }, { "epoch": 3.845528455284553, "grad_norm": 2.6110460672190823, "learning_rate": 1.41018189267125e-05, "loss": 0.9206, "step": 1892 }, { "epoch": 3.847560975609756, "grad_norm": 2.7171150908469817, "learning_rate": 1.4095813963999656e-05, "loss": 0.8465, "step": 1893 }, { "epoch": 3.8495934959349594, "grad_norm": 2.894011863426682, "learning_rate": 1.4089807226120362e-05, "loss": 0.7843, "step": 1894 }, { "epoch": 3.8516260162601625, "grad_norm": 2.5978564825426074, "learning_rate": 1.4083798715677995e-05, "loss": 0.745, "step": 1895 }, { "epoch": 3.8536585365853657, "grad_norm": 2.584501245534706, "learning_rate": 1.4077788435276701e-05, "loss": 0.7047, "step": 1896 }, { "epoch": 3.855691056910569, "grad_norm": 2.4398978388576213, "learning_rate": 1.40717763875214e-05, "loss": 0.6462, "step": 1897 }, { "epoch": 3.857723577235772, "grad_norm": 2.4375817626955745, "learning_rate": 1.4065762575017765e-05, "loss": 0.6204, "step": 1898 }, { "epoch": 3.8597560975609757, "grad_norm": 2.6120827527111916, "learning_rate": 1.4059747000372253e-05, "loss": 0.6901, "step": 1899 }, { "epoch": 3.861788617886179, "grad_norm": 2.3101747474623364, "learning_rate": 1.4053729666192067e-05, "loss": 0.6051, "step": 1900 }, { "epoch": 3.863821138211382, "grad_norm": 2.2085133527832954, "learning_rate": 1.4047710575085178e-05, "loss": 0.7025, "step": 1901 }, { "epoch": 3.8658536585365852, "grad_norm": 2.642118648581804, "learning_rate": 1.4041689729660323e-05, "loss": 0.6704, "step": 1902 }, { "epoch": 3.867886178861789, "grad_norm": 2.606992578680904, "learning_rate": 1.4035667132526998e-05, "loss": 0.7476, "step": 1903 }, { "epoch": 3.869918699186992, "grad_norm": 2.6400132075805467, "learning_rate": 1.4029642786295452e-05, "loss": 0.6154, "step": 1904 }, { "epoch": 3.8719512195121952, "grad_norm": 2.612618166835745, "learning_rate": 1.4023616693576703e-05, "loss": 0.7163, "step": 1905 }, { "epoch": 3.8739837398373984, "grad_norm": 2.5587436814633304, "learning_rate": 1.4017588856982511e-05, "loss": 0.6025, "step": 1906 }, { "epoch": 3.8760162601626016, "grad_norm": 3.141443825128378, "learning_rate": 1.4011559279125404e-05, "loss": 0.6115, "step": 1907 }, { "epoch": 3.8780487804878048, "grad_norm": 2.4075143770501417, "learning_rate": 1.400552796261866e-05, "loss": 0.7722, "step": 1908 }, { "epoch": 3.880081300813008, "grad_norm": 2.554214533985644, "learning_rate": 1.3999494910076316e-05, "loss": 0.8734, "step": 1909 }, { "epoch": 3.882113821138211, "grad_norm": 2.4266837939714097, "learning_rate": 1.3993460124113154e-05, "loss": 0.9087, "step": 1910 }, { "epoch": 3.8841463414634148, "grad_norm": 2.4104696945082758, "learning_rate": 1.398742360734471e-05, "loss": 0.7481, "step": 1911 }, { "epoch": 3.886178861788618, "grad_norm": 2.7878796909723356, "learning_rate": 1.3981385362387268e-05, "loss": 0.7601, "step": 1912 }, { "epoch": 3.888211382113821, "grad_norm": 2.340641941325179, "learning_rate": 1.3975345391857866e-05, "loss": 0.6428, "step": 1913 }, { "epoch": 3.8902439024390243, "grad_norm": 2.585691923114397, "learning_rate": 1.3969303698374286e-05, "loss": 0.8417, "step": 1914 }, { "epoch": 3.892276422764228, "grad_norm": 2.5155459452202766, "learning_rate": 1.3963260284555058e-05, "loss": 0.7774, "step": 1915 }, { "epoch": 3.894308943089431, "grad_norm": 3.480962724594879, "learning_rate": 1.3957215153019463e-05, "loss": 0.8015, "step": 1916 }, { "epoch": 3.8963414634146343, "grad_norm": 2.7395101635638395, "learning_rate": 1.3951168306387513e-05, "loss": 0.7025, "step": 1917 }, { "epoch": 3.8983739837398375, "grad_norm": 3.3045564484330314, "learning_rate": 1.3945119747279976e-05, "loss": 0.8718, "step": 1918 }, { "epoch": 3.9004065040650406, "grad_norm": 2.3517891794006687, "learning_rate": 1.3939069478318352e-05, "loss": 0.7823, "step": 1919 }, { "epoch": 3.902439024390244, "grad_norm": 2.5252369498537646, "learning_rate": 1.3933017502124897e-05, "loss": 0.8211, "step": 1920 }, { "epoch": 3.904471544715447, "grad_norm": 2.73448061663407, "learning_rate": 1.3926963821322588e-05, "loss": 0.7659, "step": 1921 }, { "epoch": 3.90650406504065, "grad_norm": 2.4668049259701834, "learning_rate": 1.3920908438535158e-05, "loss": 0.7284, "step": 1922 }, { "epoch": 3.908536585365854, "grad_norm": 2.6107530115982236, "learning_rate": 1.3914851356387068e-05, "loss": 0.6624, "step": 1923 }, { "epoch": 3.910569105691057, "grad_norm": 2.114991923935432, "learning_rate": 1.3908792577503514e-05, "loss": 0.7538, "step": 1924 }, { "epoch": 3.91260162601626, "grad_norm": 2.3464875598177617, "learning_rate": 1.3902732104510434e-05, "loss": 0.8702, "step": 1925 }, { "epoch": 3.9146341463414633, "grad_norm": 2.4790670671409796, "learning_rate": 1.3896669940034497e-05, "loss": 0.8256, "step": 1926 }, { "epoch": 3.9166666666666665, "grad_norm": 2.576509629137389, "learning_rate": 1.3890606086703104e-05, "loss": 0.8133, "step": 1927 }, { "epoch": 3.91869918699187, "grad_norm": 2.6586373959540976, "learning_rate": 1.3884540547144393e-05, "loss": 0.6705, "step": 1928 }, { "epoch": 3.9207317073170733, "grad_norm": 2.9624998857009865, "learning_rate": 1.3878473323987223e-05, "loss": 0.6617, "step": 1929 }, { "epoch": 3.9227642276422765, "grad_norm": 2.407790484923276, "learning_rate": 1.3872404419861197e-05, "loss": 0.6476, "step": 1930 }, { "epoch": 3.9247967479674797, "grad_norm": 2.2144027017637984, "learning_rate": 1.3866333837396626e-05, "loss": 0.6198, "step": 1931 }, { "epoch": 3.926829268292683, "grad_norm": 2.546800107335771, "learning_rate": 1.3860261579224574e-05, "loss": 0.7223, "step": 1932 }, { "epoch": 3.928861788617886, "grad_norm": 2.6249174929535943, "learning_rate": 1.3854187647976811e-05, "loss": 0.7752, "step": 1933 }, { "epoch": 3.930894308943089, "grad_norm": 2.338664467188678, "learning_rate": 1.384811204628584e-05, "loss": 0.6518, "step": 1934 }, { "epoch": 3.932926829268293, "grad_norm": 2.3907023426916822, "learning_rate": 1.3842034776784885e-05, "loss": 0.8758, "step": 1935 }, { "epoch": 3.934959349593496, "grad_norm": 2.2109784306840092, "learning_rate": 1.3835955842107897e-05, "loss": 0.7873, "step": 1936 }, { "epoch": 3.936991869918699, "grad_norm": 2.1266811654977475, "learning_rate": 1.3829875244889552e-05, "loss": 0.7342, "step": 1937 }, { "epoch": 3.9390243902439024, "grad_norm": 2.782244141801558, "learning_rate": 1.3823792987765235e-05, "loss": 0.7775, "step": 1938 }, { "epoch": 3.9410569105691056, "grad_norm": 2.796706613923034, "learning_rate": 1.3817709073371056e-05, "loss": 0.8063, "step": 1939 }, { "epoch": 3.943089430894309, "grad_norm": 2.3950347145372803, "learning_rate": 1.3811623504343845e-05, "loss": 0.7813, "step": 1940 }, { "epoch": 3.9451219512195124, "grad_norm": 2.463816126262798, "learning_rate": 1.3805536283321149e-05, "loss": 0.69, "step": 1941 }, { "epoch": 3.9471544715447155, "grad_norm": 2.4770984147850474, "learning_rate": 1.379944741294123e-05, "loss": 0.6957, "step": 1942 }, { "epoch": 3.9491869918699187, "grad_norm": 2.5961119392025775, "learning_rate": 1.3793356895843069e-05, "loss": 0.8217, "step": 1943 }, { "epoch": 3.951219512195122, "grad_norm": 2.569127306249681, "learning_rate": 1.378726473466635e-05, "loss": 0.7819, "step": 1944 }, { "epoch": 3.953252032520325, "grad_norm": 2.4404713073102364, "learning_rate": 1.3781170932051478e-05, "loss": 0.6414, "step": 1945 }, { "epoch": 3.9552845528455283, "grad_norm": 2.4186356771284805, "learning_rate": 1.3775075490639567e-05, "loss": 0.7814, "step": 1946 }, { "epoch": 3.9573170731707314, "grad_norm": 2.5340990670226997, "learning_rate": 1.3768978413072445e-05, "loss": 0.8251, "step": 1947 }, { "epoch": 3.959349593495935, "grad_norm": 2.5415462982514496, "learning_rate": 1.3762879701992642e-05, "loss": 0.6617, "step": 1948 }, { "epoch": 3.9613821138211383, "grad_norm": 2.5307714714243614, "learning_rate": 1.3756779360043403e-05, "loss": 0.8259, "step": 1949 }, { "epoch": 3.9634146341463414, "grad_norm": 2.426571634141852, "learning_rate": 1.3750677389868677e-05, "loss": 0.8565, "step": 1950 }, { "epoch": 3.9654471544715446, "grad_norm": 2.7231322783752665, "learning_rate": 1.3744573794113117e-05, "loss": 0.6261, "step": 1951 }, { "epoch": 3.9674796747967482, "grad_norm": 2.5545006909028656, "learning_rate": 1.373846857542208e-05, "loss": 0.6181, "step": 1952 }, { "epoch": 3.9695121951219514, "grad_norm": 2.649705877271453, "learning_rate": 1.3732361736441633e-05, "loss": 0.6446, "step": 1953 }, { "epoch": 3.9715447154471546, "grad_norm": 2.5020443907439933, "learning_rate": 1.3726253279818538e-05, "loss": 0.6155, "step": 1954 }, { "epoch": 3.9735772357723578, "grad_norm": 2.578462232891694, "learning_rate": 1.3720143208200258e-05, "loss": 0.7112, "step": 1955 }, { "epoch": 3.975609756097561, "grad_norm": 3.003970505333559, "learning_rate": 1.3714031524234965e-05, "loss": 0.7519, "step": 1956 }, { "epoch": 3.977642276422764, "grad_norm": 2.1946638895869426, "learning_rate": 1.370791823057152e-05, "loss": 0.7729, "step": 1957 }, { "epoch": 3.9796747967479673, "grad_norm": 2.818263217157233, "learning_rate": 1.3701803329859486e-05, "loss": 0.8472, "step": 1958 }, { "epoch": 3.9817073170731705, "grad_norm": 2.589151089317489, "learning_rate": 1.369568682474912e-05, "loss": 0.7843, "step": 1959 }, { "epoch": 3.983739837398374, "grad_norm": 2.5521150420131993, "learning_rate": 1.3689568717891381e-05, "loss": 0.852, "step": 1960 }, { "epoch": 3.9857723577235773, "grad_norm": 2.6446789147058345, "learning_rate": 1.3683449011937909e-05, "loss": 0.7674, "step": 1961 }, { "epoch": 3.9878048780487805, "grad_norm": 2.6552406829971766, "learning_rate": 1.367732770954105e-05, "loss": 0.7563, "step": 1962 }, { "epoch": 3.9898373983739837, "grad_norm": 2.4598729286408383, "learning_rate": 1.367120481335384e-05, "loss": 0.8161, "step": 1963 }, { "epoch": 3.991869918699187, "grad_norm": 2.4844437033396307, "learning_rate": 1.3665080326029997e-05, "loss": 0.778, "step": 1964 }, { "epoch": 3.9939024390243905, "grad_norm": 2.557436800720133, "learning_rate": 1.3658954250223941e-05, "loss": 0.7096, "step": 1965 }, { "epoch": 3.9959349593495936, "grad_norm": 2.794762062832815, "learning_rate": 1.3652826588590773e-05, "loss": 0.7756, "step": 1966 }, { "epoch": 3.997967479674797, "grad_norm": 2.7350314429944715, "learning_rate": 1.3646697343786276e-05, "loss": 0.7105, "step": 1967 }, { "epoch": 4.0, "grad_norm": 2.361273550596251, "learning_rate": 1.364056651846693e-05, "loss": 0.7454, "step": 1968 }, { "epoch": 4.0, "eval_loss": 0.9334205985069275, "eval_runtime": 132.241, "eval_samples_per_second": 7.441, "eval_steps_per_second": 0.93, "step": 1968 }, { "epoch": 4.002032520325203, "grad_norm": 2.433872517378198, "learning_rate": 1.3634434115289896e-05, "loss": 0.558, "step": 1969 }, { "epoch": 4.004065040650406, "grad_norm": 2.1568913528364906, "learning_rate": 1.362830013691302e-05, "loss": 0.8079, "step": 1970 }, { "epoch": 4.0060975609756095, "grad_norm": 2.603109813448903, "learning_rate": 1.3622164585994826e-05, "loss": 0.7253, "step": 1971 }, { "epoch": 4.008130081300813, "grad_norm": 2.310596048878363, "learning_rate": 1.3616027465194525e-05, "loss": 0.566, "step": 1972 }, { "epoch": 4.010162601626016, "grad_norm": 2.418577295016606, "learning_rate": 1.3609888777172003e-05, "loss": 0.6229, "step": 1973 }, { "epoch": 4.012195121951219, "grad_norm": 2.392067353400415, "learning_rate": 1.3603748524587833e-05, "loss": 0.762, "step": 1974 }, { "epoch": 4.014227642276423, "grad_norm": 2.7996948577960703, "learning_rate": 1.3597606710103254e-05, "loss": 0.7825, "step": 1975 }, { "epoch": 4.016260162601626, "grad_norm": 2.2519411550043795, "learning_rate": 1.35914633363802e-05, "loss": 0.5414, "step": 1976 }, { "epoch": 4.0182926829268295, "grad_norm": 2.7480093179461518, "learning_rate": 1.3585318406081262e-05, "loss": 0.8393, "step": 1977 }, { "epoch": 4.020325203252033, "grad_norm": 2.4776275428707644, "learning_rate": 1.3579171921869714e-05, "loss": 0.5637, "step": 1978 }, { "epoch": 4.022357723577236, "grad_norm": 2.3882634856988645, "learning_rate": 1.3573023886409508e-05, "loss": 0.6448, "step": 1979 }, { "epoch": 4.024390243902439, "grad_norm": 2.636645138241011, "learning_rate": 1.356687430236526e-05, "loss": 0.6665, "step": 1980 }, { "epoch": 4.026422764227642, "grad_norm": 2.6961428990646588, "learning_rate": 1.3560723172402264e-05, "loss": 0.6323, "step": 1981 }, { "epoch": 4.028455284552845, "grad_norm": 2.613731126583165, "learning_rate": 1.3554570499186477e-05, "loss": 0.6309, "step": 1982 }, { "epoch": 4.030487804878049, "grad_norm": 3.047661163324738, "learning_rate": 1.3548416285384529e-05, "loss": 0.8449, "step": 1983 }, { "epoch": 4.032520325203252, "grad_norm": 2.7065691672176637, "learning_rate": 1.3542260533663723e-05, "loss": 0.6895, "step": 1984 }, { "epoch": 4.034552845528455, "grad_norm": 2.4627878296702033, "learning_rate": 1.3536103246692014e-05, "loss": 0.6936, "step": 1985 }, { "epoch": 4.036585365853658, "grad_norm": 2.6973570337725903, "learning_rate": 1.352994442713804e-05, "loss": 0.6775, "step": 1986 }, { "epoch": 4.038617886178862, "grad_norm": 3.220907651962091, "learning_rate": 1.3523784077671089e-05, "loss": 0.8489, "step": 1987 }, { "epoch": 4.040650406504065, "grad_norm": 2.624432211022582, "learning_rate": 1.351762220096112e-05, "loss": 0.6395, "step": 1988 }, { "epoch": 4.0426829268292686, "grad_norm": 2.842058522068938, "learning_rate": 1.351145879967875e-05, "loss": 0.5965, "step": 1989 }, { "epoch": 4.044715447154472, "grad_norm": 2.9603964939063423, "learning_rate": 1.3505293876495257e-05, "loss": 0.7218, "step": 1990 }, { "epoch": 4.046747967479675, "grad_norm": 2.817879753317161, "learning_rate": 1.3499127434082588e-05, "loss": 0.5413, "step": 1991 }, { "epoch": 4.048780487804878, "grad_norm": 2.7123069564696016, "learning_rate": 1.3492959475113332e-05, "loss": 0.6418, "step": 1992 }, { "epoch": 4.050813008130081, "grad_norm": 3.155287678644206, "learning_rate": 1.3486790002260745e-05, "loss": 0.7491, "step": 1993 }, { "epoch": 4.0528455284552845, "grad_norm": 2.8662934357326675, "learning_rate": 1.348061901819874e-05, "loss": 0.6709, "step": 1994 }, { "epoch": 4.054878048780488, "grad_norm": 3.10801015724544, "learning_rate": 1.3474446525601879e-05, "loss": 0.6428, "step": 1995 }, { "epoch": 4.056910569105691, "grad_norm": 3.3603355800930474, "learning_rate": 1.3468272527145388e-05, "loss": 0.8265, "step": 1996 }, { "epoch": 4.058943089430894, "grad_norm": 2.7171935628488115, "learning_rate": 1.3462097025505134e-05, "loss": 0.6974, "step": 1997 }, { "epoch": 4.060975609756097, "grad_norm": 2.58451257883866, "learning_rate": 1.3455920023357644e-05, "loss": 0.5502, "step": 1998 }, { "epoch": 4.063008130081301, "grad_norm": 3.3566576570074527, "learning_rate": 1.344974152338009e-05, "loss": 0.6222, "step": 1999 }, { "epoch": 4.065040650406504, "grad_norm": 3.054738684357665, "learning_rate": 1.3443561528250295e-05, "loss": 0.6269, "step": 2000 }, { "epoch": 4.067073170731708, "grad_norm": 3.031050898921419, "learning_rate": 1.343738004064673e-05, "loss": 0.7913, "step": 2001 }, { "epoch": 4.069105691056911, "grad_norm": 2.966117083339379, "learning_rate": 1.3431197063248514e-05, "loss": 0.6457, "step": 2002 }, { "epoch": 4.071138211382114, "grad_norm": 2.627278564421329, "learning_rate": 1.3425012598735416e-05, "loss": 0.5957, "step": 2003 }, { "epoch": 4.073170731707317, "grad_norm": 2.889343019007643, "learning_rate": 1.3418826649787834e-05, "loss": 0.7792, "step": 2004 }, { "epoch": 4.07520325203252, "grad_norm": 3.1614042657748582, "learning_rate": 1.341263921908683e-05, "loss": 0.7144, "step": 2005 }, { "epoch": 4.0772357723577235, "grad_norm": 2.8204620099898374, "learning_rate": 1.3406450309314085e-05, "loss": 0.6899, "step": 2006 }, { "epoch": 4.079268292682927, "grad_norm": 2.6903244004139037, "learning_rate": 1.3400259923151948e-05, "loss": 0.6437, "step": 2007 }, { "epoch": 4.08130081300813, "grad_norm": 3.101098984305282, "learning_rate": 1.3394068063283387e-05, "loss": 0.6142, "step": 2008 }, { "epoch": 4.083333333333333, "grad_norm": 3.184924975741401, "learning_rate": 1.3387874732392017e-05, "loss": 0.7295, "step": 2009 }, { "epoch": 4.085365853658536, "grad_norm": 3.398470885325908, "learning_rate": 1.3381679933162088e-05, "loss": 0.7743, "step": 2010 }, { "epoch": 4.08739837398374, "grad_norm": 2.872687627871976, "learning_rate": 1.3375483668278486e-05, "loss": 0.7698, "step": 2011 }, { "epoch": 4.0894308943089435, "grad_norm": 2.3392366952914507, "learning_rate": 1.3369285940426737e-05, "loss": 0.7406, "step": 2012 }, { "epoch": 4.091463414634147, "grad_norm": 2.735915858357094, "learning_rate": 1.3363086752292996e-05, "loss": 0.5707, "step": 2013 }, { "epoch": 4.09349593495935, "grad_norm": 3.077662855035884, "learning_rate": 1.3356886106564054e-05, "loss": 0.6885, "step": 2014 }, { "epoch": 4.095528455284553, "grad_norm": 3.180560854464483, "learning_rate": 1.3350684005927332e-05, "loss": 0.5993, "step": 2015 }, { "epoch": 4.097560975609756, "grad_norm": 3.4015707396760675, "learning_rate": 1.334448045307088e-05, "loss": 0.8725, "step": 2016 }, { "epoch": 4.099593495934959, "grad_norm": 2.870851781313507, "learning_rate": 1.3338275450683384e-05, "loss": 0.7485, "step": 2017 }, { "epoch": 4.1016260162601625, "grad_norm": 2.8853750221702277, "learning_rate": 1.3332069001454146e-05, "loss": 0.5874, "step": 2018 }, { "epoch": 4.103658536585366, "grad_norm": 2.786420300384197, "learning_rate": 1.3325861108073112e-05, "loss": 0.6138, "step": 2019 }, { "epoch": 4.105691056910569, "grad_norm": 2.8928207230257517, "learning_rate": 1.331965177323084e-05, "loss": 0.5798, "step": 2020 }, { "epoch": 4.107723577235772, "grad_norm": 3.519286744247599, "learning_rate": 1.331344099961852e-05, "loss": 0.5615, "step": 2021 }, { "epoch": 4.109756097560975, "grad_norm": 2.845999079590994, "learning_rate": 1.3307228789927953e-05, "loss": 0.782, "step": 2022 }, { "epoch": 4.111788617886178, "grad_norm": 2.7610542034588157, "learning_rate": 1.3301015146851584e-05, "loss": 0.7156, "step": 2023 }, { "epoch": 4.1138211382113825, "grad_norm": 2.5831134638040325, "learning_rate": 1.3294800073082464e-05, "loss": 0.7226, "step": 2024 }, { "epoch": 4.115853658536586, "grad_norm": 3.368737646204141, "learning_rate": 1.3288583571314267e-05, "loss": 0.682, "step": 2025 }, { "epoch": 4.117886178861789, "grad_norm": 3.2701965460194202, "learning_rate": 1.3282365644241287e-05, "loss": 0.6251, "step": 2026 }, { "epoch": 4.119918699186992, "grad_norm": 2.981275865108666, "learning_rate": 1.3276146294558434e-05, "loss": 0.8952, "step": 2027 }, { "epoch": 4.121951219512195, "grad_norm": 2.7034361245980625, "learning_rate": 1.3269925524961237e-05, "loss": 0.6472, "step": 2028 }, { "epoch": 4.123983739837398, "grad_norm": 2.791351048797362, "learning_rate": 1.3263703338145842e-05, "loss": 0.6746, "step": 2029 }, { "epoch": 4.126016260162602, "grad_norm": 2.4635263229206044, "learning_rate": 1.3257479736809003e-05, "loss": 0.6322, "step": 2030 }, { "epoch": 4.128048780487805, "grad_norm": 3.08539410264463, "learning_rate": 1.3251254723648091e-05, "loss": 0.7549, "step": 2031 }, { "epoch": 4.130081300813008, "grad_norm": 3.1396423875892543, "learning_rate": 1.3245028301361086e-05, "loss": 0.8019, "step": 2032 }, { "epoch": 4.132113821138211, "grad_norm": 2.8866935140992807, "learning_rate": 1.3238800472646591e-05, "loss": 0.6169, "step": 2033 }, { "epoch": 4.134146341463414, "grad_norm": 2.645991885860708, "learning_rate": 1.3232571240203797e-05, "loss": 0.4918, "step": 2034 }, { "epoch": 4.1361788617886175, "grad_norm": 2.953197857633758, "learning_rate": 1.3226340606732524e-05, "loss": 0.6325, "step": 2035 }, { "epoch": 4.138211382113822, "grad_norm": 2.88635743477074, "learning_rate": 1.3220108574933185e-05, "loss": 0.6915, "step": 2036 }, { "epoch": 4.140243902439025, "grad_norm": 2.8757421953380455, "learning_rate": 1.3213875147506807e-05, "loss": 0.7721, "step": 2037 }, { "epoch": 4.142276422764228, "grad_norm": 3.231846975372353, "learning_rate": 1.320764032715502e-05, "loss": 0.7791, "step": 2038 }, { "epoch": 4.144308943089431, "grad_norm": 2.7468379887510745, "learning_rate": 1.3201404116580055e-05, "loss": 0.6885, "step": 2039 }, { "epoch": 4.146341463414634, "grad_norm": 3.479209956689003, "learning_rate": 1.3195166518484748e-05, "loss": 0.772, "step": 2040 }, { "epoch": 4.1483739837398375, "grad_norm": 2.8170401013834834, "learning_rate": 1.3188927535572535e-05, "loss": 0.646, "step": 2041 }, { "epoch": 4.150406504065041, "grad_norm": 3.3368581130360093, "learning_rate": 1.3182687170547458e-05, "loss": 0.7246, "step": 2042 }, { "epoch": 4.152439024390244, "grad_norm": 3.4850065939866113, "learning_rate": 1.3176445426114145e-05, "loss": 0.6605, "step": 2043 }, { "epoch": 4.154471544715447, "grad_norm": 2.7114896479956743, "learning_rate": 1.317020230497784e-05, "loss": 0.7328, "step": 2044 }, { "epoch": 4.15650406504065, "grad_norm": 2.8208089608139857, "learning_rate": 1.3163957809844366e-05, "loss": 0.8091, "step": 2045 }, { "epoch": 4.158536585365853, "grad_norm": 2.7724806383298906, "learning_rate": 1.3157711943420151e-05, "loss": 0.5781, "step": 2046 }, { "epoch": 4.1605691056910565, "grad_norm": 2.818775318166426, "learning_rate": 1.315146470841222e-05, "loss": 0.6055, "step": 2047 }, { "epoch": 4.16260162601626, "grad_norm": 3.2543653094953804, "learning_rate": 1.3145216107528178e-05, "loss": 0.6754, "step": 2048 }, { "epoch": 4.164634146341464, "grad_norm": 2.876986531294622, "learning_rate": 1.3138966143476237e-05, "loss": 0.7756, "step": 2049 }, { "epoch": 4.166666666666667, "grad_norm": 3.117925643307231, "learning_rate": 1.3132714818965193e-05, "loss": 0.7037, "step": 2050 }, { "epoch": 4.16869918699187, "grad_norm": 3.095261669823429, "learning_rate": 1.312646213670443e-05, "loss": 0.6549, "step": 2051 }, { "epoch": 4.170731707317073, "grad_norm": 2.9545185378258574, "learning_rate": 1.3120208099403926e-05, "loss": 0.716, "step": 2052 }, { "epoch": 4.1727642276422765, "grad_norm": 2.8133968582813758, "learning_rate": 1.311395270977424e-05, "loss": 0.686, "step": 2053 }, { "epoch": 4.17479674796748, "grad_norm": 2.307936316835479, "learning_rate": 1.3107695970526522e-05, "loss": 0.52, "step": 2054 }, { "epoch": 4.176829268292683, "grad_norm": 2.7956392356141393, "learning_rate": 1.3101437884372502e-05, "loss": 0.7332, "step": 2055 }, { "epoch": 4.178861788617886, "grad_norm": 2.5281892732929934, "learning_rate": 1.3095178454024496e-05, "loss": 0.5902, "step": 2056 }, { "epoch": 4.180894308943089, "grad_norm": 2.831613964756424, "learning_rate": 1.308891768219541e-05, "loss": 0.722, "step": 2057 }, { "epoch": 4.182926829268292, "grad_norm": 3.125322668161164, "learning_rate": 1.3082655571598718e-05, "loss": 0.5872, "step": 2058 }, { "epoch": 4.184959349593496, "grad_norm": 3.1790821841535903, "learning_rate": 1.3076392124948485e-05, "loss": 0.4813, "step": 2059 }, { "epoch": 4.186991869918699, "grad_norm": 3.1807087946080004, "learning_rate": 1.3070127344959348e-05, "loss": 0.6708, "step": 2060 }, { "epoch": 4.189024390243903, "grad_norm": 2.9056776683375976, "learning_rate": 1.3063861234346527e-05, "loss": 0.5406, "step": 2061 }, { "epoch": 4.191056910569106, "grad_norm": 2.8426540052873586, "learning_rate": 1.3057593795825814e-05, "loss": 0.6716, "step": 2062 }, { "epoch": 4.193089430894309, "grad_norm": 3.3532437646202533, "learning_rate": 1.3051325032113582e-05, "loss": 0.7658, "step": 2063 }, { "epoch": 4.195121951219512, "grad_norm": 2.7281382784368673, "learning_rate": 1.3045054945926775e-05, "loss": 0.5327, "step": 2064 }, { "epoch": 4.1971544715447155, "grad_norm": 2.701563571322171, "learning_rate": 1.3038783539982909e-05, "loss": 0.5595, "step": 2065 }, { "epoch": 4.199186991869919, "grad_norm": 3.1088089473453717, "learning_rate": 1.3032510817000074e-05, "loss": 0.7309, "step": 2066 }, { "epoch": 4.201219512195122, "grad_norm": 2.64707169255667, "learning_rate": 1.3026236779696932e-05, "loss": 0.4798, "step": 2067 }, { "epoch": 4.203252032520325, "grad_norm": 3.1622251108404447, "learning_rate": 1.3019961430792711e-05, "loss": 0.8057, "step": 2068 }, { "epoch": 4.205284552845528, "grad_norm": 2.985583270750285, "learning_rate": 1.3013684773007214e-05, "loss": 0.7018, "step": 2069 }, { "epoch": 4.2073170731707314, "grad_norm": 2.9384257644802383, "learning_rate": 1.3007406809060799e-05, "loss": 0.7328, "step": 2070 }, { "epoch": 4.209349593495935, "grad_norm": 2.8407187329668644, "learning_rate": 1.3001127541674406e-05, "loss": 0.6627, "step": 2071 }, { "epoch": 4.211382113821138, "grad_norm": 3.127448325287205, "learning_rate": 1.2994846973569524e-05, "loss": 0.603, "step": 2072 }, { "epoch": 4.213414634146342, "grad_norm": 3.2157860988712086, "learning_rate": 1.298856510746822e-05, "loss": 0.7376, "step": 2073 }, { "epoch": 4.215447154471545, "grad_norm": 2.9317183398881093, "learning_rate": 1.2982281946093117e-05, "loss": 0.6782, "step": 2074 }, { "epoch": 4.217479674796748, "grad_norm": 3.3478381083060524, "learning_rate": 1.2975997492167399e-05, "loss": 0.6952, "step": 2075 }, { "epoch": 4.219512195121951, "grad_norm": 3.0824960388365024, "learning_rate": 1.2969711748414804e-05, "loss": 0.6196, "step": 2076 }, { "epoch": 4.221544715447155, "grad_norm": 3.16107751282541, "learning_rate": 1.296342471755964e-05, "loss": 0.6384, "step": 2077 }, { "epoch": 4.223577235772358, "grad_norm": 3.2019120239152903, "learning_rate": 1.2957136402326776e-05, "loss": 0.8081, "step": 2078 }, { "epoch": 4.225609756097561, "grad_norm": 2.8445896211302713, "learning_rate": 1.2950846805441621e-05, "loss": 0.7224, "step": 2079 }, { "epoch": 4.227642276422764, "grad_norm": 2.922074552562058, "learning_rate": 1.2944555929630152e-05, "loss": 0.58, "step": 2080 }, { "epoch": 4.229674796747967, "grad_norm": 3.1446026831419807, "learning_rate": 1.2938263777618895e-05, "loss": 0.6354, "step": 2081 }, { "epoch": 4.2317073170731705, "grad_norm": 2.9256211613878373, "learning_rate": 1.2931970352134934e-05, "loss": 0.5993, "step": 2082 }, { "epoch": 4.233739837398374, "grad_norm": 2.8904270275932737, "learning_rate": 1.2925675655905901e-05, "loss": 0.6476, "step": 2083 }, { "epoch": 4.235772357723577, "grad_norm": 2.5218567450495977, "learning_rate": 1.2919379691659979e-05, "loss": 0.7375, "step": 2084 }, { "epoch": 4.237804878048781, "grad_norm": 3.1875024600982873, "learning_rate": 1.2913082462125908e-05, "loss": 0.6636, "step": 2085 }, { "epoch": 4.239837398373984, "grad_norm": 2.4377685184489803, "learning_rate": 1.2906783970032956e-05, "loss": 0.5443, "step": 2086 }, { "epoch": 4.241869918699187, "grad_norm": 2.5714782055849925, "learning_rate": 1.2900484218110964e-05, "loss": 0.608, "step": 2087 }, { "epoch": 4.2439024390243905, "grad_norm": 3.157387349289292, "learning_rate": 1.2894183209090304e-05, "loss": 0.6187, "step": 2088 }, { "epoch": 4.245934959349594, "grad_norm": 2.611168539530284, "learning_rate": 1.2887880945701895e-05, "loss": 0.4577, "step": 2089 }, { "epoch": 4.247967479674797, "grad_norm": 2.738917170230175, "learning_rate": 1.28815774306772e-05, "loss": 0.6829, "step": 2090 }, { "epoch": 4.25, "grad_norm": 2.9974118054076824, "learning_rate": 1.2875272666748226e-05, "loss": 0.8013, "step": 2091 }, { "epoch": 4.252032520325203, "grad_norm": 2.743045029633641, "learning_rate": 1.2868966656647522e-05, "loss": 0.4986, "step": 2092 }, { "epoch": 4.254065040650406, "grad_norm": 3.0375634296262177, "learning_rate": 1.286265940310817e-05, "loss": 0.619, "step": 2093 }, { "epoch": 4.2560975609756095, "grad_norm": 2.837258007584117, "learning_rate": 1.2856350908863802e-05, "loss": 0.6476, "step": 2094 }, { "epoch": 4.258130081300813, "grad_norm": 2.725947742982041, "learning_rate": 1.2850041176648578e-05, "loss": 0.5563, "step": 2095 }, { "epoch": 4.260162601626016, "grad_norm": 3.054594285430433, "learning_rate": 1.2843730209197203e-05, "loss": 0.7147, "step": 2096 }, { "epoch": 4.262195121951219, "grad_norm": 3.136154770661375, "learning_rate": 1.2837418009244908e-05, "loss": 0.5421, "step": 2097 }, { "epoch": 4.264227642276423, "grad_norm": 2.8681343502552057, "learning_rate": 1.2831104579527467e-05, "loss": 0.5937, "step": 2098 }, { "epoch": 4.266260162601626, "grad_norm": 3.221221689863481, "learning_rate": 1.2824789922781182e-05, "loss": 0.8534, "step": 2099 }, { "epoch": 4.2682926829268295, "grad_norm": 2.706757191088968, "learning_rate": 1.2818474041742885e-05, "loss": 0.5836, "step": 2100 }, { "epoch": 4.270325203252033, "grad_norm": 3.5779436585159328, "learning_rate": 1.2812156939149947e-05, "loss": 0.7854, "step": 2101 }, { "epoch": 4.272357723577236, "grad_norm": 3.1942835644740724, "learning_rate": 1.280583861774026e-05, "loss": 0.7086, "step": 2102 }, { "epoch": 4.274390243902439, "grad_norm": 2.9862529744373516, "learning_rate": 1.2799519080252247e-05, "loss": 0.5694, "step": 2103 }, { "epoch": 4.276422764227642, "grad_norm": 3.314373652799172, "learning_rate": 1.2793198329424858e-05, "loss": 0.6619, "step": 2104 }, { "epoch": 4.278455284552845, "grad_norm": 3.217649580482955, "learning_rate": 1.2786876367997571e-05, "loss": 0.629, "step": 2105 }, { "epoch": 4.280487804878049, "grad_norm": 3.2546259448411416, "learning_rate": 1.2780553198710388e-05, "loss": 0.5977, "step": 2106 }, { "epoch": 4.282520325203252, "grad_norm": 3.242746555073596, "learning_rate": 1.277422882430383e-05, "loss": 0.663, "step": 2107 }, { "epoch": 4.284552845528455, "grad_norm": 3.5241259609476976, "learning_rate": 1.2767903247518945e-05, "loss": 0.7188, "step": 2108 }, { "epoch": 4.286585365853658, "grad_norm": 3.187845119847995, "learning_rate": 1.2761576471097298e-05, "loss": 0.6073, "step": 2109 }, { "epoch": 4.288617886178862, "grad_norm": 3.2618292154353368, "learning_rate": 1.275524849778098e-05, "loss": 0.6754, "step": 2110 }, { "epoch": 4.290650406504065, "grad_norm": 3.1216587788457253, "learning_rate": 1.27489193303126e-05, "loss": 0.6654, "step": 2111 }, { "epoch": 4.2926829268292686, "grad_norm": 3.051801756282085, "learning_rate": 1.2742588971435276e-05, "loss": 0.6604, "step": 2112 }, { "epoch": 4.294715447154472, "grad_norm": 3.1461666170651172, "learning_rate": 1.2736257423892657e-05, "loss": 0.687, "step": 2113 }, { "epoch": 4.296747967479675, "grad_norm": 3.316040949203775, "learning_rate": 1.2729924690428887e-05, "loss": 0.7799, "step": 2114 }, { "epoch": 4.298780487804878, "grad_norm": 2.7957787812815944, "learning_rate": 1.2723590773788645e-05, "loss": 0.6587, "step": 2115 }, { "epoch": 4.300813008130081, "grad_norm": 3.137359728348412, "learning_rate": 1.2717255676717106e-05, "loss": 0.7818, "step": 2116 }, { "epoch": 4.3028455284552845, "grad_norm": 2.9390521121948225, "learning_rate": 1.2710919401959974e-05, "loss": 0.6912, "step": 2117 }, { "epoch": 4.304878048780488, "grad_norm": 3.1906201429111154, "learning_rate": 1.2704581952263443e-05, "loss": 0.6575, "step": 2118 }, { "epoch": 4.306910569105691, "grad_norm": 2.8935787765359327, "learning_rate": 1.2698243330374231e-05, "loss": 0.6815, "step": 2119 }, { "epoch": 4.308943089430894, "grad_norm": 2.7880249790129423, "learning_rate": 1.2691903539039563e-05, "loss": 0.6214, "step": 2120 }, { "epoch": 4.310975609756097, "grad_norm": 2.992355479157598, "learning_rate": 1.268556258100716e-05, "loss": 0.6764, "step": 2121 }, { "epoch": 4.3130081300813, "grad_norm": 2.9951225368188457, "learning_rate": 1.2679220459025264e-05, "loss": 0.6512, "step": 2122 }, { "epoch": 4.315040650406504, "grad_norm": 3.002483260512474, "learning_rate": 1.267287717584261e-05, "loss": 0.6732, "step": 2123 }, { "epoch": 4.317073170731708, "grad_norm": 2.664729496806343, "learning_rate": 1.2666532734208437e-05, "loss": 0.7287, "step": 2124 }, { "epoch": 4.319105691056911, "grad_norm": 2.7734505056306333, "learning_rate": 1.2660187136872495e-05, "loss": 0.7141, "step": 2125 }, { "epoch": 4.321138211382114, "grad_norm": 2.7421328599659125, "learning_rate": 1.2653840386585024e-05, "loss": 0.6667, "step": 2126 }, { "epoch": 4.323170731707317, "grad_norm": 2.8904815506508323, "learning_rate": 1.2647492486096772e-05, "loss": 0.7411, "step": 2127 }, { "epoch": 4.32520325203252, "grad_norm": 2.964545758346194, "learning_rate": 1.264114343815898e-05, "loss": 0.5901, "step": 2128 }, { "epoch": 4.3272357723577235, "grad_norm": 3.160057107296968, "learning_rate": 1.2634793245523393e-05, "loss": 0.9504, "step": 2129 }, { "epoch": 4.329268292682927, "grad_norm": 2.5999289055801382, "learning_rate": 1.262844191094224e-05, "loss": 0.6117, "step": 2130 }, { "epoch": 4.33130081300813, "grad_norm": 2.9884016197151775, "learning_rate": 1.2622089437168255e-05, "loss": 0.6877, "step": 2131 }, { "epoch": 4.333333333333333, "grad_norm": 3.0546774700128543, "learning_rate": 1.2615735826954664e-05, "loss": 0.6219, "step": 2132 }, { "epoch": 4.335365853658536, "grad_norm": 3.003555536471429, "learning_rate": 1.2609381083055188e-05, "loss": 0.5992, "step": 2133 }, { "epoch": 4.33739837398374, "grad_norm": 2.7277738177357884, "learning_rate": 1.2603025208224035e-05, "loss": 0.6827, "step": 2134 }, { "epoch": 4.3394308943089435, "grad_norm": 3.5690598237888347, "learning_rate": 1.25966682052159e-05, "loss": 0.6202, "step": 2135 }, { "epoch": 4.341463414634147, "grad_norm": 2.8163993725450287, "learning_rate": 1.2590310076785974e-05, "loss": 0.6901, "step": 2136 }, { "epoch": 4.34349593495935, "grad_norm": 2.900932449364065, "learning_rate": 1.2583950825689933e-05, "loss": 0.6636, "step": 2137 }, { "epoch": 4.345528455284553, "grad_norm": 3.3061788956827627, "learning_rate": 1.2577590454683936e-05, "loss": 0.7001, "step": 2138 }, { "epoch": 4.347560975609756, "grad_norm": 3.081506866553319, "learning_rate": 1.257122896652464e-05, "loss": 0.7048, "step": 2139 }, { "epoch": 4.349593495934959, "grad_norm": 3.1107813777175335, "learning_rate": 1.256486636396917e-05, "loss": 0.7231, "step": 2140 }, { "epoch": 4.3516260162601625, "grad_norm": 3.1544384769647547, "learning_rate": 1.2558502649775144e-05, "loss": 0.6615, "step": 2141 }, { "epoch": 4.353658536585366, "grad_norm": 3.020179394529575, "learning_rate": 1.2552137826700654e-05, "loss": 0.4696, "step": 2142 }, { "epoch": 4.355691056910569, "grad_norm": 2.8331949038691824, "learning_rate": 1.2545771897504285e-05, "loss": 0.5544, "step": 2143 }, { "epoch": 4.357723577235772, "grad_norm": 3.669688711578965, "learning_rate": 1.2539404864945087e-05, "loss": 0.7683, "step": 2144 }, { "epoch": 4.359756097560975, "grad_norm": 3.256195186029363, "learning_rate": 1.2533036731782599e-05, "loss": 0.5853, "step": 2145 }, { "epoch": 4.361788617886178, "grad_norm": 3.6490747177904685, "learning_rate": 1.2526667500776835e-05, "loss": 0.5596, "step": 2146 }, { "epoch": 4.3638211382113825, "grad_norm": 3.0826067427701624, "learning_rate": 1.2520297174688278e-05, "loss": 0.6381, "step": 2147 }, { "epoch": 4.365853658536586, "grad_norm": 3.3147536218549427, "learning_rate": 1.2513925756277894e-05, "loss": 0.5546, "step": 2148 }, { "epoch": 4.367886178861789, "grad_norm": 3.5737399532972125, "learning_rate": 1.2507553248307116e-05, "loss": 0.8026, "step": 2149 }, { "epoch": 4.369918699186992, "grad_norm": 3.0734376259863465, "learning_rate": 1.2501179653537858e-05, "loss": 0.7058, "step": 2150 }, { "epoch": 4.371951219512195, "grad_norm": 3.0559715363754356, "learning_rate": 1.2494804974732494e-05, "loss": 0.7104, "step": 2151 }, { "epoch": 4.373983739837398, "grad_norm": 3.3330732555990195, "learning_rate": 1.2488429214653871e-05, "loss": 0.598, "step": 2152 }, { "epoch": 4.376016260162602, "grad_norm": 3.1441311150613505, "learning_rate": 1.2482052376065316e-05, "loss": 0.5972, "step": 2153 }, { "epoch": 4.378048780487805, "grad_norm": 3.483928727844598, "learning_rate": 1.2475674461730604e-05, "loss": 0.673, "step": 2154 }, { "epoch": 4.380081300813008, "grad_norm": 3.3506917711951485, "learning_rate": 1.2469295474413996e-05, "loss": 0.6116, "step": 2155 }, { "epoch": 4.382113821138211, "grad_norm": 3.4343001402982285, "learning_rate": 1.24629154168802e-05, "loss": 0.7738, "step": 2156 }, { "epoch": 4.384146341463414, "grad_norm": 3.2130609200905407, "learning_rate": 1.2456534291894399e-05, "loss": 0.7399, "step": 2157 }, { "epoch": 4.3861788617886175, "grad_norm": 3.5094446167972198, "learning_rate": 1.2450152102222242e-05, "loss": 0.6341, "step": 2158 }, { "epoch": 4.388211382113822, "grad_norm": 2.9562397824542863, "learning_rate": 1.2443768850629822e-05, "loss": 0.6077, "step": 2159 }, { "epoch": 4.390243902439025, "grad_norm": 3.308134146595336, "learning_rate": 1.2437384539883715e-05, "loss": 0.5475, "step": 2160 }, { "epoch": 4.392276422764228, "grad_norm": 2.67780736019056, "learning_rate": 1.2430999172750937e-05, "loss": 0.4936, "step": 2161 }, { "epoch": 4.394308943089431, "grad_norm": 3.3085209759795307, "learning_rate": 1.2424612751998972e-05, "loss": 0.5944, "step": 2162 }, { "epoch": 4.396341463414634, "grad_norm": 2.9197031972153598, "learning_rate": 1.2418225280395757e-05, "loss": 0.8089, "step": 2163 }, { "epoch": 4.3983739837398375, "grad_norm": 3.0852317028051113, "learning_rate": 1.2411836760709686e-05, "loss": 0.6942, "step": 2164 }, { "epoch": 4.400406504065041, "grad_norm": 3.147700104056151, "learning_rate": 1.2405447195709612e-05, "loss": 0.6374, "step": 2165 }, { "epoch": 4.402439024390244, "grad_norm": 4.11990329688501, "learning_rate": 1.2399056588164831e-05, "loss": 0.6733, "step": 2166 }, { "epoch": 4.404471544715447, "grad_norm": 3.0600348785429112, "learning_rate": 1.2392664940845094e-05, "loss": 0.7694, "step": 2167 }, { "epoch": 4.40650406504065, "grad_norm": 3.149380973828591, "learning_rate": 1.2386272256520606e-05, "loss": 0.6357, "step": 2168 }, { "epoch": 4.408536585365853, "grad_norm": 3.113814468689196, "learning_rate": 1.2379878537962024e-05, "loss": 0.5784, "step": 2169 }, { "epoch": 4.4105691056910565, "grad_norm": 3.186477629977461, "learning_rate": 1.2373483787940444e-05, "loss": 0.4877, "step": 2170 }, { "epoch": 4.41260162601626, "grad_norm": 3.002791167624794, "learning_rate": 1.2367088009227415e-05, "loss": 0.5846, "step": 2171 }, { "epoch": 4.414634146341464, "grad_norm": 3.4124207100865727, "learning_rate": 1.2360691204594937e-05, "loss": 0.6174, "step": 2172 }, { "epoch": 4.416666666666667, "grad_norm": 3.4195807061880217, "learning_rate": 1.2354293376815438e-05, "loss": 0.6348, "step": 2173 }, { "epoch": 4.41869918699187, "grad_norm": 3.082587183252866, "learning_rate": 1.234789452866181e-05, "loss": 0.6187, "step": 2174 }, { "epoch": 4.420731707317073, "grad_norm": 3.228220596066274, "learning_rate": 1.234149466290737e-05, "loss": 0.7761, "step": 2175 }, { "epoch": 4.4227642276422765, "grad_norm": 2.8656099445755934, "learning_rate": 1.2335093782325889e-05, "loss": 0.5623, "step": 2176 }, { "epoch": 4.42479674796748, "grad_norm": 2.7400766612439567, "learning_rate": 1.2328691889691566e-05, "loss": 0.7071, "step": 2177 }, { "epoch": 4.426829268292683, "grad_norm": 2.967521868148167, "learning_rate": 1.2322288987779055e-05, "loss": 0.6245, "step": 2178 }, { "epoch": 4.428861788617886, "grad_norm": 3.0136489266663533, "learning_rate": 1.2315885079363427e-05, "loss": 0.6206, "step": 2179 }, { "epoch": 4.430894308943089, "grad_norm": 3.151975307837421, "learning_rate": 1.2309480167220203e-05, "loss": 0.6049, "step": 2180 }, { "epoch": 4.432926829268292, "grad_norm": 2.713893656771352, "learning_rate": 1.2303074254125335e-05, "loss": 0.747, "step": 2181 }, { "epoch": 4.434959349593496, "grad_norm": 2.6808248354610926, "learning_rate": 1.2296667342855213e-05, "loss": 0.6505, "step": 2182 }, { "epoch": 4.4369918699187, "grad_norm": 2.7305101077901828, "learning_rate": 1.2290259436186653e-05, "loss": 0.6363, "step": 2183 }, { "epoch": 4.439024390243903, "grad_norm": 2.539848252407869, "learning_rate": 1.2283850536896907e-05, "loss": 0.8602, "step": 2184 }, { "epoch": 4.441056910569106, "grad_norm": 3.2285807939591527, "learning_rate": 1.227744064776365e-05, "loss": 0.7298, "step": 2185 }, { "epoch": 4.443089430894309, "grad_norm": 3.0502327661115283, "learning_rate": 1.2271029771565e-05, "loss": 0.5414, "step": 2186 }, { "epoch": 4.445121951219512, "grad_norm": 2.579008366224955, "learning_rate": 1.2264617911079486e-05, "loss": 0.6975, "step": 2187 }, { "epoch": 4.4471544715447155, "grad_norm": 2.88719327111785, "learning_rate": 1.2258205069086082e-05, "loss": 0.7459, "step": 2188 }, { "epoch": 4.449186991869919, "grad_norm": 3.464836412705549, "learning_rate": 1.2251791248364169e-05, "loss": 0.7826, "step": 2189 }, { "epoch": 4.451219512195122, "grad_norm": 2.8893016353309746, "learning_rate": 1.2245376451693564e-05, "loss": 0.5768, "step": 2190 }, { "epoch": 4.453252032520325, "grad_norm": 3.3840462623802603, "learning_rate": 1.2238960681854502e-05, "loss": 0.5909, "step": 2191 }, { "epoch": 4.455284552845528, "grad_norm": 2.877837054142557, "learning_rate": 1.2232543941627641e-05, "loss": 0.6734, "step": 2192 }, { "epoch": 4.4573170731707314, "grad_norm": 3.0324877384826308, "learning_rate": 1.2226126233794064e-05, "loss": 0.7811, "step": 2193 }, { "epoch": 4.459349593495935, "grad_norm": 2.646241596804522, "learning_rate": 1.2219707561135268e-05, "loss": 0.7663, "step": 2194 }, { "epoch": 4.461382113821138, "grad_norm": 3.035558583472837, "learning_rate": 1.2213287926433167e-05, "loss": 0.7436, "step": 2195 }, { "epoch": 4.463414634146342, "grad_norm": 3.007041545969289, "learning_rate": 1.2206867332470091e-05, "loss": 0.7178, "step": 2196 }, { "epoch": 4.465447154471545, "grad_norm": 2.976626196927109, "learning_rate": 1.2200445782028795e-05, "loss": 0.5896, "step": 2197 }, { "epoch": 4.467479674796748, "grad_norm": 2.9192928789216683, "learning_rate": 1.2194023277892447e-05, "loss": 0.5022, "step": 2198 }, { "epoch": 4.469512195121951, "grad_norm": 2.922689999323099, "learning_rate": 1.2187599822844616e-05, "loss": 0.6876, "step": 2199 }, { "epoch": 4.471544715447155, "grad_norm": 2.8985039036869886, "learning_rate": 1.2181175419669293e-05, "loss": 0.6001, "step": 2200 }, { "epoch": 4.473577235772358, "grad_norm": 3.1455949038981403, "learning_rate": 1.2174750071150878e-05, "loss": 0.613, "step": 2201 }, { "epoch": 4.475609756097561, "grad_norm": 3.479851406816009, "learning_rate": 1.2168323780074184e-05, "loss": 0.6275, "step": 2202 }, { "epoch": 4.477642276422764, "grad_norm": 3.014972995204455, "learning_rate": 1.2161896549224425e-05, "loss": 0.7136, "step": 2203 }, { "epoch": 4.479674796747967, "grad_norm": 2.9634554122904566, "learning_rate": 1.215546838138723e-05, "loss": 0.724, "step": 2204 }, { "epoch": 4.4817073170731705, "grad_norm": 2.9492107344453182, "learning_rate": 1.2149039279348632e-05, "loss": 0.6343, "step": 2205 }, { "epoch": 4.483739837398374, "grad_norm": 2.808139762594353, "learning_rate": 1.2142609245895061e-05, "loss": 0.7152, "step": 2206 }, { "epoch": 4.485772357723577, "grad_norm": 2.8700589057195742, "learning_rate": 1.2136178283813365e-05, "loss": 0.5592, "step": 2207 }, { "epoch": 4.487804878048781, "grad_norm": 3.1824544554929965, "learning_rate": 1.212974639589078e-05, "loss": 0.7084, "step": 2208 }, { "epoch": 4.489837398373984, "grad_norm": 2.9113143241782726, "learning_rate": 1.2123313584914958e-05, "loss": 0.5949, "step": 2209 }, { "epoch": 4.491869918699187, "grad_norm": 3.1053606368948974, "learning_rate": 1.2116879853673938e-05, "loss": 0.8011, "step": 2210 }, { "epoch": 4.4939024390243905, "grad_norm": 3.3020995538693767, "learning_rate": 1.211044520495616e-05, "loss": 0.7948, "step": 2211 }, { "epoch": 4.495934959349594, "grad_norm": 2.9577723125547712, "learning_rate": 1.2104009641550472e-05, "loss": 0.5828, "step": 2212 }, { "epoch": 4.497967479674797, "grad_norm": 2.8466484582812464, "learning_rate": 1.2097573166246107e-05, "loss": 0.5974, "step": 2213 }, { "epoch": 4.5, "grad_norm": 3.0628889473086667, "learning_rate": 1.2091135781832697e-05, "loss": 0.8011, "step": 2214 }, { "epoch": 4.502032520325203, "grad_norm": 3.295718282001525, "learning_rate": 1.2084697491100268e-05, "loss": 0.6787, "step": 2215 }, { "epoch": 4.504065040650406, "grad_norm": 2.776196655947554, "learning_rate": 1.2078258296839245e-05, "loss": 0.5858, "step": 2216 }, { "epoch": 4.5060975609756095, "grad_norm": 2.604306179743555, "learning_rate": 1.207181820184043e-05, "loss": 0.4972, "step": 2217 }, { "epoch": 4.508130081300813, "grad_norm": 3.754625435483793, "learning_rate": 1.206537720889503e-05, "loss": 0.6323, "step": 2218 }, { "epoch": 4.510162601626016, "grad_norm": 3.14891039497955, "learning_rate": 1.2058935320794633e-05, "loss": 0.6952, "step": 2219 }, { "epoch": 4.512195121951219, "grad_norm": 3.1992023776053555, "learning_rate": 1.2052492540331218e-05, "loss": 0.6783, "step": 2220 }, { "epoch": 4.514227642276423, "grad_norm": 3.2693347070762657, "learning_rate": 1.2046048870297152e-05, "loss": 0.7158, "step": 2221 }, { "epoch": 4.516260162601626, "grad_norm": 2.5485300210537725, "learning_rate": 1.2039604313485183e-05, "loss": 0.5234, "step": 2222 }, { "epoch": 4.5182926829268295, "grad_norm": 3.3472929818233315, "learning_rate": 1.2033158872688446e-05, "loss": 0.7022, "step": 2223 }, { "epoch": 4.520325203252033, "grad_norm": 3.3197485200115864, "learning_rate": 1.2026712550700457e-05, "loss": 0.7629, "step": 2224 }, { "epoch": 4.522357723577236, "grad_norm": 3.0804045846427317, "learning_rate": 1.202026535031512e-05, "loss": 0.7156, "step": 2225 }, { "epoch": 4.524390243902439, "grad_norm": 3.367920678535504, "learning_rate": 1.2013817274326716e-05, "loss": 0.6688, "step": 2226 }, { "epoch": 4.526422764227642, "grad_norm": 3.429723926372333, "learning_rate": 1.2007368325529902e-05, "loss": 0.6641, "step": 2227 }, { "epoch": 4.528455284552845, "grad_norm": 3.419214101264249, "learning_rate": 1.200091850671972e-05, "loss": 0.6339, "step": 2228 }, { "epoch": 4.530487804878049, "grad_norm": 2.990193994939024, "learning_rate": 1.199446782069158e-05, "loss": 0.6271, "step": 2229 }, { "epoch": 4.532520325203252, "grad_norm": 2.6006234569416717, "learning_rate": 1.1988016270241278e-05, "loss": 0.5829, "step": 2230 }, { "epoch": 4.534552845528455, "grad_norm": 2.5542274663985074, "learning_rate": 1.1981563858164979e-05, "loss": 0.6175, "step": 2231 }, { "epoch": 4.536585365853659, "grad_norm": 2.7929345255138553, "learning_rate": 1.1975110587259222e-05, "loss": 0.664, "step": 2232 }, { "epoch": 4.538617886178862, "grad_norm": 3.077944427504673, "learning_rate": 1.1968656460320922e-05, "loss": 0.5811, "step": 2233 }, { "epoch": 4.540650406504065, "grad_norm": 2.5088793962362455, "learning_rate": 1.1962201480147355e-05, "loss": 0.5425, "step": 2234 }, { "epoch": 4.5426829268292686, "grad_norm": 2.706444618829862, "learning_rate": 1.1955745649536178e-05, "loss": 0.6558, "step": 2235 }, { "epoch": 4.544715447154472, "grad_norm": 3.273855420874858, "learning_rate": 1.1949288971285411e-05, "loss": 0.6785, "step": 2236 }, { "epoch": 4.546747967479675, "grad_norm": 2.9386743087860183, "learning_rate": 1.1942831448193444e-05, "loss": 0.5942, "step": 2237 }, { "epoch": 4.548780487804878, "grad_norm": 2.885561312539037, "learning_rate": 1.1936373083059032e-05, "loss": 0.734, "step": 2238 }, { "epoch": 4.550813008130081, "grad_norm": 2.9252073771626077, "learning_rate": 1.192991387868129e-05, "loss": 0.6495, "step": 2239 }, { "epoch": 4.5528455284552845, "grad_norm": 3.380247635002831, "learning_rate": 1.1923453837859706e-05, "loss": 0.7059, "step": 2240 }, { "epoch": 4.554878048780488, "grad_norm": 3.407941666416515, "learning_rate": 1.1916992963394123e-05, "loss": 0.7136, "step": 2241 }, { "epoch": 4.556910569105691, "grad_norm": 3.3953015102562625, "learning_rate": 1.191053125808475e-05, "loss": 0.6801, "step": 2242 }, { "epoch": 4.558943089430894, "grad_norm": 2.8013004761521123, "learning_rate": 1.1904068724732156e-05, "loss": 0.5734, "step": 2243 }, { "epoch": 4.560975609756097, "grad_norm": 2.8620476401175114, "learning_rate": 1.1897605366137264e-05, "loss": 0.6815, "step": 2244 }, { "epoch": 4.5630081300813, "grad_norm": 3.2541896462165494, "learning_rate": 1.1891141185101355e-05, "loss": 0.7196, "step": 2245 }, { "epoch": 4.565040650406504, "grad_norm": 3.190278175120848, "learning_rate": 1.1884676184426074e-05, "loss": 0.6177, "step": 2246 }, { "epoch": 4.567073170731708, "grad_norm": 2.9439001993137124, "learning_rate": 1.1878210366913419e-05, "loss": 0.7091, "step": 2247 }, { "epoch": 4.569105691056911, "grad_norm": 3.268585329151592, "learning_rate": 1.1871743735365735e-05, "loss": 0.6624, "step": 2248 }, { "epoch": 4.571138211382114, "grad_norm": 2.9882814050077173, "learning_rate": 1.1865276292585723e-05, "loss": 0.5832, "step": 2249 }, { "epoch": 4.573170731707317, "grad_norm": 3.0153058396212002, "learning_rate": 1.1858808041376439e-05, "loss": 0.6279, "step": 2250 }, { "epoch": 4.57520325203252, "grad_norm": 3.32571331785145, "learning_rate": 1.185233898454129e-05, "loss": 0.5572, "step": 2251 }, { "epoch": 4.5772357723577235, "grad_norm": 3.184060766708109, "learning_rate": 1.1845869124884027e-05, "loss": 0.6808, "step": 2252 }, { "epoch": 4.579268292682927, "grad_norm": 3.1392426577334116, "learning_rate": 1.1839398465208748e-05, "loss": 0.6388, "step": 2253 }, { "epoch": 4.58130081300813, "grad_norm": 3.4949335448159218, "learning_rate": 1.183292700831991e-05, "loss": 0.7887, "step": 2254 }, { "epoch": 4.583333333333333, "grad_norm": 3.2120726101189376, "learning_rate": 1.1826454757022297e-05, "loss": 0.7577, "step": 2255 }, { "epoch": 4.585365853658536, "grad_norm": 3.0161108951293314, "learning_rate": 1.1819981714121054e-05, "loss": 0.6653, "step": 2256 }, { "epoch": 4.58739837398374, "grad_norm": 3.4176847665094923, "learning_rate": 1.1813507882421657e-05, "loss": 0.5948, "step": 2257 }, { "epoch": 4.5894308943089435, "grad_norm": 3.518469690509393, "learning_rate": 1.1807033264729932e-05, "loss": 0.766, "step": 2258 }, { "epoch": 4.591463414634147, "grad_norm": 3.5233967420030847, "learning_rate": 1.1800557863852043e-05, "loss": 0.6742, "step": 2259 }, { "epoch": 4.59349593495935, "grad_norm": 3.0210409545631416, "learning_rate": 1.1794081682594491e-05, "loss": 0.7086, "step": 2260 }, { "epoch": 4.595528455284553, "grad_norm": 3.1353596777625485, "learning_rate": 1.1787604723764122e-05, "loss": 0.6403, "step": 2261 }, { "epoch": 4.597560975609756, "grad_norm": 2.632603066922333, "learning_rate": 1.1781126990168106e-05, "loss": 0.7259, "step": 2262 }, { "epoch": 4.599593495934959, "grad_norm": 2.893985218446589, "learning_rate": 1.1774648484613963e-05, "loss": 0.7175, "step": 2263 }, { "epoch": 4.6016260162601625, "grad_norm": 3.576648413336792, "learning_rate": 1.176816920990954e-05, "loss": 0.7375, "step": 2264 }, { "epoch": 4.603658536585366, "grad_norm": 2.8125478090230387, "learning_rate": 1.1761689168863022e-05, "loss": 0.6502, "step": 2265 }, { "epoch": 4.605691056910569, "grad_norm": 2.80072725945583, "learning_rate": 1.175520836428292e-05, "loss": 0.7196, "step": 2266 }, { "epoch": 4.607723577235772, "grad_norm": 2.9812355676613667, "learning_rate": 1.1748726798978076e-05, "loss": 0.7311, "step": 2267 }, { "epoch": 4.609756097560975, "grad_norm": 3.2033629957298055, "learning_rate": 1.174224447575767e-05, "loss": 0.6652, "step": 2268 }, { "epoch": 4.611788617886178, "grad_norm": 2.52848131147767, "learning_rate": 1.1735761397431205e-05, "loss": 0.5231, "step": 2269 }, { "epoch": 4.613821138211382, "grad_norm": 3.240467755420282, "learning_rate": 1.172927756680851e-05, "loss": 0.593, "step": 2270 }, { "epoch": 4.615853658536586, "grad_norm": 3.5985012461433055, "learning_rate": 1.1722792986699743e-05, "loss": 0.8391, "step": 2271 }, { "epoch": 4.617886178861789, "grad_norm": 3.2877125300712984, "learning_rate": 1.171630765991538e-05, "loss": 0.6657, "step": 2272 }, { "epoch": 4.619918699186992, "grad_norm": 2.3812722592980635, "learning_rate": 1.1709821589266236e-05, "loss": 0.534, "step": 2273 }, { "epoch": 4.621951219512195, "grad_norm": 3.2045142767677777, "learning_rate": 1.1703334777563426e-05, "loss": 0.5503, "step": 2274 }, { "epoch": 4.623983739837398, "grad_norm": 3.1632131757114914, "learning_rate": 1.1696847227618408e-05, "loss": 0.497, "step": 2275 }, { "epoch": 4.626016260162602, "grad_norm": 2.9209819673642277, "learning_rate": 1.169035894224295e-05, "loss": 0.5757, "step": 2276 }, { "epoch": 4.628048780487805, "grad_norm": 3.4731124141261884, "learning_rate": 1.1683869924249138e-05, "loss": 0.8661, "step": 2277 }, { "epoch": 4.630081300813008, "grad_norm": 2.989423369686288, "learning_rate": 1.1677380176449372e-05, "loss": 0.618, "step": 2278 }, { "epoch": 4.632113821138211, "grad_norm": 3.056312544367182, "learning_rate": 1.1670889701656378e-05, "loss": 0.7843, "step": 2279 }, { "epoch": 4.634146341463414, "grad_norm": 3.025382160409105, "learning_rate": 1.1664398502683194e-05, "loss": 0.6191, "step": 2280 }, { "epoch": 4.636178861788618, "grad_norm": 3.810266683364539, "learning_rate": 1.1657906582343166e-05, "loss": 0.7094, "step": 2281 }, { "epoch": 4.638211382113822, "grad_norm": 3.454393229892853, "learning_rate": 1.1651413943449958e-05, "loss": 0.5715, "step": 2282 }, { "epoch": 4.640243902439025, "grad_norm": 3.031161062772086, "learning_rate": 1.1644920588817546e-05, "loss": 0.6047, "step": 2283 }, { "epoch": 4.642276422764228, "grad_norm": 2.6495742166309957, "learning_rate": 1.1638426521260211e-05, "loss": 0.6348, "step": 2284 }, { "epoch": 4.644308943089431, "grad_norm": 3.2796512593814815, "learning_rate": 1.1631931743592548e-05, "loss": 0.6085, "step": 2285 }, { "epoch": 4.646341463414634, "grad_norm": 3.239747509965858, "learning_rate": 1.1625436258629461e-05, "loss": 0.6306, "step": 2286 }, { "epoch": 4.6483739837398375, "grad_norm": 2.9347705812466343, "learning_rate": 1.1618940069186155e-05, "loss": 0.4799, "step": 2287 }, { "epoch": 4.650406504065041, "grad_norm": 3.3780369129671226, "learning_rate": 1.1612443178078138e-05, "loss": 0.6602, "step": 2288 }, { "epoch": 4.652439024390244, "grad_norm": 3.2008197738848474, "learning_rate": 1.1605945588121238e-05, "loss": 0.6712, "step": 2289 }, { "epoch": 4.654471544715447, "grad_norm": 3.2374333668300346, "learning_rate": 1.1599447302131568e-05, "loss": 0.641, "step": 2290 }, { "epoch": 4.65650406504065, "grad_norm": 3.6726598362217677, "learning_rate": 1.159294832292555e-05, "loss": 0.7398, "step": 2291 }, { "epoch": 4.658536585365853, "grad_norm": 3.224472013594856, "learning_rate": 1.1586448653319908e-05, "loss": 0.6871, "step": 2292 }, { "epoch": 4.6605691056910565, "grad_norm": 3.292939847388614, "learning_rate": 1.1579948296131662e-05, "loss": 0.5114, "step": 2293 }, { "epoch": 4.66260162601626, "grad_norm": 3.637342294824386, "learning_rate": 1.1573447254178132e-05, "loss": 0.7232, "step": 2294 }, { "epoch": 4.664634146341464, "grad_norm": 3.3869034620438, "learning_rate": 1.1566945530276934e-05, "loss": 0.7157, "step": 2295 }, { "epoch": 4.666666666666667, "grad_norm": 2.7728161355851433, "learning_rate": 1.156044312724598e-05, "loss": 0.5829, "step": 2296 }, { "epoch": 4.66869918699187, "grad_norm": 2.6246738305365813, "learning_rate": 1.1553940047903476e-05, "loss": 0.6291, "step": 2297 }, { "epoch": 4.670731707317073, "grad_norm": 3.028680297397905, "learning_rate": 1.1547436295067923e-05, "loss": 0.6497, "step": 2298 }, { "epoch": 4.6727642276422765, "grad_norm": 3.578605206184993, "learning_rate": 1.1540931871558105e-05, "loss": 0.768, "step": 2299 }, { "epoch": 4.67479674796748, "grad_norm": 3.7228112628637637, "learning_rate": 1.153442678019311e-05, "loss": 0.6187, "step": 2300 }, { "epoch": 4.676829268292683, "grad_norm": 3.2604486005472686, "learning_rate": 1.152792102379231e-05, "loss": 0.6976, "step": 2301 }, { "epoch": 4.678861788617886, "grad_norm": 3.4719632419538806, "learning_rate": 1.152141460517536e-05, "loss": 0.6947, "step": 2302 }, { "epoch": 4.680894308943089, "grad_norm": 3.3592955835252765, "learning_rate": 1.1514907527162217e-05, "loss": 0.6855, "step": 2303 }, { "epoch": 4.682926829268292, "grad_norm": 3.373221354137555, "learning_rate": 1.1508399792573095e-05, "loss": 0.6543, "step": 2304 }, { "epoch": 4.684959349593496, "grad_norm": 2.9502727898615766, "learning_rate": 1.1501891404228525e-05, "loss": 0.5457, "step": 2305 }, { "epoch": 4.6869918699187, "grad_norm": 3.1964550155481684, "learning_rate": 1.1495382364949303e-05, "loss": 0.6838, "step": 2306 }, { "epoch": 4.689024390243903, "grad_norm": 3.2788075827905914, "learning_rate": 1.1488872677556511e-05, "loss": 0.5193, "step": 2307 }, { "epoch": 4.691056910569106, "grad_norm": 3.3340629464793277, "learning_rate": 1.1482362344871514e-05, "loss": 0.8757, "step": 2308 }, { "epoch": 4.693089430894309, "grad_norm": 2.9616866340876657, "learning_rate": 1.1475851369715954e-05, "loss": 0.6885, "step": 2309 }, { "epoch": 4.695121951219512, "grad_norm": 2.8262946937744666, "learning_rate": 1.146933975491175e-05, "loss": 0.593, "step": 2310 }, { "epoch": 4.6971544715447155, "grad_norm": 2.712886377858222, "learning_rate": 1.1462827503281099e-05, "loss": 0.5652, "step": 2311 }, { "epoch": 4.699186991869919, "grad_norm": 3.409042390068692, "learning_rate": 1.1456314617646482e-05, "loss": 0.646, "step": 2312 }, { "epoch": 4.701219512195122, "grad_norm": 2.8640763657226307, "learning_rate": 1.1449801100830646e-05, "loss": 0.7424, "step": 2313 }, { "epoch": 4.703252032520325, "grad_norm": 3.226582945240935, "learning_rate": 1.1443286955656614e-05, "loss": 0.6455, "step": 2314 }, { "epoch": 4.705284552845528, "grad_norm": 3.0076159541686307, "learning_rate": 1.143677218494768e-05, "loss": 0.7681, "step": 2315 }, { "epoch": 4.7073170731707314, "grad_norm": 3.094853715304811, "learning_rate": 1.1430256791527406e-05, "loss": 0.575, "step": 2316 }, { "epoch": 4.709349593495935, "grad_norm": 3.0508421860169865, "learning_rate": 1.1423740778219636e-05, "loss": 0.7007, "step": 2317 }, { "epoch": 4.711382113821138, "grad_norm": 3.4794870273139415, "learning_rate": 1.1417224147848471e-05, "loss": 0.6442, "step": 2318 }, { "epoch": 4.713414634146341, "grad_norm": 3.012720398510299, "learning_rate": 1.1410706903238285e-05, "loss": 0.5984, "step": 2319 }, { "epoch": 4.715447154471545, "grad_norm": 2.5988099157367435, "learning_rate": 1.1404189047213716e-05, "loss": 0.5593, "step": 2320 }, { "epoch": 4.717479674796748, "grad_norm": 3.546161406089007, "learning_rate": 1.1397670582599665e-05, "loss": 0.6703, "step": 2321 }, { "epoch": 4.719512195121951, "grad_norm": 3.2037031919407637, "learning_rate": 1.1391151512221304e-05, "loss": 0.7368, "step": 2322 }, { "epoch": 4.721544715447155, "grad_norm": 3.1047520135333837, "learning_rate": 1.1384631838904056e-05, "loss": 0.6559, "step": 2323 }, { "epoch": 4.723577235772358, "grad_norm": 2.8022769189626433, "learning_rate": 1.137811156547362e-05, "loss": 0.6017, "step": 2324 }, { "epoch": 4.725609756097561, "grad_norm": 3.3412193551928415, "learning_rate": 1.1371590694755943e-05, "loss": 0.7177, "step": 2325 }, { "epoch": 4.727642276422764, "grad_norm": 2.6186347889642807, "learning_rate": 1.1365069229577234e-05, "loss": 0.6187, "step": 2326 }, { "epoch": 4.729674796747967, "grad_norm": 3.380688388664635, "learning_rate": 1.1358547172763965e-05, "loss": 0.7361, "step": 2327 }, { "epoch": 4.7317073170731705, "grad_norm": 2.864317609212535, "learning_rate": 1.1352024527142855e-05, "loss": 0.6515, "step": 2328 }, { "epoch": 4.733739837398374, "grad_norm": 3.2025655353804257, "learning_rate": 1.1345501295540887e-05, "loss": 0.6241, "step": 2329 }, { "epoch": 4.735772357723577, "grad_norm": 2.829080244789038, "learning_rate": 1.1338977480785298e-05, "loss": 0.6053, "step": 2330 }, { "epoch": 4.737804878048781, "grad_norm": 3.518003051540819, "learning_rate": 1.1332453085703568e-05, "loss": 0.6663, "step": 2331 }, { "epoch": 4.739837398373984, "grad_norm": 3.136216920735255, "learning_rate": 1.1325928113123431e-05, "loss": 0.5556, "step": 2332 }, { "epoch": 4.741869918699187, "grad_norm": 2.886600601271744, "learning_rate": 1.1319402565872884e-05, "loss": 0.5996, "step": 2333 }, { "epoch": 4.7439024390243905, "grad_norm": 3.4235520560136035, "learning_rate": 1.1312876446780158e-05, "loss": 0.6632, "step": 2334 }, { "epoch": 4.745934959349594, "grad_norm": 3.0123279098116913, "learning_rate": 1.1306349758673744e-05, "loss": 0.5685, "step": 2335 }, { "epoch": 4.747967479674797, "grad_norm": 3.2395748525525643, "learning_rate": 1.129982250438237e-05, "loss": 0.7397, "step": 2336 }, { "epoch": 4.75, "grad_norm": 2.915326705563588, "learning_rate": 1.1293294686735006e-05, "loss": 0.7008, "step": 2337 }, { "epoch": 4.752032520325203, "grad_norm": 3.0575661733597563, "learning_rate": 1.1286766308560884e-05, "loss": 0.6274, "step": 2338 }, { "epoch": 4.754065040650406, "grad_norm": 2.755015093880398, "learning_rate": 1.1280237372689462e-05, "loss": 0.7095, "step": 2339 }, { "epoch": 4.7560975609756095, "grad_norm": 3.184731179980934, "learning_rate": 1.1273707881950445e-05, "loss": 0.7049, "step": 2340 }, { "epoch": 4.758130081300813, "grad_norm": 2.995064939303926, "learning_rate": 1.1267177839173786e-05, "loss": 0.5418, "step": 2341 }, { "epoch": 4.760162601626016, "grad_norm": 2.662635997869429, "learning_rate": 1.1260647247189663e-05, "loss": 0.6709, "step": 2342 }, { "epoch": 4.762195121951219, "grad_norm": 2.946215730448694, "learning_rate": 1.12541161088285e-05, "loss": 0.9019, "step": 2343 }, { "epoch": 4.764227642276423, "grad_norm": 3.233275336124828, "learning_rate": 1.1247584426920962e-05, "loss": 0.5525, "step": 2344 }, { "epoch": 4.766260162601626, "grad_norm": 3.041248016808087, "learning_rate": 1.124105220429794e-05, "loss": 0.7961, "step": 2345 }, { "epoch": 4.7682926829268295, "grad_norm": 3.3900861280669794, "learning_rate": 1.1234519443790566e-05, "loss": 0.493, "step": 2346 }, { "epoch": 4.770325203252033, "grad_norm": 2.881623669662241, "learning_rate": 1.1227986148230205e-05, "loss": 0.706, "step": 2347 }, { "epoch": 4.772357723577236, "grad_norm": 3.770642151502543, "learning_rate": 1.1221452320448449e-05, "loss": 0.7342, "step": 2348 }, { "epoch": 4.774390243902439, "grad_norm": 3.136022467278505, "learning_rate": 1.1214917963277125e-05, "loss": 0.5932, "step": 2349 }, { "epoch": 4.776422764227642, "grad_norm": 3.1176399401585444, "learning_rate": 1.1208383079548292e-05, "loss": 0.6011, "step": 2350 }, { "epoch": 4.778455284552845, "grad_norm": 2.5906612212354654, "learning_rate": 1.1201847672094227e-05, "loss": 0.5253, "step": 2351 }, { "epoch": 4.780487804878049, "grad_norm": 3.4053340936463363, "learning_rate": 1.1195311743747445e-05, "loss": 0.7138, "step": 2352 }, { "epoch": 4.782520325203252, "grad_norm": 4.092218441428071, "learning_rate": 1.1188775297340681e-05, "loss": 0.6661, "step": 2353 }, { "epoch": 4.784552845528455, "grad_norm": 2.7806903066455737, "learning_rate": 1.1182238335706896e-05, "loss": 0.5796, "step": 2354 }, { "epoch": 4.786585365853659, "grad_norm": 2.99898519233595, "learning_rate": 1.1175700861679276e-05, "loss": 0.6531, "step": 2355 }, { "epoch": 4.788617886178862, "grad_norm": 3.1264126423106853, "learning_rate": 1.116916287809122e-05, "loss": 0.6139, "step": 2356 }, { "epoch": 4.790650406504065, "grad_norm": 3.808695836581929, "learning_rate": 1.1162624387776363e-05, "loss": 0.732, "step": 2357 }, { "epoch": 4.7926829268292686, "grad_norm": 2.9537473921293103, "learning_rate": 1.115608539356855e-05, "loss": 0.6696, "step": 2358 }, { "epoch": 4.794715447154472, "grad_norm": 3.0858991632224053, "learning_rate": 1.1149545898301841e-05, "loss": 0.5764, "step": 2359 }, { "epoch": 4.796747967479675, "grad_norm": 3.0373573625837422, "learning_rate": 1.1143005904810527e-05, "loss": 0.6049, "step": 2360 }, { "epoch": 4.798780487804878, "grad_norm": 2.845281482901104, "learning_rate": 1.1136465415929096e-05, "loss": 0.5911, "step": 2361 }, { "epoch": 4.800813008130081, "grad_norm": 3.381118654967506, "learning_rate": 1.112992443449227e-05, "loss": 0.697, "step": 2362 }, { "epoch": 4.8028455284552845, "grad_norm": 3.5230839043618136, "learning_rate": 1.112338296333497e-05, "loss": 0.727, "step": 2363 }, { "epoch": 4.804878048780488, "grad_norm": 3.186199364081253, "learning_rate": 1.1116841005292339e-05, "loss": 0.6661, "step": 2364 }, { "epoch": 4.806910569105691, "grad_norm": 2.8277135834890745, "learning_rate": 1.1110298563199717e-05, "loss": 0.5792, "step": 2365 }, { "epoch": 4.808943089430894, "grad_norm": 3.228905178006054, "learning_rate": 1.1103755639892672e-05, "loss": 0.8373, "step": 2366 }, { "epoch": 4.810975609756097, "grad_norm": 3.0111454761147476, "learning_rate": 1.1097212238206972e-05, "loss": 0.6263, "step": 2367 }, { "epoch": 4.8130081300813, "grad_norm": 3.5435480235127756, "learning_rate": 1.1090668360978589e-05, "loss": 0.7377, "step": 2368 }, { "epoch": 4.815040650406504, "grad_norm": 2.708885482078627, "learning_rate": 1.1084124011043707e-05, "loss": 0.5866, "step": 2369 }, { "epoch": 4.817073170731708, "grad_norm": 3.3038070615301907, "learning_rate": 1.1077579191238707e-05, "loss": 0.6098, "step": 2370 }, { "epoch": 4.819105691056911, "grad_norm": 3.381133992680285, "learning_rate": 1.1071033904400186e-05, "loss": 0.7049, "step": 2371 }, { "epoch": 4.821138211382114, "grad_norm": 3.298142405248073, "learning_rate": 1.106448815336493e-05, "loss": 0.677, "step": 2372 }, { "epoch": 4.823170731707317, "grad_norm": 3.09327243594827, "learning_rate": 1.1057941940969938e-05, "loss": 0.6447, "step": 2373 }, { "epoch": 4.82520325203252, "grad_norm": 3.1062507047037027, "learning_rate": 1.1051395270052401e-05, "loss": 0.5734, "step": 2374 }, { "epoch": 4.8272357723577235, "grad_norm": 3.7157582625391425, "learning_rate": 1.104484814344971e-05, "loss": 0.7276, "step": 2375 }, { "epoch": 4.829268292682927, "grad_norm": 3.4525608625085185, "learning_rate": 1.1038300563999455e-05, "loss": 0.7276, "step": 2376 }, { "epoch": 4.83130081300813, "grad_norm": 3.5278531056113573, "learning_rate": 1.1031752534539424e-05, "loss": 0.7383, "step": 2377 }, { "epoch": 4.833333333333333, "grad_norm": 3.177543053953585, "learning_rate": 1.1025204057907597e-05, "loss": 0.7751, "step": 2378 }, { "epoch": 4.835365853658536, "grad_norm": 3.367628723120567, "learning_rate": 1.101865513694215e-05, "loss": 0.7787, "step": 2379 }, { "epoch": 4.83739837398374, "grad_norm": 3.383183628453972, "learning_rate": 1.1012105774481446e-05, "loss": 0.5484, "step": 2380 }, { "epoch": 4.8394308943089435, "grad_norm": 3.129980864010534, "learning_rate": 1.1005555973364048e-05, "loss": 0.6993, "step": 2381 }, { "epoch": 4.841463414634147, "grad_norm": 3.5220013551338787, "learning_rate": 1.0999005736428703e-05, "loss": 0.7467, "step": 2382 }, { "epoch": 4.84349593495935, "grad_norm": 3.318953318703048, "learning_rate": 1.0992455066514352e-05, "loss": 0.7607, "step": 2383 }, { "epoch": 4.845528455284553, "grad_norm": 3.058061204440119, "learning_rate": 1.0985903966460115e-05, "loss": 0.6876, "step": 2384 }, { "epoch": 4.847560975609756, "grad_norm": 2.617594358913442, "learning_rate": 1.0979352439105311e-05, "loss": 0.5269, "step": 2385 }, { "epoch": 4.849593495934959, "grad_norm": 2.495340945587686, "learning_rate": 1.0972800487289426e-05, "loss": 0.5025, "step": 2386 }, { "epoch": 4.8516260162601625, "grad_norm": 3.2546109820473887, "learning_rate": 1.096624811385215e-05, "loss": 0.6783, "step": 2387 }, { "epoch": 4.853658536585366, "grad_norm": 3.1646944958440857, "learning_rate": 1.0959695321633346e-05, "loss": 0.5889, "step": 2388 }, { "epoch": 4.855691056910569, "grad_norm": 3.21726902782112, "learning_rate": 1.0953142113473057e-05, "loss": 0.6833, "step": 2389 }, { "epoch": 4.857723577235772, "grad_norm": 3.516881401805942, "learning_rate": 1.0946588492211508e-05, "loss": 0.6298, "step": 2390 }, { "epoch": 4.859756097560975, "grad_norm": 3.237989828841033, "learning_rate": 1.0940034460689106e-05, "loss": 0.6419, "step": 2391 }, { "epoch": 4.861788617886178, "grad_norm": 2.767876163964359, "learning_rate": 1.0933480021746432e-05, "loss": 0.6452, "step": 2392 }, { "epoch": 4.863821138211382, "grad_norm": 3.1317863995737216, "learning_rate": 1.0926925178224244e-05, "loss": 0.7787, "step": 2393 }, { "epoch": 4.865853658536586, "grad_norm": 2.9544430689591565, "learning_rate": 1.0920369932963477e-05, "loss": 0.5212, "step": 2394 }, { "epoch": 4.867886178861789, "grad_norm": 3.5432564727857674, "learning_rate": 1.091381428880524e-05, "loss": 0.666, "step": 2395 }, { "epoch": 4.869918699186992, "grad_norm": 4.08032859195243, "learning_rate": 1.0907258248590816e-05, "loss": 0.7006, "step": 2396 }, { "epoch": 4.871951219512195, "grad_norm": 3.2519452619947313, "learning_rate": 1.0900701815161654e-05, "loss": 0.6181, "step": 2397 }, { "epoch": 4.873983739837398, "grad_norm": 3.5479622971610074, "learning_rate": 1.0894144991359379e-05, "loss": 0.7982, "step": 2398 }, { "epoch": 4.876016260162602, "grad_norm": 3.4411678592932966, "learning_rate": 1.0887587780025784e-05, "loss": 0.7683, "step": 2399 }, { "epoch": 4.878048780487805, "grad_norm": 2.9137159025700052, "learning_rate": 1.0881030184002827e-05, "loss": 0.7501, "step": 2400 }, { "epoch": 4.880081300813008, "grad_norm": 3.577204023071614, "learning_rate": 1.0874472206132643e-05, "loss": 0.7621, "step": 2401 }, { "epoch": 4.882113821138211, "grad_norm": 3.087744841440461, "learning_rate": 1.0867913849257517e-05, "loss": 0.7127, "step": 2402 }, { "epoch": 4.884146341463414, "grad_norm": 2.7416950807701532, "learning_rate": 1.0861355116219904e-05, "loss": 0.5253, "step": 2403 }, { "epoch": 4.886178861788618, "grad_norm": 2.9496865391460925, "learning_rate": 1.0854796009862434e-05, "loss": 0.5661, "step": 2404 }, { "epoch": 4.888211382113822, "grad_norm": 3.0557180012121705, "learning_rate": 1.084823653302788e-05, "loss": 0.8033, "step": 2405 }, { "epoch": 4.890243902439025, "grad_norm": 3.3279382828962656, "learning_rate": 1.0841676688559191e-05, "loss": 0.6832, "step": 2406 }, { "epoch": 4.892276422764228, "grad_norm": 2.876992014346207, "learning_rate": 1.0835116479299464e-05, "loss": 0.6215, "step": 2407 }, { "epoch": 4.894308943089431, "grad_norm": 2.8020525394955085, "learning_rate": 1.0828555908091958e-05, "loss": 0.6242, "step": 2408 }, { "epoch": 4.896341463414634, "grad_norm": 2.948638586422426, "learning_rate": 1.0821994977780093e-05, "loss": 0.6085, "step": 2409 }, { "epoch": 4.8983739837398375, "grad_norm": 2.8517787124474094, "learning_rate": 1.081543369120744e-05, "loss": 0.6817, "step": 2410 }, { "epoch": 4.900406504065041, "grad_norm": 3.2974878823165112, "learning_rate": 1.0808872051217727e-05, "loss": 0.6994, "step": 2411 }, { "epoch": 4.902439024390244, "grad_norm": 3.161305584877977, "learning_rate": 1.0802310060654832e-05, "loss": 0.5495, "step": 2412 }, { "epoch": 4.904471544715447, "grad_norm": 3.076378926265912, "learning_rate": 1.0795747722362788e-05, "loss": 0.7159, "step": 2413 }, { "epoch": 4.90650406504065, "grad_norm": 3.302275705860661, "learning_rate": 1.078918503918578e-05, "loss": 0.7439, "step": 2414 }, { "epoch": 4.908536585365853, "grad_norm": 3.4099829786771285, "learning_rate": 1.0782622013968133e-05, "loss": 0.5074, "step": 2415 }, { "epoch": 4.9105691056910565, "grad_norm": 3.087914113840506, "learning_rate": 1.0776058649554336e-05, "loss": 0.4813, "step": 2416 }, { "epoch": 4.91260162601626, "grad_norm": 3.054303086294716, "learning_rate": 1.0769494948789012e-05, "loss": 0.6002, "step": 2417 }, { "epoch": 4.914634146341464, "grad_norm": 3.3534384991216917, "learning_rate": 1.0762930914516933e-05, "loss": 0.7108, "step": 2418 }, { "epoch": 4.916666666666667, "grad_norm": 2.7911683906096774, "learning_rate": 1.0756366549583018e-05, "loss": 0.7311, "step": 2419 }, { "epoch": 4.91869918699187, "grad_norm": 3.1430129205474833, "learning_rate": 1.0749801856832325e-05, "loss": 0.671, "step": 2420 }, { "epoch": 4.920731707317073, "grad_norm": 3.7389225142814806, "learning_rate": 1.0743236839110065e-05, "loss": 0.6373, "step": 2421 }, { "epoch": 4.9227642276422765, "grad_norm": 2.8502234448500317, "learning_rate": 1.0736671499261576e-05, "loss": 0.5436, "step": 2422 }, { "epoch": 4.92479674796748, "grad_norm": 3.4678304294902675, "learning_rate": 1.073010584013234e-05, "loss": 0.7868, "step": 2423 }, { "epoch": 4.926829268292683, "grad_norm": 3.135033934710596, "learning_rate": 1.0723539864567983e-05, "loss": 0.7259, "step": 2424 }, { "epoch": 4.928861788617886, "grad_norm": 3.275774580547805, "learning_rate": 1.0716973575414262e-05, "loss": 0.6955, "step": 2425 }, { "epoch": 4.930894308943089, "grad_norm": 3.091293048375871, "learning_rate": 1.0710406975517069e-05, "loss": 0.7283, "step": 2426 }, { "epoch": 4.932926829268292, "grad_norm": 3.3225568058841555, "learning_rate": 1.0703840067722439e-05, "loss": 0.6051, "step": 2427 }, { "epoch": 4.934959349593496, "grad_norm": 3.345181891370543, "learning_rate": 1.0697272854876537e-05, "loss": 0.6834, "step": 2428 }, { "epoch": 4.9369918699187, "grad_norm": 3.312384641605906, "learning_rate": 1.0690705339825649e-05, "loss": 0.7128, "step": 2429 }, { "epoch": 4.939024390243903, "grad_norm": 3.870529877523309, "learning_rate": 1.0684137525416209e-05, "loss": 0.6198, "step": 2430 }, { "epoch": 4.941056910569106, "grad_norm": 3.047330500569165, "learning_rate": 1.067756941449477e-05, "loss": 0.6775, "step": 2431 }, { "epoch": 4.943089430894309, "grad_norm": 3.1147634055916336, "learning_rate": 1.0671001009908015e-05, "loss": 0.6588, "step": 2432 }, { "epoch": 4.945121951219512, "grad_norm": 2.772170019864765, "learning_rate": 1.0664432314502757e-05, "loss": 0.5882, "step": 2433 }, { "epoch": 4.9471544715447155, "grad_norm": 2.962794050505434, "learning_rate": 1.0657863331125937e-05, "loss": 0.5351, "step": 2434 }, { "epoch": 4.949186991869919, "grad_norm": 3.380624602812432, "learning_rate": 1.0651294062624615e-05, "loss": 0.7105, "step": 2435 }, { "epoch": 4.951219512195122, "grad_norm": 3.1290441535395894, "learning_rate": 1.0644724511845976e-05, "loss": 0.7062, "step": 2436 }, { "epoch": 4.953252032520325, "grad_norm": 3.3463052119796934, "learning_rate": 1.0638154681637333e-05, "loss": 0.591, "step": 2437 }, { "epoch": 4.955284552845528, "grad_norm": 2.8674787261108454, "learning_rate": 1.063158457484611e-05, "loss": 0.5721, "step": 2438 }, { "epoch": 4.9573170731707314, "grad_norm": 3.2374896065300787, "learning_rate": 1.062501419431986e-05, "loss": 0.6258, "step": 2439 }, { "epoch": 4.959349593495935, "grad_norm": 3.9114855347417166, "learning_rate": 1.0618443542906251e-05, "loss": 0.6203, "step": 2440 }, { "epoch": 4.961382113821138, "grad_norm": 3.518521705938725, "learning_rate": 1.0611872623453066e-05, "loss": 0.6867, "step": 2441 }, { "epoch": 4.963414634146341, "grad_norm": 3.0316219226875063, "learning_rate": 1.0605301438808211e-05, "loss": 0.6512, "step": 2442 }, { "epoch": 4.965447154471545, "grad_norm": 3.1873709705624598, "learning_rate": 1.0598729991819696e-05, "loss": 0.7092, "step": 2443 }, { "epoch": 4.967479674796748, "grad_norm": 3.375217199504291, "learning_rate": 1.059215828533566e-05, "loss": 0.7816, "step": 2444 }, { "epoch": 4.969512195121951, "grad_norm": 3.5339039183989565, "learning_rate": 1.0585586322204343e-05, "loss": 0.6913, "step": 2445 }, { "epoch": 4.971544715447155, "grad_norm": 3.0485968751979167, "learning_rate": 1.0579014105274094e-05, "loss": 0.6893, "step": 2446 }, { "epoch": 4.973577235772358, "grad_norm": 3.6807218226664817, "learning_rate": 1.057244163739338e-05, "loss": 0.6752, "step": 2447 }, { "epoch": 4.975609756097561, "grad_norm": 3.6077873677560914, "learning_rate": 1.0565868921410776e-05, "loss": 0.7857, "step": 2448 }, { "epoch": 4.977642276422764, "grad_norm": 3.0242706003622635, "learning_rate": 1.0559295960174962e-05, "loss": 0.7559, "step": 2449 }, { "epoch": 4.979674796747967, "grad_norm": 3.3073876439430645, "learning_rate": 1.0552722756534726e-05, "loss": 0.7188, "step": 2450 }, { "epoch": 4.9817073170731705, "grad_norm": 2.946828048157694, "learning_rate": 1.0546149313338956e-05, "loss": 0.5823, "step": 2451 }, { "epoch": 4.983739837398374, "grad_norm": 3.2892121696161802, "learning_rate": 1.0539575633436645e-05, "loss": 0.5825, "step": 2452 }, { "epoch": 4.985772357723577, "grad_norm": 2.814621275090556, "learning_rate": 1.05330017196769e-05, "loss": 0.6139, "step": 2453 }, { "epoch": 4.987804878048781, "grad_norm": 3.37182026790423, "learning_rate": 1.0526427574908912e-05, "loss": 0.6714, "step": 2454 }, { "epoch": 4.989837398373984, "grad_norm": 3.432634849533678, "learning_rate": 1.0519853201981986e-05, "loss": 0.5784, "step": 2455 }, { "epoch": 4.991869918699187, "grad_norm": 3.0820588865291776, "learning_rate": 1.0513278603745523e-05, "loss": 0.6245, "step": 2456 }, { "epoch": 4.9939024390243905, "grad_norm": 3.3279343912160813, "learning_rate": 1.0506703783049009e-05, "loss": 0.8459, "step": 2457 }, { "epoch": 4.995934959349594, "grad_norm": 3.39983528116009, "learning_rate": 1.0500128742742046e-05, "loss": 0.6619, "step": 2458 }, { "epoch": 4.997967479674797, "grad_norm": 3.4851762032715934, "learning_rate": 1.0493553485674315e-05, "loss": 0.6582, "step": 2459 }, { "epoch": 5.0, "grad_norm": 3.537821436609617, "learning_rate": 1.0486978014695606e-05, "loss": 0.615, "step": 2460 }, { "epoch": 5.0, "eval_loss": 0.943859875202179, "eval_runtime": 132.1277, "eval_samples_per_second": 7.447, "eval_steps_per_second": 0.931, "step": 2460 }, { "epoch": 5.002032520325203, "grad_norm": 2.636397466185468, "learning_rate": 1.0480402332655786e-05, "loss": 0.6135, "step": 2461 }, { "epoch": 5.004065040650406, "grad_norm": 3.1728687565109364, "learning_rate": 1.0473826442404821e-05, "loss": 0.4207, "step": 2462 }, { "epoch": 5.0060975609756095, "grad_norm": 3.2092484538763553, "learning_rate": 1.0467250346792775e-05, "loss": 0.5124, "step": 2463 }, { "epoch": 5.008130081300813, "grad_norm": 3.1371627468159473, "learning_rate": 1.0460674048669783e-05, "loss": 0.535, "step": 2464 }, { "epoch": 5.010162601626016, "grad_norm": 3.0388311968326596, "learning_rate": 1.0454097550886087e-05, "loss": 0.5494, "step": 2465 }, { "epoch": 5.012195121951219, "grad_norm": 2.710119844030016, "learning_rate": 1.0447520856291999e-05, "loss": 0.6556, "step": 2466 }, { "epoch": 5.014227642276423, "grad_norm": 3.242907474712844, "learning_rate": 1.0440943967737929e-05, "loss": 0.5375, "step": 2467 }, { "epoch": 5.016260162601626, "grad_norm": 3.183947784006189, "learning_rate": 1.0434366888074363e-05, "loss": 0.5499, "step": 2468 }, { "epoch": 5.0182926829268295, "grad_norm": 3.502884440645061, "learning_rate": 1.0427789620151871e-05, "loss": 0.6187, "step": 2469 }, { "epoch": 5.020325203252033, "grad_norm": 3.80166228494771, "learning_rate": 1.0421212166821113e-05, "loss": 0.5872, "step": 2470 }, { "epoch": 5.022357723577236, "grad_norm": 3.2796209940230776, "learning_rate": 1.0414634530932814e-05, "loss": 0.8112, "step": 2471 }, { "epoch": 5.024390243902439, "grad_norm": 3.757316122943901, "learning_rate": 1.0408056715337797e-05, "loss": 0.6461, "step": 2472 }, { "epoch": 5.026422764227642, "grad_norm": 3.7177694826351337, "learning_rate": 1.0401478722886942e-05, "loss": 0.5552, "step": 2473 }, { "epoch": 5.028455284552845, "grad_norm": 3.612432647294577, "learning_rate": 1.039490055643122e-05, "loss": 0.6448, "step": 2474 }, { "epoch": 5.030487804878049, "grad_norm": 3.694981439101527, "learning_rate": 1.0388322218821675e-05, "loss": 0.6524, "step": 2475 }, { "epoch": 5.032520325203252, "grad_norm": 3.4541185689283953, "learning_rate": 1.0381743712909424e-05, "loss": 0.593, "step": 2476 }, { "epoch": 5.034552845528455, "grad_norm": 3.854886706865119, "learning_rate": 1.0375165041545658e-05, "loss": 0.4922, "step": 2477 }, { "epoch": 5.036585365853658, "grad_norm": 3.002822576303026, "learning_rate": 1.0368586207581637e-05, "loss": 0.4729, "step": 2478 }, { "epoch": 5.038617886178862, "grad_norm": 3.0610704944451514, "learning_rate": 1.0362007213868691e-05, "loss": 0.6434, "step": 2479 }, { "epoch": 5.040650406504065, "grad_norm": 3.603139341831079, "learning_rate": 1.0355428063258224e-05, "loss": 0.6157, "step": 2480 }, { "epoch": 5.0426829268292686, "grad_norm": 3.113650240712849, "learning_rate": 1.0348848758601703e-05, "loss": 0.6348, "step": 2481 }, { "epoch": 5.044715447154472, "grad_norm": 3.63013615228171, "learning_rate": 1.0342269302750664e-05, "loss": 0.5816, "step": 2482 }, { "epoch": 5.046747967479675, "grad_norm": 3.0770270202172454, "learning_rate": 1.0335689698556713e-05, "loss": 0.3905, "step": 2483 }, { "epoch": 5.048780487804878, "grad_norm": 4.170885828785549, "learning_rate": 1.0329109948871512e-05, "loss": 0.6394, "step": 2484 }, { "epoch": 5.050813008130081, "grad_norm": 3.5575534317365194, "learning_rate": 1.0322530056546786e-05, "loss": 0.5616, "step": 2485 }, { "epoch": 5.0528455284552845, "grad_norm": 3.846576763978446, "learning_rate": 1.0315950024434336e-05, "loss": 0.5348, "step": 2486 }, { "epoch": 5.054878048780488, "grad_norm": 3.591749145220846, "learning_rate": 1.0309369855386002e-05, "loss": 0.4674, "step": 2487 }, { "epoch": 5.056910569105691, "grad_norm": 3.9827598292530615, "learning_rate": 1.0302789552253702e-05, "loss": 0.598, "step": 2488 }, { "epoch": 5.058943089430894, "grad_norm": 3.566877879882359, "learning_rate": 1.0296209117889405e-05, "loss": 0.5594, "step": 2489 }, { "epoch": 5.060975609756097, "grad_norm": 3.1765238635541246, "learning_rate": 1.028962855514513e-05, "loss": 0.5744, "step": 2490 }, { "epoch": 5.063008130081301, "grad_norm": 3.127929428335895, "learning_rate": 1.0283047866872966e-05, "loss": 0.552, "step": 2491 }, { "epoch": 5.065040650406504, "grad_norm": 3.7275335851213636, "learning_rate": 1.0276467055925044e-05, "loss": 0.5674, "step": 2492 }, { "epoch": 5.067073170731708, "grad_norm": 4.1672140992659585, "learning_rate": 1.0269886125153554e-05, "loss": 0.6447, "step": 2493 }, { "epoch": 5.069105691056911, "grad_norm": 3.7272396801685286, "learning_rate": 1.0263305077410737e-05, "loss": 0.4845, "step": 2494 }, { "epoch": 5.071138211382114, "grad_norm": 3.6347786404368785, "learning_rate": 1.0256723915548882e-05, "loss": 0.4742, "step": 2495 }, { "epoch": 5.073170731707317, "grad_norm": 3.2807642364631553, "learning_rate": 1.0250142642420335e-05, "loss": 0.489, "step": 2496 }, { "epoch": 5.07520325203252, "grad_norm": 3.1614097705319573, "learning_rate": 1.0243561260877479e-05, "loss": 0.5099, "step": 2497 }, { "epoch": 5.0772357723577235, "grad_norm": 3.8157917046288703, "learning_rate": 1.0236979773772757e-05, "loss": 0.6432, "step": 2498 }, { "epoch": 5.079268292682927, "grad_norm": 3.4036019201980223, "learning_rate": 1.0230398183958647e-05, "loss": 0.6173, "step": 2499 }, { "epoch": 5.08130081300813, "grad_norm": 3.0828210357632746, "learning_rate": 1.0223816494287675e-05, "loss": 0.5794, "step": 2500 }, { "epoch": 5.083333333333333, "grad_norm": 3.485945080444728, "learning_rate": 1.0217234707612409e-05, "loss": 0.7421, "step": 2501 }, { "epoch": 5.085365853658536, "grad_norm": 3.621214537322825, "learning_rate": 1.0210652826785466e-05, "loss": 0.5242, "step": 2502 }, { "epoch": 5.08739837398374, "grad_norm": 3.8472019835440405, "learning_rate": 1.0204070854659494e-05, "loss": 0.5961, "step": 2503 }, { "epoch": 5.0894308943089435, "grad_norm": 3.8260773948733875, "learning_rate": 1.0197488794087188e-05, "loss": 0.5467, "step": 2504 }, { "epoch": 5.091463414634147, "grad_norm": 3.7151235747973756, "learning_rate": 1.0190906647921277e-05, "loss": 0.6876, "step": 2505 }, { "epoch": 5.09349593495935, "grad_norm": 3.762180234242759, "learning_rate": 1.0184324419014526e-05, "loss": 0.6745, "step": 2506 }, { "epoch": 5.095528455284553, "grad_norm": 3.967883363282315, "learning_rate": 1.0177742110219746e-05, "loss": 0.5329, "step": 2507 }, { "epoch": 5.097560975609756, "grad_norm": 3.553620941738808, "learning_rate": 1.0171159724389766e-05, "loss": 0.5191, "step": 2508 }, { "epoch": 5.099593495934959, "grad_norm": 3.318261735622568, "learning_rate": 1.016457726437746e-05, "loss": 0.5247, "step": 2509 }, { "epoch": 5.1016260162601625, "grad_norm": 3.612249283384346, "learning_rate": 1.0157994733035738e-05, "loss": 0.6671, "step": 2510 }, { "epoch": 5.103658536585366, "grad_norm": 3.610676915190277, "learning_rate": 1.0151412133217527e-05, "loss": 0.4733, "step": 2511 }, { "epoch": 5.105691056910569, "grad_norm": 3.4870044294308085, "learning_rate": 1.0144829467775794e-05, "loss": 0.6275, "step": 2512 }, { "epoch": 5.107723577235772, "grad_norm": 3.3384300535909435, "learning_rate": 1.013824673956353e-05, "loss": 0.694, "step": 2513 }, { "epoch": 5.109756097560975, "grad_norm": 3.4223938137991623, "learning_rate": 1.0131663951433755e-05, "loss": 0.5077, "step": 2514 }, { "epoch": 5.111788617886178, "grad_norm": 3.5860612907273928, "learning_rate": 1.0125081106239518e-05, "loss": 0.4923, "step": 2515 }, { "epoch": 5.1138211382113825, "grad_norm": 3.281506840508952, "learning_rate": 1.0118498206833886e-05, "loss": 0.5916, "step": 2516 }, { "epoch": 5.115853658536586, "grad_norm": 3.424240922109997, "learning_rate": 1.0111915256069957e-05, "loss": 0.5899, "step": 2517 }, { "epoch": 5.117886178861789, "grad_norm": 3.781142573778258, "learning_rate": 1.0105332256800842e-05, "loss": 0.5836, "step": 2518 }, { "epoch": 5.119918699186992, "grad_norm": 3.452022995352727, "learning_rate": 1.0098749211879682e-05, "loss": 0.5354, "step": 2519 }, { "epoch": 5.121951219512195, "grad_norm": 3.690583033256952, "learning_rate": 1.0092166124159628e-05, "loss": 0.6249, "step": 2520 }, { "epoch": 5.123983739837398, "grad_norm": 3.4964007267033574, "learning_rate": 1.0085582996493865e-05, "loss": 0.5577, "step": 2521 }, { "epoch": 5.126016260162602, "grad_norm": 3.4687223700284746, "learning_rate": 1.0078999831735583e-05, "loss": 0.576, "step": 2522 }, { "epoch": 5.128048780487805, "grad_norm": 3.6248415639864655, "learning_rate": 1.0072416632737985e-05, "loss": 0.5582, "step": 2523 }, { "epoch": 5.130081300813008, "grad_norm": 3.4561053793110212, "learning_rate": 1.0065833402354302e-05, "loss": 0.5294, "step": 2524 }, { "epoch": 5.132113821138211, "grad_norm": 3.6293564672831713, "learning_rate": 1.0059250143437764e-05, "loss": 0.5177, "step": 2525 }, { "epoch": 5.134146341463414, "grad_norm": 3.764844113572759, "learning_rate": 1.0052666858841632e-05, "loss": 0.6291, "step": 2526 }, { "epoch": 5.1361788617886175, "grad_norm": 3.199098611407848, "learning_rate": 1.0046083551419157e-05, "loss": 0.66, "step": 2527 }, { "epoch": 5.138211382113822, "grad_norm": 3.837774864260081, "learning_rate": 1.003950022402361e-05, "loss": 0.6415, "step": 2528 }, { "epoch": 5.140243902439025, "grad_norm": 3.3942775359204713, "learning_rate": 1.0032916879508279e-05, "loss": 0.5039, "step": 2529 }, { "epoch": 5.142276422764228, "grad_norm": 3.51072978352436, "learning_rate": 1.0026333520726441e-05, "loss": 0.5671, "step": 2530 }, { "epoch": 5.144308943089431, "grad_norm": 3.1928600650341257, "learning_rate": 1.0019750150531395e-05, "loss": 0.6541, "step": 2531 }, { "epoch": 5.146341463414634, "grad_norm": 4.024389687632778, "learning_rate": 1.0013166771776441e-05, "loss": 0.6162, "step": 2532 }, { "epoch": 5.1483739837398375, "grad_norm": 3.786412166275772, "learning_rate": 1.0006583387314874e-05, "loss": 0.5289, "step": 2533 }, { "epoch": 5.150406504065041, "grad_norm": 3.534128878367121, "learning_rate": 1e-05, "loss": 0.5952, "step": 2534 }, { "epoch": 5.152439024390244, "grad_norm": 3.526197486147381, "learning_rate": 9.993416612685128e-06, "loss": 0.5956, "step": 2535 }, { "epoch": 5.154471544715447, "grad_norm": 3.9710939172291226, "learning_rate": 9.986833228223562e-06, "loss": 0.563, "step": 2536 }, { "epoch": 5.15650406504065, "grad_norm": 3.281744096770357, "learning_rate": 9.980249849468606e-06, "loss": 0.5721, "step": 2537 }, { "epoch": 5.158536585365853, "grad_norm": 3.7054509246348606, "learning_rate": 9.973666479273562e-06, "loss": 0.456, "step": 2538 }, { "epoch": 5.1605691056910565, "grad_norm": 3.480182408395269, "learning_rate": 9.967083120491726e-06, "loss": 0.5525, "step": 2539 }, { "epoch": 5.16260162601626, "grad_norm": 3.4020471237718084, "learning_rate": 9.96049977597639e-06, "loss": 0.5867, "step": 2540 }, { "epoch": 5.164634146341464, "grad_norm": 3.6587483721853604, "learning_rate": 9.953916448580844e-06, "loss": 0.5197, "step": 2541 }, { "epoch": 5.166666666666667, "grad_norm": 3.6654445510291502, "learning_rate": 9.947333141158373e-06, "loss": 0.504, "step": 2542 }, { "epoch": 5.16869918699187, "grad_norm": 4.121484816008554, "learning_rate": 9.94074985656224e-06, "loss": 0.571, "step": 2543 }, { "epoch": 5.170731707317073, "grad_norm": 3.4707180473190373, "learning_rate": 9.934166597645703e-06, "loss": 0.4801, "step": 2544 }, { "epoch": 5.1727642276422765, "grad_norm": 4.558598788457441, "learning_rate": 9.927583367262015e-06, "loss": 0.6903, "step": 2545 }, { "epoch": 5.17479674796748, "grad_norm": 4.157234373967582, "learning_rate": 9.921000168264424e-06, "loss": 0.5342, "step": 2546 }, { "epoch": 5.176829268292683, "grad_norm": 3.7044714801351577, "learning_rate": 9.914417003506136e-06, "loss": 0.6119, "step": 2547 }, { "epoch": 5.178861788617886, "grad_norm": 3.569878831062616, "learning_rate": 9.907833875840374e-06, "loss": 0.5748, "step": 2548 }, { "epoch": 5.180894308943089, "grad_norm": 3.749894682863454, "learning_rate": 9.901250788120323e-06, "loss": 0.6934, "step": 2549 }, { "epoch": 5.182926829268292, "grad_norm": 3.7993079777027403, "learning_rate": 9.89466774319916e-06, "loss": 0.5966, "step": 2550 }, { "epoch": 5.184959349593496, "grad_norm": 3.645472157178223, "learning_rate": 9.88808474393005e-06, "loss": 0.599, "step": 2551 }, { "epoch": 5.186991869918699, "grad_norm": 3.0944643282111084, "learning_rate": 9.881501793166117e-06, "loss": 0.5699, "step": 2552 }, { "epoch": 5.189024390243903, "grad_norm": 3.58200474251695, "learning_rate": 9.874918893760486e-06, "loss": 0.5858, "step": 2553 }, { "epoch": 5.191056910569106, "grad_norm": 3.3289507171283166, "learning_rate": 9.868336048566245e-06, "loss": 0.6712, "step": 2554 }, { "epoch": 5.193089430894309, "grad_norm": 3.4117794561756694, "learning_rate": 9.861753260436472e-06, "loss": 0.4862, "step": 2555 }, { "epoch": 5.195121951219512, "grad_norm": 4.6910561942944815, "learning_rate": 9.85517053222421e-06, "loss": 0.6392, "step": 2556 }, { "epoch": 5.1971544715447155, "grad_norm": 3.6216100683076595, "learning_rate": 9.848587866782476e-06, "loss": 0.5836, "step": 2557 }, { "epoch": 5.199186991869919, "grad_norm": 3.5048409739150057, "learning_rate": 9.842005266964263e-06, "loss": 0.5626, "step": 2558 }, { "epoch": 5.201219512195122, "grad_norm": 3.745034055964563, "learning_rate": 9.83542273562254e-06, "loss": 0.6074, "step": 2559 }, { "epoch": 5.203252032520325, "grad_norm": 3.260478746637701, "learning_rate": 9.82884027561024e-06, "loss": 0.5952, "step": 2560 }, { "epoch": 5.205284552845528, "grad_norm": 3.414174829512448, "learning_rate": 9.82225788978026e-06, "loss": 0.5645, "step": 2561 }, { "epoch": 5.2073170731707314, "grad_norm": 2.9874261446654002, "learning_rate": 9.815675580985475e-06, "loss": 0.5842, "step": 2562 }, { "epoch": 5.209349593495935, "grad_norm": 3.954623538974978, "learning_rate": 9.809093352078726e-06, "loss": 0.6383, "step": 2563 }, { "epoch": 5.211382113821138, "grad_norm": 3.92430778467177, "learning_rate": 9.802511205912815e-06, "loss": 0.6358, "step": 2564 }, { "epoch": 5.213414634146342, "grad_norm": 3.7546412099454654, "learning_rate": 9.795929145340511e-06, "loss": 0.6413, "step": 2565 }, { "epoch": 5.215447154471545, "grad_norm": 3.2369984672661807, "learning_rate": 9.789347173214538e-06, "loss": 0.5582, "step": 2566 }, { "epoch": 5.217479674796748, "grad_norm": 3.252891972959923, "learning_rate": 9.782765292387593e-06, "loss": 0.6402, "step": 2567 }, { "epoch": 5.219512195121951, "grad_norm": 3.8306330966780227, "learning_rate": 9.776183505712327e-06, "loss": 0.5641, "step": 2568 }, { "epoch": 5.221544715447155, "grad_norm": 3.8700632001605366, "learning_rate": 9.769601816041355e-06, "loss": 0.5943, "step": 2569 }, { "epoch": 5.223577235772358, "grad_norm": 3.375553232430163, "learning_rate": 9.763020226227246e-06, "loss": 0.4849, "step": 2570 }, { "epoch": 5.225609756097561, "grad_norm": 3.7924817373190187, "learning_rate": 9.756438739122523e-06, "loss": 0.6408, "step": 2571 }, { "epoch": 5.227642276422764, "grad_norm": 3.865756690108456, "learning_rate": 9.749857357579667e-06, "loss": 0.5986, "step": 2572 }, { "epoch": 5.229674796747967, "grad_norm": 4.157140316800948, "learning_rate": 9.74327608445112e-06, "loss": 0.4431, "step": 2573 }, { "epoch": 5.2317073170731705, "grad_norm": 4.70994144456786, "learning_rate": 9.736694922589268e-06, "loss": 0.7363, "step": 2574 }, { "epoch": 5.233739837398374, "grad_norm": 4.252519201550621, "learning_rate": 9.730113874846449e-06, "loss": 0.5733, "step": 2575 }, { "epoch": 5.235772357723577, "grad_norm": 3.3870151793255103, "learning_rate": 9.723532944074961e-06, "loss": 0.5379, "step": 2576 }, { "epoch": 5.237804878048781, "grad_norm": 4.145748846374422, "learning_rate": 9.716952133127037e-06, "loss": 0.6269, "step": 2577 }, { "epoch": 5.239837398373984, "grad_norm": 3.6549095350817975, "learning_rate": 9.71037144485487e-06, "loss": 0.5858, "step": 2578 }, { "epoch": 5.241869918699187, "grad_norm": 3.553104863155404, "learning_rate": 9.7037908821106e-06, "loss": 0.5943, "step": 2579 }, { "epoch": 5.2439024390243905, "grad_norm": 3.2399377767431243, "learning_rate": 9.6972104477463e-06, "loss": 0.6655, "step": 2580 }, { "epoch": 5.245934959349594, "grad_norm": 4.200029780995932, "learning_rate": 9.690630144614e-06, "loss": 0.6715, "step": 2581 }, { "epoch": 5.247967479674797, "grad_norm": 3.089056436202186, "learning_rate": 9.684049975565669e-06, "loss": 0.5293, "step": 2582 }, { "epoch": 5.25, "grad_norm": 4.043551939564905, "learning_rate": 9.677469943453213e-06, "loss": 0.6577, "step": 2583 }, { "epoch": 5.252032520325203, "grad_norm": 4.16807372013878, "learning_rate": 9.670890051128493e-06, "loss": 0.5257, "step": 2584 }, { "epoch": 5.254065040650406, "grad_norm": 3.8359223961869486, "learning_rate": 9.664310301443289e-06, "loss": 0.5106, "step": 2585 }, { "epoch": 5.2560975609756095, "grad_norm": 4.064878217535024, "learning_rate": 9.657730697249337e-06, "loss": 0.517, "step": 2586 }, { "epoch": 5.258130081300813, "grad_norm": 3.6468078984957044, "learning_rate": 9.651151241398299e-06, "loss": 0.584, "step": 2587 }, { "epoch": 5.260162601626016, "grad_norm": 3.165376729238993, "learning_rate": 9.644571936741778e-06, "loss": 0.6115, "step": 2588 }, { "epoch": 5.262195121951219, "grad_norm": 3.3938802535431325, "learning_rate": 9.637992786131312e-06, "loss": 0.7171, "step": 2589 }, { "epoch": 5.264227642276423, "grad_norm": 3.698916499210059, "learning_rate": 9.631413792418366e-06, "loss": 0.5788, "step": 2590 }, { "epoch": 5.266260162601626, "grad_norm": 4.256094343335853, "learning_rate": 9.624834958454344e-06, "loss": 0.6721, "step": 2591 }, { "epoch": 5.2682926829268295, "grad_norm": 4.07058643877208, "learning_rate": 9.618256287090576e-06, "loss": 0.4739, "step": 2592 }, { "epoch": 5.270325203252033, "grad_norm": 4.147443361349221, "learning_rate": 9.611677781178328e-06, "loss": 0.6743, "step": 2593 }, { "epoch": 5.272357723577236, "grad_norm": 4.405214831268133, "learning_rate": 9.605099443568784e-06, "loss": 0.5974, "step": 2594 }, { "epoch": 5.274390243902439, "grad_norm": 4.141945950259477, "learning_rate": 9.598521277113062e-06, "loss": 0.638, "step": 2595 }, { "epoch": 5.276422764227642, "grad_norm": 3.9645231999711927, "learning_rate": 9.591943284662206e-06, "loss": 0.6034, "step": 2596 }, { "epoch": 5.278455284552845, "grad_norm": 3.5774165388001187, "learning_rate": 9.585365469067186e-06, "loss": 0.7651, "step": 2597 }, { "epoch": 5.280487804878049, "grad_norm": 3.7582146809524852, "learning_rate": 9.578787833178893e-06, "loss": 0.6754, "step": 2598 }, { "epoch": 5.282520325203252, "grad_norm": 3.044514450630015, "learning_rate": 9.57221037984813e-06, "loss": 0.4793, "step": 2599 }, { "epoch": 5.284552845528455, "grad_norm": 3.8953125322159283, "learning_rate": 9.56563311192564e-06, "loss": 0.5678, "step": 2600 }, { "epoch": 5.286585365853658, "grad_norm": 3.0538047411746447, "learning_rate": 9.559056032262073e-06, "loss": 0.4842, "step": 2601 }, { "epoch": 5.288617886178862, "grad_norm": 3.720269220113188, "learning_rate": 9.552479143708003e-06, "loss": 0.5501, "step": 2602 }, { "epoch": 5.290650406504065, "grad_norm": 3.260484633845254, "learning_rate": 9.545902449113918e-06, "loss": 0.4976, "step": 2603 }, { "epoch": 5.2926829268292686, "grad_norm": 3.753285301952384, "learning_rate": 9.53932595133022e-06, "loss": 0.5755, "step": 2604 }, { "epoch": 5.294715447154472, "grad_norm": 3.346718585727812, "learning_rate": 9.532749653207228e-06, "loss": 0.8559, "step": 2605 }, { "epoch": 5.296747967479675, "grad_norm": 3.6939865085490613, "learning_rate": 9.526173557595179e-06, "loss": 0.5653, "step": 2606 }, { "epoch": 5.298780487804878, "grad_norm": 3.431602406415012, "learning_rate": 9.51959766734422e-06, "loss": 0.582, "step": 2607 }, { "epoch": 5.300813008130081, "grad_norm": 4.1860472514193665, "learning_rate": 9.513021985304399e-06, "loss": 0.5552, "step": 2608 }, { "epoch": 5.3028455284552845, "grad_norm": 3.68419224466781, "learning_rate": 9.506446514325687e-06, "loss": 0.4745, "step": 2609 }, { "epoch": 5.304878048780488, "grad_norm": 3.6948949517017686, "learning_rate": 9.499871257257958e-06, "loss": 0.5459, "step": 2610 }, { "epoch": 5.306910569105691, "grad_norm": 3.5410526446294295, "learning_rate": 9.493296216950993e-06, "loss": 0.5695, "step": 2611 }, { "epoch": 5.308943089430894, "grad_norm": 4.2901946010793965, "learning_rate": 9.486721396254484e-06, "loss": 0.458, "step": 2612 }, { "epoch": 5.310975609756097, "grad_norm": 3.807130532666538, "learning_rate": 9.480146798018015e-06, "loss": 0.7303, "step": 2613 }, { "epoch": 5.3130081300813, "grad_norm": 4.045983919718652, "learning_rate": 9.473572425091091e-06, "loss": 0.72, "step": 2614 }, { "epoch": 5.315040650406504, "grad_norm": 3.860384115477893, "learning_rate": 9.466998280323103e-06, "loss": 0.6146, "step": 2615 }, { "epoch": 5.317073170731708, "grad_norm": 3.884550100970543, "learning_rate": 9.460424366563355e-06, "loss": 0.4467, "step": 2616 }, { "epoch": 5.319105691056911, "grad_norm": 3.611823930264421, "learning_rate": 9.453850686661051e-06, "loss": 0.6362, "step": 2617 }, { "epoch": 5.321138211382114, "grad_norm": 4.170323291756602, "learning_rate": 9.447277243465278e-06, "loss": 0.6847, "step": 2618 }, { "epoch": 5.323170731707317, "grad_norm": 3.3575151585625957, "learning_rate": 9.44070403982504e-06, "loss": 0.5249, "step": 2619 }, { "epoch": 5.32520325203252, "grad_norm": 3.323048292320722, "learning_rate": 9.434131078589224e-06, "loss": 0.5673, "step": 2620 }, { "epoch": 5.3272357723577235, "grad_norm": 3.2495007671629526, "learning_rate": 9.427558362606624e-06, "loss": 0.5367, "step": 2621 }, { "epoch": 5.329268292682927, "grad_norm": 3.5782109051312236, "learning_rate": 9.42098589472591e-06, "loss": 0.625, "step": 2622 }, { "epoch": 5.33130081300813, "grad_norm": 3.8270870383238544, "learning_rate": 9.41441367779566e-06, "loss": 0.5496, "step": 2623 }, { "epoch": 5.333333333333333, "grad_norm": 6.410841409243114, "learning_rate": 9.407841714664343e-06, "loss": 0.558, "step": 2624 }, { "epoch": 5.335365853658536, "grad_norm": 3.9271549217472175, "learning_rate": 9.401270008180304e-06, "loss": 0.4871, "step": 2625 }, { "epoch": 5.33739837398374, "grad_norm": 3.6574376283139287, "learning_rate": 9.394698561191796e-06, "loss": 0.6349, "step": 2626 }, { "epoch": 5.3394308943089435, "grad_norm": 3.0354578782533324, "learning_rate": 9.388127376546938e-06, "loss": 0.5193, "step": 2627 }, { "epoch": 5.341463414634147, "grad_norm": 4.081711299745366, "learning_rate": 9.381556457093752e-06, "loss": 0.5868, "step": 2628 }, { "epoch": 5.34349593495935, "grad_norm": 3.0529312965355873, "learning_rate": 9.374985805680143e-06, "loss": 0.4454, "step": 2629 }, { "epoch": 5.345528455284553, "grad_norm": 3.762448629943356, "learning_rate": 9.368415425153891e-06, "loss": 0.652, "step": 2630 }, { "epoch": 5.347560975609756, "grad_norm": 4.561034993524711, "learning_rate": 9.361845318362672e-06, "loss": 0.6896, "step": 2631 }, { "epoch": 5.349593495934959, "grad_norm": 3.7819000175948663, "learning_rate": 9.355275488154025e-06, "loss": 0.5972, "step": 2632 }, { "epoch": 5.3516260162601625, "grad_norm": 3.4305402745640796, "learning_rate": 9.348705937375387e-06, "loss": 0.4932, "step": 2633 }, { "epoch": 5.353658536585366, "grad_norm": 3.3971565013483715, "learning_rate": 9.342136668874063e-06, "loss": 0.5692, "step": 2634 }, { "epoch": 5.355691056910569, "grad_norm": 3.605690123093172, "learning_rate": 9.335567685497246e-06, "loss": 0.5306, "step": 2635 }, { "epoch": 5.357723577235772, "grad_norm": 3.7673531586081657, "learning_rate": 9.32899899009199e-06, "loss": 0.6937, "step": 2636 }, { "epoch": 5.359756097560975, "grad_norm": 3.645420591086862, "learning_rate": 9.322430585505236e-06, "loss": 0.52, "step": 2637 }, { "epoch": 5.361788617886178, "grad_norm": 3.9802651385932695, "learning_rate": 9.315862474583795e-06, "loss": 0.5949, "step": 2638 }, { "epoch": 5.3638211382113825, "grad_norm": 3.6275536669109396, "learning_rate": 9.309294660174351e-06, "loss": 0.5381, "step": 2639 }, { "epoch": 5.365853658536586, "grad_norm": 3.6882879666668082, "learning_rate": 9.30272714512347e-06, "loss": 0.489, "step": 2640 }, { "epoch": 5.367886178861789, "grad_norm": 3.759231646648565, "learning_rate": 9.296159932277564e-06, "loss": 0.5419, "step": 2641 }, { "epoch": 5.369918699186992, "grad_norm": 3.52168071523663, "learning_rate": 9.289593024482934e-06, "loss": 0.4397, "step": 2642 }, { "epoch": 5.371951219512195, "grad_norm": 3.00133142720939, "learning_rate": 9.283026424585741e-06, "loss": 0.42, "step": 2643 }, { "epoch": 5.373983739837398, "grad_norm": 4.148819912845683, "learning_rate": 9.276460135432019e-06, "loss": 0.7367, "step": 2644 }, { "epoch": 5.376016260162602, "grad_norm": 4.30223292499144, "learning_rate": 9.269894159867665e-06, "loss": 0.6186, "step": 2645 }, { "epoch": 5.378048780487805, "grad_norm": 3.751394285068468, "learning_rate": 9.263328500738428e-06, "loss": 0.4814, "step": 2646 }, { "epoch": 5.380081300813008, "grad_norm": 3.355676254939446, "learning_rate": 9.256763160889939e-06, "loss": 0.4721, "step": 2647 }, { "epoch": 5.382113821138211, "grad_norm": 3.858001220942699, "learning_rate": 9.250198143167675e-06, "loss": 0.5805, "step": 2648 }, { "epoch": 5.384146341463414, "grad_norm": 3.7618644003058948, "learning_rate": 9.243633450416984e-06, "loss": 0.5538, "step": 2649 }, { "epoch": 5.3861788617886175, "grad_norm": 3.806926383316063, "learning_rate": 9.237069085483072e-06, "loss": 0.4978, "step": 2650 }, { "epoch": 5.388211382113822, "grad_norm": 4.231485149934423, "learning_rate": 9.230505051210991e-06, "loss": 0.7198, "step": 2651 }, { "epoch": 5.390243902439025, "grad_norm": 4.331528990230237, "learning_rate": 9.223941350445666e-06, "loss": 0.52, "step": 2652 }, { "epoch": 5.392276422764228, "grad_norm": 3.323941212051932, "learning_rate": 9.217377986031867e-06, "loss": 0.4791, "step": 2653 }, { "epoch": 5.394308943089431, "grad_norm": 3.5082028342854494, "learning_rate": 9.210814960814226e-06, "loss": 0.5214, "step": 2654 }, { "epoch": 5.396341463414634, "grad_norm": 3.690733104385041, "learning_rate": 9.204252277637215e-06, "loss": 0.5937, "step": 2655 }, { "epoch": 5.3983739837398375, "grad_norm": 3.6364935094918946, "learning_rate": 9.19768993934517e-06, "loss": 0.5497, "step": 2656 }, { "epoch": 5.400406504065041, "grad_norm": 3.257349374011521, "learning_rate": 9.191127948782277e-06, "loss": 0.575, "step": 2657 }, { "epoch": 5.402439024390244, "grad_norm": 3.755863646425733, "learning_rate": 9.184566308792561e-06, "loss": 0.4879, "step": 2658 }, { "epoch": 5.404471544715447, "grad_norm": 3.4019845914189615, "learning_rate": 9.178005022219912e-06, "loss": 0.4602, "step": 2659 }, { "epoch": 5.40650406504065, "grad_norm": 4.760156926554227, "learning_rate": 9.171444091908046e-06, "loss": 0.5673, "step": 2660 }, { "epoch": 5.408536585365853, "grad_norm": 3.746851584153235, "learning_rate": 9.16488352070054e-06, "loss": 0.5382, "step": 2661 }, { "epoch": 5.4105691056910565, "grad_norm": 3.58456284548545, "learning_rate": 9.158323311440812e-06, "loss": 0.5277, "step": 2662 }, { "epoch": 5.41260162601626, "grad_norm": 3.5960217670035215, "learning_rate": 9.15176346697212e-06, "loss": 0.6263, "step": 2663 }, { "epoch": 5.414634146341464, "grad_norm": 4.032052881158552, "learning_rate": 9.145203990137571e-06, "loss": 0.6947, "step": 2664 }, { "epoch": 5.416666666666667, "grad_norm": 3.272292655379001, "learning_rate": 9.138644883780097e-06, "loss": 0.5416, "step": 2665 }, { "epoch": 5.41869918699187, "grad_norm": 4.625170096616691, "learning_rate": 9.132086150742486e-06, "loss": 0.6175, "step": 2666 }, { "epoch": 5.420731707317073, "grad_norm": 3.6867567480425043, "learning_rate": 9.12552779386736e-06, "loss": 0.4783, "step": 2667 }, { "epoch": 5.4227642276422765, "grad_norm": 4.238984151912853, "learning_rate": 9.118969815997174e-06, "loss": 0.5145, "step": 2668 }, { "epoch": 5.42479674796748, "grad_norm": 3.8364941788364946, "learning_rate": 9.11241221997422e-06, "loss": 0.6658, "step": 2669 }, { "epoch": 5.426829268292683, "grad_norm": 4.063036621845115, "learning_rate": 9.105855008640624e-06, "loss": 0.7333, "step": 2670 }, { "epoch": 5.428861788617886, "grad_norm": 3.44843091918468, "learning_rate": 9.09929818483835e-06, "loss": 0.5371, "step": 2671 }, { "epoch": 5.430894308943089, "grad_norm": 3.7183302158964593, "learning_rate": 9.092741751409186e-06, "loss": 0.531, "step": 2672 }, { "epoch": 5.432926829268292, "grad_norm": 3.862342553513551, "learning_rate": 9.086185711194763e-06, "loss": 0.5079, "step": 2673 }, { "epoch": 5.434959349593496, "grad_norm": 3.7610404141761378, "learning_rate": 9.079630067036526e-06, "loss": 0.5654, "step": 2674 }, { "epoch": 5.4369918699187, "grad_norm": 3.6538339393259243, "learning_rate": 9.07307482177576e-06, "loss": 0.4401, "step": 2675 }, { "epoch": 5.439024390243903, "grad_norm": 4.041855174360427, "learning_rate": 9.06651997825357e-06, "loss": 0.5554, "step": 2676 }, { "epoch": 5.441056910569106, "grad_norm": 4.075115336884467, "learning_rate": 9.059965539310894e-06, "loss": 0.5161, "step": 2677 }, { "epoch": 5.443089430894309, "grad_norm": 3.8111994484313723, "learning_rate": 9.053411507788494e-06, "loss": 0.4764, "step": 2678 }, { "epoch": 5.445121951219512, "grad_norm": 3.887834826443285, "learning_rate": 9.046857886526948e-06, "loss": 0.575, "step": 2679 }, { "epoch": 5.4471544715447155, "grad_norm": 3.4388796260546686, "learning_rate": 9.040304678366658e-06, "loss": 0.7128, "step": 2680 }, { "epoch": 5.449186991869919, "grad_norm": 3.7760387472916137, "learning_rate": 9.03375188614785e-06, "loss": 0.5257, "step": 2681 }, { "epoch": 5.451219512195122, "grad_norm": 3.554473485531629, "learning_rate": 9.027199512710579e-06, "loss": 0.4501, "step": 2682 }, { "epoch": 5.453252032520325, "grad_norm": 3.930759817348624, "learning_rate": 9.020647560894696e-06, "loss": 0.5802, "step": 2683 }, { "epoch": 5.455284552845528, "grad_norm": 3.8389415874318886, "learning_rate": 9.014096033539889e-06, "loss": 0.5643, "step": 2684 }, { "epoch": 5.4573170731707314, "grad_norm": 3.8284278084906855, "learning_rate": 9.007544933485651e-06, "loss": 0.6512, "step": 2685 }, { "epoch": 5.459349593495935, "grad_norm": 3.2069105077130384, "learning_rate": 9.000994263571297e-06, "loss": 0.4047, "step": 2686 }, { "epoch": 5.461382113821138, "grad_norm": 3.840954525784949, "learning_rate": 8.994444026635956e-06, "loss": 0.5825, "step": 2687 }, { "epoch": 5.463414634146342, "grad_norm": 4.106239995227786, "learning_rate": 8.987894225518556e-06, "loss": 0.5013, "step": 2688 }, { "epoch": 5.465447154471545, "grad_norm": 3.909148559929816, "learning_rate": 8.981344863057852e-06, "loss": 0.5291, "step": 2689 }, { "epoch": 5.467479674796748, "grad_norm": 3.4493083634445116, "learning_rate": 8.974795942092404e-06, "loss": 0.5716, "step": 2690 }, { "epoch": 5.469512195121951, "grad_norm": 3.4452185185915734, "learning_rate": 8.968247465460576e-06, "loss": 0.551, "step": 2691 }, { "epoch": 5.471544715447155, "grad_norm": 3.9531128215037654, "learning_rate": 8.961699436000548e-06, "loss": 0.6284, "step": 2692 }, { "epoch": 5.473577235772358, "grad_norm": 3.5753782125762608, "learning_rate": 8.955151856550294e-06, "loss": 0.581, "step": 2693 }, { "epoch": 5.475609756097561, "grad_norm": 3.45848596386055, "learning_rate": 8.9486047299476e-06, "loss": 0.5412, "step": 2694 }, { "epoch": 5.477642276422764, "grad_norm": 3.82591331987142, "learning_rate": 8.942058059030064e-06, "loss": 0.5819, "step": 2695 }, { "epoch": 5.479674796747967, "grad_norm": 3.218100354325049, "learning_rate": 8.93551184663507e-06, "loss": 0.5714, "step": 2696 }, { "epoch": 5.4817073170731705, "grad_norm": 3.743373012495574, "learning_rate": 8.928966095599819e-06, "loss": 0.7025, "step": 2697 }, { "epoch": 5.483739837398374, "grad_norm": 2.988029732799845, "learning_rate": 8.922420808761296e-06, "loss": 0.6036, "step": 2698 }, { "epoch": 5.485772357723577, "grad_norm": 3.395414158054255, "learning_rate": 8.915875988956296e-06, "loss": 0.4526, "step": 2699 }, { "epoch": 5.487804878048781, "grad_norm": 3.5875780490281133, "learning_rate": 8.909331639021414e-06, "loss": 0.5829, "step": 2700 }, { "epoch": 5.489837398373984, "grad_norm": 3.435525020398749, "learning_rate": 8.902787761793032e-06, "loss": 0.6283, "step": 2701 }, { "epoch": 5.491869918699187, "grad_norm": 3.374025110687133, "learning_rate": 8.896244360107331e-06, "loss": 0.7161, "step": 2702 }, { "epoch": 5.4939024390243905, "grad_norm": 3.3687662124833864, "learning_rate": 8.889701436800285e-06, "loss": 0.5006, "step": 2703 }, { "epoch": 5.495934959349594, "grad_norm": 4.093706907843455, "learning_rate": 8.883158994707666e-06, "loss": 0.5395, "step": 2704 }, { "epoch": 5.497967479674797, "grad_norm": 3.562808568575615, "learning_rate": 8.876617036665031e-06, "loss": 0.6409, "step": 2705 }, { "epoch": 5.5, "grad_norm": 4.395823018152493, "learning_rate": 8.870075565507734e-06, "loss": 0.6196, "step": 2706 }, { "epoch": 5.502032520325203, "grad_norm": 3.6813927988881017, "learning_rate": 8.863534584070907e-06, "loss": 0.5289, "step": 2707 }, { "epoch": 5.504065040650406, "grad_norm": 3.6748957885842493, "learning_rate": 8.856994095189477e-06, "loss": 0.5679, "step": 2708 }, { "epoch": 5.5060975609756095, "grad_norm": 4.450692158220494, "learning_rate": 8.850454101698159e-06, "loss": 0.7632, "step": 2709 }, { "epoch": 5.508130081300813, "grad_norm": 3.3734062408256675, "learning_rate": 8.843914606431453e-06, "loss": 0.57, "step": 2710 }, { "epoch": 5.510162601626016, "grad_norm": 3.694886757772571, "learning_rate": 8.83737561222364e-06, "loss": 0.6507, "step": 2711 }, { "epoch": 5.512195121951219, "grad_norm": 3.6784318510414367, "learning_rate": 8.830837121908783e-06, "loss": 0.6382, "step": 2712 }, { "epoch": 5.514227642276423, "grad_norm": 3.2878511579543837, "learning_rate": 8.82429913832073e-06, "loss": 0.642, "step": 2713 }, { "epoch": 5.516260162601626, "grad_norm": 3.507746178167954, "learning_rate": 8.817761664293106e-06, "loss": 0.5128, "step": 2714 }, { "epoch": 5.5182926829268295, "grad_norm": 3.333454721065789, "learning_rate": 8.811224702659322e-06, "loss": 0.5427, "step": 2715 }, { "epoch": 5.520325203252033, "grad_norm": 3.730976527773014, "learning_rate": 8.804688256252557e-06, "loss": 0.7393, "step": 2716 }, { "epoch": 5.522357723577236, "grad_norm": 3.8475188953272923, "learning_rate": 8.798152327905775e-06, "loss": 0.6117, "step": 2717 }, { "epoch": 5.524390243902439, "grad_norm": 3.73993527709263, "learning_rate": 8.791616920451711e-06, "loss": 0.5541, "step": 2718 }, { "epoch": 5.526422764227642, "grad_norm": 4.453556796791605, "learning_rate": 8.785082036722875e-06, "loss": 0.7754, "step": 2719 }, { "epoch": 5.528455284552845, "grad_norm": 3.6521406735495283, "learning_rate": 8.778547679551555e-06, "loss": 0.5128, "step": 2720 }, { "epoch": 5.530487804878049, "grad_norm": 4.005245655878767, "learning_rate": 8.772013851769798e-06, "loss": 0.6082, "step": 2721 }, { "epoch": 5.532520325203252, "grad_norm": 3.872709392255697, "learning_rate": 8.765480556209438e-06, "loss": 0.618, "step": 2722 }, { "epoch": 5.534552845528455, "grad_norm": 3.4569909435475212, "learning_rate": 8.758947795702063e-06, "loss": 0.5293, "step": 2723 }, { "epoch": 5.536585365853659, "grad_norm": 3.6521587521904375, "learning_rate": 8.75241557307904e-06, "loss": 0.6173, "step": 2724 }, { "epoch": 5.538617886178862, "grad_norm": 3.6808927863754866, "learning_rate": 8.745883891171504e-06, "loss": 0.6366, "step": 2725 }, { "epoch": 5.540650406504065, "grad_norm": 4.11193655835135, "learning_rate": 8.739352752810342e-06, "loss": 0.5685, "step": 2726 }, { "epoch": 5.5426829268292686, "grad_norm": 3.596071553655811, "learning_rate": 8.732822160826216e-06, "loss": 0.5503, "step": 2727 }, { "epoch": 5.544715447154472, "grad_norm": 3.405988085920095, "learning_rate": 8.726292118049555e-06, "loss": 0.7299, "step": 2728 }, { "epoch": 5.546747967479675, "grad_norm": 3.6610895239027523, "learning_rate": 8.719762627310544e-06, "loss": 0.5622, "step": 2729 }, { "epoch": 5.548780487804878, "grad_norm": 3.715120902042622, "learning_rate": 8.71323369143912e-06, "loss": 0.5944, "step": 2730 }, { "epoch": 5.550813008130081, "grad_norm": 3.882738330119974, "learning_rate": 8.706705313264996e-06, "loss": 0.5844, "step": 2731 }, { "epoch": 5.5528455284552845, "grad_norm": 3.664117273583059, "learning_rate": 8.700177495617635e-06, "loss": 0.6207, "step": 2732 }, { "epoch": 5.554878048780488, "grad_norm": 3.969469340782468, "learning_rate": 8.693650241326259e-06, "loss": 0.4415, "step": 2733 }, { "epoch": 5.556910569105691, "grad_norm": 3.3682334668309246, "learning_rate": 8.687123553219845e-06, "loss": 0.5202, "step": 2734 }, { "epoch": 5.558943089430894, "grad_norm": 4.120401356419461, "learning_rate": 8.680597434127121e-06, "loss": 0.5771, "step": 2735 }, { "epoch": 5.560975609756097, "grad_norm": 3.8757007169342272, "learning_rate": 8.674071886876572e-06, "loss": 0.7574, "step": 2736 }, { "epoch": 5.5630081300813, "grad_norm": 3.5389076819756773, "learning_rate": 8.667546914296437e-06, "loss": 0.542, "step": 2737 }, { "epoch": 5.565040650406504, "grad_norm": 3.795960505082603, "learning_rate": 8.661022519214706e-06, "loss": 0.5666, "step": 2738 }, { "epoch": 5.567073170731708, "grad_norm": 4.254923773840668, "learning_rate": 8.654498704459114e-06, "loss": 0.4675, "step": 2739 }, { "epoch": 5.569105691056911, "grad_norm": 4.532665524642624, "learning_rate": 8.647975472857148e-06, "loss": 0.6102, "step": 2740 }, { "epoch": 5.571138211382114, "grad_norm": 3.8527845066921635, "learning_rate": 8.641452827236038e-06, "loss": 0.6047, "step": 2741 }, { "epoch": 5.573170731707317, "grad_norm": 3.6503088383045053, "learning_rate": 8.634930770422766e-06, "loss": 0.6889, "step": 2742 }, { "epoch": 5.57520325203252, "grad_norm": 3.6226334950469505, "learning_rate": 8.62840930524406e-06, "loss": 0.6702, "step": 2743 }, { "epoch": 5.5772357723577235, "grad_norm": 3.4608165185503723, "learning_rate": 8.621888434526382e-06, "loss": 0.6169, "step": 2744 }, { "epoch": 5.579268292682927, "grad_norm": 3.7835489133300366, "learning_rate": 8.615368161095947e-06, "loss": 0.7366, "step": 2745 }, { "epoch": 5.58130081300813, "grad_norm": 3.9151680516595513, "learning_rate": 8.608848487778701e-06, "loss": 0.6771, "step": 2746 }, { "epoch": 5.583333333333333, "grad_norm": 3.8246235421151873, "learning_rate": 8.602329417400335e-06, "loss": 0.5458, "step": 2747 }, { "epoch": 5.585365853658536, "grad_norm": 4.2949533356611145, "learning_rate": 8.595810952786289e-06, "loss": 0.7307, "step": 2748 }, { "epoch": 5.58739837398374, "grad_norm": 3.3666414384147214, "learning_rate": 8.589293096761717e-06, "loss": 0.5231, "step": 2749 }, { "epoch": 5.5894308943089435, "grad_norm": 3.5810358965465325, "learning_rate": 8.582775852151532e-06, "loss": 0.4907, "step": 2750 }, { "epoch": 5.591463414634147, "grad_norm": 4.299616443606635, "learning_rate": 8.576259221780365e-06, "loss": 0.5682, "step": 2751 }, { "epoch": 5.59349593495935, "grad_norm": 3.9271074674368895, "learning_rate": 8.569743208472594e-06, "loss": 0.5356, "step": 2752 }, { "epoch": 5.595528455284553, "grad_norm": 3.2068949893003533, "learning_rate": 8.563227815052326e-06, "loss": 0.4576, "step": 2753 }, { "epoch": 5.597560975609756, "grad_norm": 3.8457730958071266, "learning_rate": 8.55671304434339e-06, "loss": 0.6851, "step": 2754 }, { "epoch": 5.599593495934959, "grad_norm": 3.6335568662717272, "learning_rate": 8.550198899169356e-06, "loss": 0.52, "step": 2755 }, { "epoch": 5.6016260162601625, "grad_norm": 3.9233789204052916, "learning_rate": 8.543685382353518e-06, "loss": 0.7699, "step": 2756 }, { "epoch": 5.603658536585366, "grad_norm": 3.7167600801678033, "learning_rate": 8.5371724967189e-06, "loss": 0.5201, "step": 2757 }, { "epoch": 5.605691056910569, "grad_norm": 3.5690417166323347, "learning_rate": 8.530660245088257e-06, "loss": 0.4615, "step": 2758 }, { "epoch": 5.607723577235772, "grad_norm": 4.180373013461464, "learning_rate": 8.52414863028405e-06, "loss": 0.6553, "step": 2759 }, { "epoch": 5.609756097560975, "grad_norm": 3.7379477809098542, "learning_rate": 8.51763765512849e-06, "loss": 0.5526, "step": 2760 }, { "epoch": 5.611788617886178, "grad_norm": 3.4976508206214927, "learning_rate": 8.51112732244349e-06, "loss": 0.6384, "step": 2761 }, { "epoch": 5.613821138211382, "grad_norm": 3.751556165265369, "learning_rate": 8.504617635050702e-06, "loss": 0.4644, "step": 2762 }, { "epoch": 5.615853658536586, "grad_norm": 3.6740666050341257, "learning_rate": 8.49810859577148e-06, "loss": 0.4717, "step": 2763 }, { "epoch": 5.617886178861789, "grad_norm": 3.65068225063718, "learning_rate": 8.491600207426907e-06, "loss": 0.5822, "step": 2764 }, { "epoch": 5.619918699186992, "grad_norm": 3.8808806397043694, "learning_rate": 8.485092472837789e-06, "loss": 0.55, "step": 2765 }, { "epoch": 5.621951219512195, "grad_norm": 3.73349092324901, "learning_rate": 8.47858539482464e-06, "loss": 0.5008, "step": 2766 }, { "epoch": 5.623983739837398, "grad_norm": 4.153180537829123, "learning_rate": 8.472078976207696e-06, "loss": 0.604, "step": 2767 }, { "epoch": 5.626016260162602, "grad_norm": 3.5978259985240744, "learning_rate": 8.465573219806893e-06, "loss": 0.5898, "step": 2768 }, { "epoch": 5.628048780487805, "grad_norm": 4.20984142044793, "learning_rate": 8.459068128441897e-06, "loss": 0.7225, "step": 2769 }, { "epoch": 5.630081300813008, "grad_norm": 4.422588652101442, "learning_rate": 8.45256370493208e-06, "loss": 0.5825, "step": 2770 }, { "epoch": 5.632113821138211, "grad_norm": 4.1772136389733285, "learning_rate": 8.446059952096525e-06, "loss": 0.5391, "step": 2771 }, { "epoch": 5.634146341463414, "grad_norm": 3.7525296967133293, "learning_rate": 8.439556872754025e-06, "loss": 0.5148, "step": 2772 }, { "epoch": 5.636178861788618, "grad_norm": 3.6494588953137406, "learning_rate": 8.43305446972307e-06, "loss": 0.5953, "step": 2773 }, { "epoch": 5.638211382113822, "grad_norm": 3.746332432564448, "learning_rate": 8.42655274582187e-06, "loss": 0.6593, "step": 2774 }, { "epoch": 5.640243902439025, "grad_norm": 3.4994071474196797, "learning_rate": 8.42005170386834e-06, "loss": 0.4977, "step": 2775 }, { "epoch": 5.642276422764228, "grad_norm": 3.4849387365710855, "learning_rate": 8.413551346680095e-06, "loss": 0.6206, "step": 2776 }, { "epoch": 5.644308943089431, "grad_norm": 3.886412269060091, "learning_rate": 8.407051677074452e-06, "loss": 0.6528, "step": 2777 }, { "epoch": 5.646341463414634, "grad_norm": 4.251916048593202, "learning_rate": 8.400552697868435e-06, "loss": 0.6247, "step": 2778 }, { "epoch": 5.6483739837398375, "grad_norm": 3.965946117303312, "learning_rate": 8.394054411878763e-06, "loss": 0.546, "step": 2779 }, { "epoch": 5.650406504065041, "grad_norm": 3.800427724889804, "learning_rate": 8.38755682192186e-06, "loss": 0.5526, "step": 2780 }, { "epoch": 5.652439024390244, "grad_norm": 3.3765606829596004, "learning_rate": 8.381059930813852e-06, "loss": 0.4908, "step": 2781 }, { "epoch": 5.654471544715447, "grad_norm": 3.5079096031579553, "learning_rate": 8.374563741370542e-06, "loss": 0.4967, "step": 2782 }, { "epoch": 5.65650406504065, "grad_norm": 3.68477883398733, "learning_rate": 8.368068256407454e-06, "loss": 0.6077, "step": 2783 }, { "epoch": 5.658536585365853, "grad_norm": 4.176050643612972, "learning_rate": 8.36157347873979e-06, "loss": 0.4754, "step": 2784 }, { "epoch": 5.6605691056910565, "grad_norm": 3.8732469803548684, "learning_rate": 8.355079411182456e-06, "loss": 0.5316, "step": 2785 }, { "epoch": 5.66260162601626, "grad_norm": 3.914026403669791, "learning_rate": 8.348586056550046e-06, "loss": 0.5468, "step": 2786 }, { "epoch": 5.664634146341464, "grad_norm": 3.505890464139434, "learning_rate": 8.342093417656837e-06, "loss": 0.5783, "step": 2787 }, { "epoch": 5.666666666666667, "grad_norm": 4.0655195543925045, "learning_rate": 8.335601497316809e-06, "loss": 0.5849, "step": 2788 }, { "epoch": 5.66869918699187, "grad_norm": 4.5349846996143315, "learning_rate": 8.329110298343624e-06, "loss": 0.6551, "step": 2789 }, { "epoch": 5.670731707317073, "grad_norm": 3.9501176679705456, "learning_rate": 8.322619823550633e-06, "loss": 0.6021, "step": 2790 }, { "epoch": 5.6727642276422765, "grad_norm": 3.7486747205903743, "learning_rate": 8.316130075750868e-06, "loss": 0.5062, "step": 2791 }, { "epoch": 5.67479674796748, "grad_norm": 3.675651717660583, "learning_rate": 8.309641057757052e-06, "loss": 0.4219, "step": 2792 }, { "epoch": 5.676829268292683, "grad_norm": 4.616826856069048, "learning_rate": 8.303152772381593e-06, "loss": 0.7817, "step": 2793 }, { "epoch": 5.678861788617886, "grad_norm": 4.333188483505051, "learning_rate": 8.296665222436575e-06, "loss": 0.5299, "step": 2794 }, { "epoch": 5.680894308943089, "grad_norm": 4.486047173846124, "learning_rate": 8.290178410733771e-06, "loss": 0.5721, "step": 2795 }, { "epoch": 5.682926829268292, "grad_norm": 4.410695113728438, "learning_rate": 8.283692340084623e-06, "loss": 0.7445, "step": 2796 }, { "epoch": 5.684959349593496, "grad_norm": 3.4840107327540353, "learning_rate": 8.27720701330026e-06, "loss": 0.6607, "step": 2797 }, { "epoch": 5.6869918699187, "grad_norm": 4.035346203823392, "learning_rate": 8.270722433191494e-06, "loss": 0.6387, "step": 2798 }, { "epoch": 5.689024390243903, "grad_norm": 4.0804455638884765, "learning_rate": 8.264238602568797e-06, "loss": 0.5846, "step": 2799 }, { "epoch": 5.691056910569106, "grad_norm": 3.5133921728165958, "learning_rate": 8.257755524242333e-06, "loss": 0.6132, "step": 2800 }, { "epoch": 5.693089430894309, "grad_norm": 4.006507216302821, "learning_rate": 8.251273201021926e-06, "loss": 0.5547, "step": 2801 }, { "epoch": 5.695121951219512, "grad_norm": 4.200485641610277, "learning_rate": 8.244791635717085e-06, "loss": 0.5981, "step": 2802 }, { "epoch": 5.6971544715447155, "grad_norm": 4.37789983569244, "learning_rate": 8.23831083113698e-06, "loss": 0.6315, "step": 2803 }, { "epoch": 5.699186991869919, "grad_norm": 3.7915239993352197, "learning_rate": 8.231830790090461e-06, "loss": 0.6272, "step": 2804 }, { "epoch": 5.701219512195122, "grad_norm": 3.738397474281649, "learning_rate": 8.225351515386042e-06, "loss": 0.4507, "step": 2805 }, { "epoch": 5.703252032520325, "grad_norm": 4.399962581743998, "learning_rate": 8.218873009831897e-06, "loss": 0.6414, "step": 2806 }, { "epoch": 5.705284552845528, "grad_norm": 3.9254618128381464, "learning_rate": 8.212395276235881e-06, "loss": 0.5598, "step": 2807 }, { "epoch": 5.7073170731707314, "grad_norm": 3.5834212035418678, "learning_rate": 8.205918317405508e-06, "loss": 0.5111, "step": 2808 }, { "epoch": 5.709349593495935, "grad_norm": 3.8951843378067683, "learning_rate": 8.199442136147959e-06, "loss": 0.6242, "step": 2809 }, { "epoch": 5.711382113821138, "grad_norm": 3.7387118396922703, "learning_rate": 8.192966735270072e-06, "loss": 0.5806, "step": 2810 }, { "epoch": 5.713414634146341, "grad_norm": 4.090683104343949, "learning_rate": 8.186492117578346e-06, "loss": 0.643, "step": 2811 }, { "epoch": 5.715447154471545, "grad_norm": 4.116524297651816, "learning_rate": 8.18001828587895e-06, "loss": 0.7087, "step": 2812 }, { "epoch": 5.717479674796748, "grad_norm": 4.066149368080705, "learning_rate": 8.173545242977703e-06, "loss": 0.4865, "step": 2813 }, { "epoch": 5.719512195121951, "grad_norm": 3.5006554841972206, "learning_rate": 8.167072991680095e-06, "loss": 0.5326, "step": 2814 }, { "epoch": 5.721544715447155, "grad_norm": 3.0249057956239804, "learning_rate": 8.160601534791255e-06, "loss": 0.593, "step": 2815 }, { "epoch": 5.723577235772358, "grad_norm": 3.5932309938177682, "learning_rate": 8.154130875115978e-06, "loss": 0.6658, "step": 2816 }, { "epoch": 5.725609756097561, "grad_norm": 4.48821661174173, "learning_rate": 8.147661015458712e-06, "loss": 0.6366, "step": 2817 }, { "epoch": 5.727642276422764, "grad_norm": 4.033396920743622, "learning_rate": 8.14119195862356e-06, "loss": 0.6592, "step": 2818 }, { "epoch": 5.729674796747967, "grad_norm": 3.345754666749418, "learning_rate": 8.13472370741428e-06, "loss": 0.5675, "step": 2819 }, { "epoch": 5.7317073170731705, "grad_norm": 3.8111665690899983, "learning_rate": 8.12825626463427e-06, "loss": 0.5459, "step": 2820 }, { "epoch": 5.733739837398374, "grad_norm": 3.9729628958015817, "learning_rate": 8.121789633086584e-06, "loss": 0.6341, "step": 2821 }, { "epoch": 5.735772357723577, "grad_norm": 3.825621998228838, "learning_rate": 8.115323815573926e-06, "loss": 0.6043, "step": 2822 }, { "epoch": 5.737804878048781, "grad_norm": 4.0965394592657525, "learning_rate": 8.10885881489865e-06, "loss": 0.6179, "step": 2823 }, { "epoch": 5.739837398373984, "grad_norm": 3.8485481651988636, "learning_rate": 8.102394633862743e-06, "loss": 0.6903, "step": 2824 }, { "epoch": 5.741869918699187, "grad_norm": 3.4886556526854067, "learning_rate": 8.095931275267847e-06, "loss": 0.5696, "step": 2825 }, { "epoch": 5.7439024390243905, "grad_norm": 3.820369776637836, "learning_rate": 8.089468741915252e-06, "loss": 0.5633, "step": 2826 }, { "epoch": 5.745934959349594, "grad_norm": 3.695476091049313, "learning_rate": 8.083007036605878e-06, "loss": 0.5757, "step": 2827 }, { "epoch": 5.747967479674797, "grad_norm": 3.558177460707829, "learning_rate": 8.0765461621403e-06, "loss": 0.5176, "step": 2828 }, { "epoch": 5.75, "grad_norm": 3.8167911319165646, "learning_rate": 8.070086121318714e-06, "loss": 0.4917, "step": 2829 }, { "epoch": 5.752032520325203, "grad_norm": 4.446251191860514, "learning_rate": 8.063626916940971e-06, "loss": 0.584, "step": 2830 }, { "epoch": 5.754065040650406, "grad_norm": 3.82193794745456, "learning_rate": 8.057168551806558e-06, "loss": 0.6046, "step": 2831 }, { "epoch": 5.7560975609756095, "grad_norm": 3.188000180026993, "learning_rate": 8.050711028714589e-06, "loss": 0.5568, "step": 2832 }, { "epoch": 5.758130081300813, "grad_norm": 4.253625140284592, "learning_rate": 8.044254350463827e-06, "loss": 0.5937, "step": 2833 }, { "epoch": 5.760162601626016, "grad_norm": 3.933285543764125, "learning_rate": 8.037798519852649e-06, "loss": 0.6123, "step": 2834 }, { "epoch": 5.762195121951219, "grad_norm": 4.03007208560439, "learning_rate": 8.031343539679082e-06, "loss": 0.6597, "step": 2835 }, { "epoch": 5.764227642276423, "grad_norm": 4.023421748732679, "learning_rate": 8.02488941274078e-06, "loss": 0.5836, "step": 2836 }, { "epoch": 5.766260162601626, "grad_norm": 4.046990287411728, "learning_rate": 8.018436141835024e-06, "loss": 0.5801, "step": 2837 }, { "epoch": 5.7682926829268295, "grad_norm": 3.8109029742380067, "learning_rate": 8.011983729758726e-06, "loss": 0.522, "step": 2838 }, { "epoch": 5.770325203252033, "grad_norm": 3.8031072872536877, "learning_rate": 8.005532179308422e-06, "loss": 0.5688, "step": 2839 }, { "epoch": 5.772357723577236, "grad_norm": 3.7418953801097428, "learning_rate": 7.999081493280283e-06, "loss": 0.5114, "step": 2840 }, { "epoch": 5.774390243902439, "grad_norm": 3.6738664485888064, "learning_rate": 7.9926316744701e-06, "loss": 0.6186, "step": 2841 }, { "epoch": 5.776422764227642, "grad_norm": 3.4788531308332837, "learning_rate": 7.986182725673288e-06, "loss": 0.6196, "step": 2842 }, { "epoch": 5.778455284552845, "grad_norm": 3.024126485423535, "learning_rate": 7.979734649684883e-06, "loss": 0.641, "step": 2843 }, { "epoch": 5.780487804878049, "grad_norm": 4.368867877892656, "learning_rate": 7.973287449299545e-06, "loss": 0.632, "step": 2844 }, { "epoch": 5.782520325203252, "grad_norm": 3.3516907893679773, "learning_rate": 7.966841127311556e-06, "loss": 0.5354, "step": 2845 }, { "epoch": 5.784552845528455, "grad_norm": 3.610684004977063, "learning_rate": 7.960395686514817e-06, "loss": 0.66, "step": 2846 }, { "epoch": 5.786585365853659, "grad_norm": 3.665239472439892, "learning_rate": 7.95395112970285e-06, "loss": 0.6526, "step": 2847 }, { "epoch": 5.788617886178862, "grad_norm": 4.410227136688374, "learning_rate": 7.947507459668784e-06, "loss": 0.5812, "step": 2848 }, { "epoch": 5.790650406504065, "grad_norm": 3.609371857037139, "learning_rate": 7.941064679205369e-06, "loss": 0.6663, "step": 2849 }, { "epoch": 5.7926829268292686, "grad_norm": 3.8469468069424955, "learning_rate": 7.934622791104972e-06, "loss": 0.5955, "step": 2850 }, { "epoch": 5.794715447154472, "grad_norm": 3.9133056718234527, "learning_rate": 7.928181798159569e-06, "loss": 0.5428, "step": 2851 }, { "epoch": 5.796747967479675, "grad_norm": 3.856715719027327, "learning_rate": 7.921741703160758e-06, "loss": 0.6602, "step": 2852 }, { "epoch": 5.798780487804878, "grad_norm": 3.958419074768021, "learning_rate": 7.915302508899733e-06, "loss": 0.6311, "step": 2853 }, { "epoch": 5.800813008130081, "grad_norm": 4.128861349303884, "learning_rate": 7.908864218167306e-06, "loss": 0.5261, "step": 2854 }, { "epoch": 5.8028455284552845, "grad_norm": 4.4033637551210285, "learning_rate": 7.902426833753895e-06, "loss": 0.6649, "step": 2855 }, { "epoch": 5.804878048780488, "grad_norm": 4.1785961538902745, "learning_rate": 7.895990358449533e-06, "loss": 0.5548, "step": 2856 }, { "epoch": 5.806910569105691, "grad_norm": 3.9486537019006995, "learning_rate": 7.889554795043843e-06, "loss": 0.5794, "step": 2857 }, { "epoch": 5.808943089430894, "grad_norm": 3.594823236070886, "learning_rate": 7.883120146326067e-06, "loss": 0.6089, "step": 2858 }, { "epoch": 5.810975609756097, "grad_norm": 4.285034890559545, "learning_rate": 7.876686415085047e-06, "loss": 0.6652, "step": 2859 }, { "epoch": 5.8130081300813, "grad_norm": 3.787844793490848, "learning_rate": 7.87025360410922e-06, "loss": 0.7457, "step": 2860 }, { "epoch": 5.815040650406504, "grad_norm": 4.119579840820023, "learning_rate": 7.86382171618664e-06, "loss": 0.6112, "step": 2861 }, { "epoch": 5.817073170731708, "grad_norm": 3.87007607915065, "learning_rate": 7.857390754104942e-06, "loss": 0.6831, "step": 2862 }, { "epoch": 5.819105691056911, "grad_norm": 4.2623071058484445, "learning_rate": 7.850960720651371e-06, "loss": 0.4409, "step": 2863 }, { "epoch": 5.821138211382114, "grad_norm": 4.085371997872743, "learning_rate": 7.844531618612772e-06, "loss": 0.4623, "step": 2864 }, { "epoch": 5.823170731707317, "grad_norm": 3.678420403480457, "learning_rate": 7.838103450775576e-06, "loss": 0.5294, "step": 2865 }, { "epoch": 5.82520325203252, "grad_norm": 4.142444452891971, "learning_rate": 7.83167621992582e-06, "loss": 0.5818, "step": 2866 }, { "epoch": 5.8272357723577235, "grad_norm": 3.5501692916141017, "learning_rate": 7.825249928849125e-06, "loss": 0.6636, "step": 2867 }, { "epoch": 5.829268292682927, "grad_norm": 3.4257009518660397, "learning_rate": 7.81882458033071e-06, "loss": 0.5282, "step": 2868 }, { "epoch": 5.83130081300813, "grad_norm": 4.133322682956119, "learning_rate": 7.812400177155387e-06, "loss": 0.5937, "step": 2869 }, { "epoch": 5.833333333333333, "grad_norm": 3.6069330810813764, "learning_rate": 7.805976722107557e-06, "loss": 0.6482, "step": 2870 }, { "epoch": 5.835365853658536, "grad_norm": 3.559283348893208, "learning_rate": 7.799554217971206e-06, "loss": 0.5794, "step": 2871 }, { "epoch": 5.83739837398374, "grad_norm": 4.035866028128326, "learning_rate": 7.79313266752991e-06, "loss": 0.5824, "step": 2872 }, { "epoch": 5.8394308943089435, "grad_norm": 3.696303888636798, "learning_rate": 7.786712073566838e-06, "loss": 0.5547, "step": 2873 }, { "epoch": 5.841463414634147, "grad_norm": 4.6358484961116435, "learning_rate": 7.780292438864737e-06, "loss": 0.7279, "step": 2874 }, { "epoch": 5.84349593495935, "grad_norm": 3.499168120459152, "learning_rate": 7.773873766205938e-06, "loss": 0.5918, "step": 2875 }, { "epoch": 5.845528455284553, "grad_norm": 3.4841959316701225, "learning_rate": 7.767456058372362e-06, "loss": 0.5497, "step": 2876 }, { "epoch": 5.847560975609756, "grad_norm": 3.300640498529557, "learning_rate": 7.761039318145501e-06, "loss": 0.6659, "step": 2877 }, { "epoch": 5.849593495934959, "grad_norm": 3.9054207763879094, "learning_rate": 7.754623548306438e-06, "loss": 0.4425, "step": 2878 }, { "epoch": 5.8516260162601625, "grad_norm": 4.245005804780363, "learning_rate": 7.748208751635834e-06, "loss": 0.5794, "step": 2879 }, { "epoch": 5.853658536585366, "grad_norm": 3.662051735011541, "learning_rate": 7.741794930913922e-06, "loss": 0.5557, "step": 2880 }, { "epoch": 5.855691056910569, "grad_norm": 3.9436391997466416, "learning_rate": 7.735382088920516e-06, "loss": 0.5169, "step": 2881 }, { "epoch": 5.857723577235772, "grad_norm": 4.471415155796559, "learning_rate": 7.728970228435004e-06, "loss": 0.6826, "step": 2882 }, { "epoch": 5.859756097560975, "grad_norm": 3.8257945439645464, "learning_rate": 7.722559352236351e-06, "loss": 0.4461, "step": 2883 }, { "epoch": 5.861788617886178, "grad_norm": 3.871096912142702, "learning_rate": 7.7161494631031e-06, "loss": 0.558, "step": 2884 }, { "epoch": 5.863821138211382, "grad_norm": 3.1598378984380826, "learning_rate": 7.70974056381335e-06, "loss": 0.4571, "step": 2885 }, { "epoch": 5.865853658536586, "grad_norm": 4.380208230858509, "learning_rate": 7.703332657144792e-06, "loss": 0.6182, "step": 2886 }, { "epoch": 5.867886178861789, "grad_norm": 3.985715266302571, "learning_rate": 7.696925745874666e-06, "loss": 0.545, "step": 2887 }, { "epoch": 5.869918699186992, "grad_norm": 4.408278486781634, "learning_rate": 7.690519832779799e-06, "loss": 0.7428, "step": 2888 }, { "epoch": 5.871951219512195, "grad_norm": 3.74367184922964, "learning_rate": 7.684114920636578e-06, "loss": 0.5565, "step": 2889 }, { "epoch": 5.873983739837398, "grad_norm": 3.3303663822130813, "learning_rate": 7.67771101222095e-06, "loss": 0.4877, "step": 2890 }, { "epoch": 5.876016260162602, "grad_norm": 3.720951511193761, "learning_rate": 7.671308110308436e-06, "loss": 0.6831, "step": 2891 }, { "epoch": 5.878048780487805, "grad_norm": 3.8790926847124547, "learning_rate": 7.664906217674115e-06, "loss": 0.6113, "step": 2892 }, { "epoch": 5.880081300813008, "grad_norm": 3.9324158817687933, "learning_rate": 7.658505337092631e-06, "loss": 0.5674, "step": 2893 }, { "epoch": 5.882113821138211, "grad_norm": 3.817374132850335, "learning_rate": 7.652105471338197e-06, "loss": 0.5608, "step": 2894 }, { "epoch": 5.884146341463414, "grad_norm": 3.9398795613268023, "learning_rate": 7.645706623184564e-06, "loss": 0.5275, "step": 2895 }, { "epoch": 5.886178861788618, "grad_norm": 4.07117473362583, "learning_rate": 7.639308795405066e-06, "loss": 0.6067, "step": 2896 }, { "epoch": 5.888211382113822, "grad_norm": 3.8485737774110294, "learning_rate": 7.632911990772586e-06, "loss": 0.6977, "step": 2897 }, { "epoch": 5.890243902439025, "grad_norm": 4.306524227389426, "learning_rate": 7.626516212059557e-06, "loss": 0.5432, "step": 2898 }, { "epoch": 5.892276422764228, "grad_norm": 3.823560330572198, "learning_rate": 7.62012146203798e-06, "loss": 0.6525, "step": 2899 }, { "epoch": 5.894308943089431, "grad_norm": 3.930530833016262, "learning_rate": 7.613727743479395e-06, "loss": 0.6458, "step": 2900 }, { "epoch": 5.896341463414634, "grad_norm": 4.2715033059107075, "learning_rate": 7.607335059154908e-06, "loss": 0.5835, "step": 2901 }, { "epoch": 5.8983739837398375, "grad_norm": 4.196375946099025, "learning_rate": 7.600943411835173e-06, "loss": 0.5966, "step": 2902 }, { "epoch": 5.900406504065041, "grad_norm": 3.835191421424086, "learning_rate": 7.594552804290394e-06, "loss": 0.6085, "step": 2903 }, { "epoch": 5.902439024390244, "grad_norm": 3.9772826625757043, "learning_rate": 7.588163239290316e-06, "loss": 0.5631, "step": 2904 }, { "epoch": 5.904471544715447, "grad_norm": 3.768805013878676, "learning_rate": 7.581774719604246e-06, "loss": 0.6021, "step": 2905 }, { "epoch": 5.90650406504065, "grad_norm": 4.082312780119078, "learning_rate": 7.575387248001031e-06, "loss": 0.5691, "step": 2906 }, { "epoch": 5.908536585365853, "grad_norm": 3.722095252910045, "learning_rate": 7.569000827249067e-06, "loss": 0.5106, "step": 2907 }, { "epoch": 5.9105691056910565, "grad_norm": 4.471280836552605, "learning_rate": 7.562615460116289e-06, "loss": 0.5921, "step": 2908 }, { "epoch": 5.91260162601626, "grad_norm": 4.02352538285785, "learning_rate": 7.556231149370181e-06, "loss": 0.6692, "step": 2909 }, { "epoch": 5.914634146341464, "grad_norm": 3.563054084501784, "learning_rate": 7.549847897777762e-06, "loss": 0.5268, "step": 2910 }, { "epoch": 5.916666666666667, "grad_norm": 3.879890227248012, "learning_rate": 7.543465708105599e-06, "loss": 0.5096, "step": 2911 }, { "epoch": 5.91869918699187, "grad_norm": 4.383906581272743, "learning_rate": 7.537084583119802e-06, "loss": 0.5213, "step": 2912 }, { "epoch": 5.920731707317073, "grad_norm": 3.4113840162093756, "learning_rate": 7.530704525586008e-06, "loss": 0.5541, "step": 2913 }, { "epoch": 5.9227642276422765, "grad_norm": 4.608806260491124, "learning_rate": 7.524325538269398e-06, "loss": 0.5402, "step": 2914 }, { "epoch": 5.92479674796748, "grad_norm": 3.9223158066155985, "learning_rate": 7.517947623934687e-06, "loss": 0.6992, "step": 2915 }, { "epoch": 5.926829268292683, "grad_norm": 3.7400472067429367, "learning_rate": 7.511570785346129e-06, "loss": 0.5025, "step": 2916 }, { "epoch": 5.928861788617886, "grad_norm": 3.7550676016293765, "learning_rate": 7.505195025267512e-06, "loss": 0.5524, "step": 2917 }, { "epoch": 5.930894308943089, "grad_norm": 3.687598044262704, "learning_rate": 7.498820346462145e-06, "loss": 0.5808, "step": 2918 }, { "epoch": 5.932926829268292, "grad_norm": 4.28193839307765, "learning_rate": 7.492446751692887e-06, "loss": 0.5965, "step": 2919 }, { "epoch": 5.934959349593496, "grad_norm": 4.224358495886196, "learning_rate": 7.486074243722109e-06, "loss": 0.6729, "step": 2920 }, { "epoch": 5.9369918699187, "grad_norm": 3.4649833440620235, "learning_rate": 7.4797028253117235e-06, "loss": 0.5393, "step": 2921 }, { "epoch": 5.939024390243903, "grad_norm": 4.4932356553301815, "learning_rate": 7.473332499223169e-06, "loss": 0.6024, "step": 2922 }, { "epoch": 5.941056910569106, "grad_norm": 3.118500405865074, "learning_rate": 7.466963268217402e-06, "loss": 0.5105, "step": 2923 }, { "epoch": 5.943089430894309, "grad_norm": 4.3431146652088914, "learning_rate": 7.460595135054916e-06, "loss": 0.5658, "step": 2924 }, { "epoch": 5.945121951219512, "grad_norm": 4.0153075118534005, "learning_rate": 7.454228102495718e-06, "loss": 0.506, "step": 2925 }, { "epoch": 5.9471544715447155, "grad_norm": 4.446320964877436, "learning_rate": 7.447862173299346e-06, "loss": 0.6465, "step": 2926 }, { "epoch": 5.949186991869919, "grad_norm": 3.1874874102426207, "learning_rate": 7.441497350224861e-06, "loss": 0.5863, "step": 2927 }, { "epoch": 5.951219512195122, "grad_norm": 3.961432923187537, "learning_rate": 7.435133636030831e-06, "loss": 0.6518, "step": 2928 }, { "epoch": 5.953252032520325, "grad_norm": 3.98528595634835, "learning_rate": 7.428771033475363e-06, "loss": 0.6572, "step": 2929 }, { "epoch": 5.955284552845528, "grad_norm": 3.8073404982092365, "learning_rate": 7.422409545316062e-06, "loss": 0.5738, "step": 2930 }, { "epoch": 5.9573170731707314, "grad_norm": 3.7418243850546213, "learning_rate": 7.416049174310073e-06, "loss": 0.4424, "step": 2931 }, { "epoch": 5.959349593495935, "grad_norm": 4.128054102151345, "learning_rate": 7.4096899232140295e-06, "loss": 0.7321, "step": 2932 }, { "epoch": 5.961382113821138, "grad_norm": 4.649110999364099, "learning_rate": 7.403331794784103e-06, "loss": 0.6664, "step": 2933 }, { "epoch": 5.963414634146341, "grad_norm": 3.132362637298108, "learning_rate": 7.396974791775967e-06, "loss": 0.4762, "step": 2934 }, { "epoch": 5.965447154471545, "grad_norm": 4.362763622112354, "learning_rate": 7.390618916944813e-06, "loss": 0.6324, "step": 2935 }, { "epoch": 5.967479674796748, "grad_norm": 3.7388180813696064, "learning_rate": 7.384264173045339e-06, "loss": 0.5832, "step": 2936 }, { "epoch": 5.969512195121951, "grad_norm": 4.213330617660272, "learning_rate": 7.377910562831749e-06, "loss": 0.436, "step": 2937 }, { "epoch": 5.971544715447155, "grad_norm": 3.178096002000682, "learning_rate": 7.371558089057764e-06, "loss": 0.6041, "step": 2938 }, { "epoch": 5.973577235772358, "grad_norm": 4.4051930926426595, "learning_rate": 7.36520675447661e-06, "loss": 0.6791, "step": 2939 }, { "epoch": 5.975609756097561, "grad_norm": 3.9525951560688677, "learning_rate": 7.358856561841021e-06, "loss": 0.583, "step": 2940 }, { "epoch": 5.977642276422764, "grad_norm": 3.7335171083054606, "learning_rate": 7.352507513903232e-06, "loss": 0.5599, "step": 2941 }, { "epoch": 5.979674796747967, "grad_norm": 3.775666254468488, "learning_rate": 7.346159613414978e-06, "loss": 0.5681, "step": 2942 }, { "epoch": 5.9817073170731705, "grad_norm": 4.2987668701386, "learning_rate": 7.339812863127507e-06, "loss": 0.649, "step": 2943 }, { "epoch": 5.983739837398374, "grad_norm": 3.9635569589940594, "learning_rate": 7.333467265791563e-06, "loss": 0.5689, "step": 2944 }, { "epoch": 5.985772357723577, "grad_norm": 4.018905859856519, "learning_rate": 7.327122824157395e-06, "loss": 0.6354, "step": 2945 }, { "epoch": 5.987804878048781, "grad_norm": 3.553152852694179, "learning_rate": 7.32077954097474e-06, "loss": 0.6266, "step": 2946 }, { "epoch": 5.989837398373984, "grad_norm": 4.544534988802589, "learning_rate": 7.314437418992843e-06, "loss": 0.6971, "step": 2947 }, { "epoch": 5.991869918699187, "grad_norm": 3.701679808554489, "learning_rate": 7.308096460960441e-06, "loss": 0.5103, "step": 2948 }, { "epoch": 5.9939024390243905, "grad_norm": 4.704748094353002, "learning_rate": 7.301756669625769e-06, "loss": 0.6781, "step": 2949 }, { "epoch": 5.995934959349594, "grad_norm": 3.9302399928853946, "learning_rate": 7.295418047736561e-06, "loss": 0.5744, "step": 2950 }, { "epoch": 5.997967479674797, "grad_norm": 4.2539863361384675, "learning_rate": 7.28908059804003e-06, "loss": 0.6801, "step": 2951 }, { "epoch": 6.0, "grad_norm": 3.976609975932851, "learning_rate": 7.282744323282895e-06, "loss": 0.654, "step": 2952 }, { "epoch": 6.0, "eval_loss": 0.9610751271247864, "eval_runtime": 132.3801, "eval_samples_per_second": 7.433, "eval_steps_per_second": 0.929, "step": 2952 }, { "epoch": 6.002032520325203, "grad_norm": 3.349175917053979, "learning_rate": 7.276409226211358e-06, "loss": 0.5156, "step": 2953 }, { "epoch": 6.004065040650406, "grad_norm": 3.600961362665429, "learning_rate": 7.270075309571114e-06, "loss": 0.5347, "step": 2954 }, { "epoch": 6.0060975609756095, "grad_norm": 3.735351483827181, "learning_rate": 7.263742576107349e-06, "loss": 0.6367, "step": 2955 }, { "epoch": 6.008130081300813, "grad_norm": 3.5627733485965742, "learning_rate": 7.2574110285647244e-06, "loss": 0.4966, "step": 2956 }, { "epoch": 6.010162601626016, "grad_norm": 3.937900211892577, "learning_rate": 7.251080669687403e-06, "loss": 0.5364, "step": 2957 }, { "epoch": 6.012195121951219, "grad_norm": 3.46454770597545, "learning_rate": 7.244751502219021e-06, "loss": 0.3741, "step": 2958 }, { "epoch": 6.014227642276423, "grad_norm": 2.977687580537351, "learning_rate": 7.238423528902702e-06, "loss": 0.4089, "step": 2959 }, { "epoch": 6.016260162601626, "grad_norm": 3.52162701829753, "learning_rate": 7.232096752481061e-06, "loss": 0.5345, "step": 2960 }, { "epoch": 6.0182926829268295, "grad_norm": 3.3226447835069437, "learning_rate": 7.2257711756961735e-06, "loss": 0.565, "step": 2961 }, { "epoch": 6.020325203252033, "grad_norm": 3.4288184719335484, "learning_rate": 7.2194468012896155e-06, "loss": 0.4088, "step": 2962 }, { "epoch": 6.022357723577236, "grad_norm": 4.4840876771242595, "learning_rate": 7.21312363200243e-06, "loss": 0.4216, "step": 2963 }, { "epoch": 6.024390243902439, "grad_norm": 3.624153863250077, "learning_rate": 7.206801670575145e-06, "loss": 0.4441, "step": 2964 }, { "epoch": 6.026422764227642, "grad_norm": 3.5103728033112005, "learning_rate": 7.200480919747756e-06, "loss": 0.4629, "step": 2965 }, { "epoch": 6.028455284552845, "grad_norm": 4.30169855960421, "learning_rate": 7.194161382259742e-06, "loss": 0.5831, "step": 2966 }, { "epoch": 6.030487804878049, "grad_norm": 4.462796574066753, "learning_rate": 7.187843060850055e-06, "loss": 0.5338, "step": 2967 }, { "epoch": 6.032520325203252, "grad_norm": 4.051212742992804, "learning_rate": 7.181525958257116e-06, "loss": 0.5423, "step": 2968 }, { "epoch": 6.034552845528455, "grad_norm": 3.763787628006617, "learning_rate": 7.175210077218824e-06, "loss": 0.5202, "step": 2969 }, { "epoch": 6.036585365853658, "grad_norm": 3.764048109917195, "learning_rate": 7.168895420472537e-06, "loss": 0.4985, "step": 2970 }, { "epoch": 6.038617886178862, "grad_norm": 4.244951961391266, "learning_rate": 7.162581990755094e-06, "loss": 0.4747, "step": 2971 }, { "epoch": 6.040650406504065, "grad_norm": 4.31635317668806, "learning_rate": 7.156269790802801e-06, "loss": 0.556, "step": 2972 }, { "epoch": 6.0426829268292686, "grad_norm": 4.025660144926452, "learning_rate": 7.149958823351424e-06, "loss": 0.3743, "step": 2973 }, { "epoch": 6.044715447154472, "grad_norm": 3.5451754675103913, "learning_rate": 7.1436490911362046e-06, "loss": 0.4949, "step": 2974 }, { "epoch": 6.046747967479675, "grad_norm": 3.7367452239946424, "learning_rate": 7.137340596891833e-06, "loss": 0.5499, "step": 2975 }, { "epoch": 6.048780487804878, "grad_norm": 3.832952332243938, "learning_rate": 7.131033343352483e-06, "loss": 0.6339, "step": 2976 }, { "epoch": 6.050813008130081, "grad_norm": 4.053121611065836, "learning_rate": 7.124727333251775e-06, "loss": 0.6352, "step": 2977 }, { "epoch": 6.0528455284552845, "grad_norm": 4.302264273234119, "learning_rate": 7.118422569322804e-06, "loss": 0.5049, "step": 2978 }, { "epoch": 6.054878048780488, "grad_norm": 4.309000823273214, "learning_rate": 7.11211905429811e-06, "loss": 0.454, "step": 2979 }, { "epoch": 6.056910569105691, "grad_norm": 4.186030180662754, "learning_rate": 7.105816790909699e-06, "loss": 0.4266, "step": 2980 }, { "epoch": 6.058943089430894, "grad_norm": 4.110223464254434, "learning_rate": 7.099515781889037e-06, "loss": 0.5683, "step": 2981 }, { "epoch": 6.060975609756097, "grad_norm": 4.209517292171371, "learning_rate": 7.093216029967044e-06, "loss": 0.4852, "step": 2982 }, { "epoch": 6.063008130081301, "grad_norm": 3.134169839449209, "learning_rate": 7.0869175378740985e-06, "loss": 0.5397, "step": 2983 }, { "epoch": 6.065040650406504, "grad_norm": 3.831336756756402, "learning_rate": 7.080620308340024e-06, "loss": 0.5175, "step": 2984 }, { "epoch": 6.067073170731708, "grad_norm": 3.705567498656212, "learning_rate": 7.0743243440941015e-06, "loss": 0.5241, "step": 2985 }, { "epoch": 6.069105691056911, "grad_norm": 3.785549468874142, "learning_rate": 7.0680296478650675e-06, "loss": 0.5868, "step": 2986 }, { "epoch": 6.071138211382114, "grad_norm": 4.415840600380693, "learning_rate": 7.0617362223811055e-06, "loss": 0.5761, "step": 2987 }, { "epoch": 6.073170731707317, "grad_norm": 4.900533946861144, "learning_rate": 7.055444070369852e-06, "loss": 0.5385, "step": 2988 }, { "epoch": 6.07520325203252, "grad_norm": 4.9051038954923465, "learning_rate": 7.049153194558382e-06, "loss": 0.6272, "step": 2989 }, { "epoch": 6.0772357723577235, "grad_norm": 4.6323504165046545, "learning_rate": 7.0428635976732285e-06, "loss": 0.633, "step": 2990 }, { "epoch": 6.079268292682927, "grad_norm": 3.780420498595647, "learning_rate": 7.0365752824403585e-06, "loss": 0.4404, "step": 2991 }, { "epoch": 6.08130081300813, "grad_norm": 4.210762803040122, "learning_rate": 7.0302882515852025e-06, "loss": 0.4667, "step": 2992 }, { "epoch": 6.083333333333333, "grad_norm": 4.108779125666895, "learning_rate": 7.024002507832608e-06, "loss": 0.5273, "step": 2993 }, { "epoch": 6.085365853658536, "grad_norm": 4.70128376001504, "learning_rate": 7.017718053906886e-06, "loss": 0.5931, "step": 2994 }, { "epoch": 6.08739837398374, "grad_norm": 3.959444264154983, "learning_rate": 7.011434892531781e-06, "loss": 0.4455, "step": 2995 }, { "epoch": 6.0894308943089435, "grad_norm": 4.597604298290493, "learning_rate": 7.005153026430476e-06, "loss": 0.5817, "step": 2996 }, { "epoch": 6.091463414634147, "grad_norm": 4.154377765711609, "learning_rate": 6.998872458325599e-06, "loss": 0.5533, "step": 2997 }, { "epoch": 6.09349593495935, "grad_norm": 3.945335498068827, "learning_rate": 6.992593190939203e-06, "loss": 0.5009, "step": 2998 }, { "epoch": 6.095528455284553, "grad_norm": 4.920583635425672, "learning_rate": 6.986315226992789e-06, "loss": 0.6464, "step": 2999 }, { "epoch": 6.097560975609756, "grad_norm": 4.284630311676868, "learning_rate": 6.980038569207291e-06, "loss": 0.4333, "step": 3000 }, { "epoch": 6.099593495934959, "grad_norm": 4.8489886978766865, "learning_rate": 6.973763220303069e-06, "loss": 0.5754, "step": 3001 }, { "epoch": 6.1016260162601625, "grad_norm": 4.090324951684486, "learning_rate": 6.967489182999929e-06, "loss": 0.5171, "step": 3002 }, { "epoch": 6.103658536585366, "grad_norm": 4.210685533630323, "learning_rate": 6.961216460017094e-06, "loss": 0.5043, "step": 3003 }, { "epoch": 6.105691056910569, "grad_norm": 4.405899798168565, "learning_rate": 6.954945054073228e-06, "loss": 0.5091, "step": 3004 }, { "epoch": 6.107723577235772, "grad_norm": 3.942898644138106, "learning_rate": 6.9486749678864196e-06, "loss": 0.4872, "step": 3005 }, { "epoch": 6.109756097560975, "grad_norm": 4.6908757541807455, "learning_rate": 6.942406204174187e-06, "loss": 0.4476, "step": 3006 }, { "epoch": 6.111788617886178, "grad_norm": 3.925281925058619, "learning_rate": 6.936138765653478e-06, "loss": 0.5085, "step": 3007 }, { "epoch": 6.1138211382113825, "grad_norm": 3.930675816553143, "learning_rate": 6.929872655040655e-06, "loss": 0.442, "step": 3008 }, { "epoch": 6.115853658536586, "grad_norm": 3.831467439772286, "learning_rate": 6.9236078750515165e-06, "loss": 0.5959, "step": 3009 }, { "epoch": 6.117886178861789, "grad_norm": 4.735608013146452, "learning_rate": 6.917344428401284e-06, "loss": 0.6033, "step": 3010 }, { "epoch": 6.119918699186992, "grad_norm": 3.9586129406366775, "learning_rate": 6.911082317804594e-06, "loss": 0.4899, "step": 3011 }, { "epoch": 6.121951219512195, "grad_norm": 4.1425435720766695, "learning_rate": 6.904821545975507e-06, "loss": 0.5731, "step": 3012 }, { "epoch": 6.123983739837398, "grad_norm": 4.189198275941813, "learning_rate": 6.898562115627502e-06, "loss": 0.5756, "step": 3013 }, { "epoch": 6.126016260162602, "grad_norm": 4.107670432887526, "learning_rate": 6.89230402947348e-06, "loss": 0.6097, "step": 3014 }, { "epoch": 6.128048780487805, "grad_norm": 4.7753762662101025, "learning_rate": 6.8860472902257606e-06, "loss": 0.4394, "step": 3015 }, { "epoch": 6.130081300813008, "grad_norm": 3.9592905478369, "learning_rate": 6.879791900596077e-06, "loss": 0.5993, "step": 3016 }, { "epoch": 6.132113821138211, "grad_norm": 3.8072546219208983, "learning_rate": 6.873537863295573e-06, "loss": 0.4987, "step": 3017 }, { "epoch": 6.134146341463414, "grad_norm": 4.584199459251497, "learning_rate": 6.8672851810348095e-06, "loss": 0.505, "step": 3018 }, { "epoch": 6.1361788617886175, "grad_norm": 4.456684030484783, "learning_rate": 6.8610338565237645e-06, "loss": 0.4581, "step": 3019 }, { "epoch": 6.138211382113822, "grad_norm": 4.870004100623928, "learning_rate": 6.854783892471823e-06, "loss": 0.498, "step": 3020 }, { "epoch": 6.140243902439025, "grad_norm": 3.991721353742403, "learning_rate": 6.848535291587786e-06, "loss": 0.5415, "step": 3021 }, { "epoch": 6.142276422764228, "grad_norm": 4.90686547954649, "learning_rate": 6.842288056579853e-06, "loss": 0.5555, "step": 3022 }, { "epoch": 6.144308943089431, "grad_norm": 3.7721390826610772, "learning_rate": 6.8360421901556374e-06, "loss": 0.5133, "step": 3023 }, { "epoch": 6.146341463414634, "grad_norm": 3.5545476277012935, "learning_rate": 6.829797695022163e-06, "loss": 0.4693, "step": 3024 }, { "epoch": 6.1483739837398375, "grad_norm": 4.220541428362268, "learning_rate": 6.823554573885858e-06, "loss": 0.4695, "step": 3025 }, { "epoch": 6.150406504065041, "grad_norm": 3.91799543111688, "learning_rate": 6.817312829452547e-06, "loss": 0.5124, "step": 3026 }, { "epoch": 6.152439024390244, "grad_norm": 4.404592828833676, "learning_rate": 6.811072464427466e-06, "loss": 0.3972, "step": 3027 }, { "epoch": 6.154471544715447, "grad_norm": 4.2313416948220395, "learning_rate": 6.804833481515256e-06, "loss": 0.5159, "step": 3028 }, { "epoch": 6.15650406504065, "grad_norm": 4.255132290403546, "learning_rate": 6.798595883419947e-06, "loss": 0.5518, "step": 3029 }, { "epoch": 6.158536585365853, "grad_norm": 4.1009722805933055, "learning_rate": 6.792359672844985e-06, "loss": 0.4834, "step": 3030 }, { "epoch": 6.1605691056910565, "grad_norm": 3.8268936442297834, "learning_rate": 6.786124852493195e-06, "loss": 0.528, "step": 3031 }, { "epoch": 6.16260162601626, "grad_norm": 4.350394790800419, "learning_rate": 6.7798914250668154e-06, "loss": 0.4982, "step": 3032 }, { "epoch": 6.164634146341464, "grad_norm": 4.177286271122697, "learning_rate": 6.773659393267479e-06, "loss": 0.4886, "step": 3033 }, { "epoch": 6.166666666666667, "grad_norm": 4.760378286930278, "learning_rate": 6.767428759796203e-06, "loss": 0.5854, "step": 3034 }, { "epoch": 6.16869918699187, "grad_norm": 4.178571375606512, "learning_rate": 6.761199527353415e-06, "loss": 0.5761, "step": 3035 }, { "epoch": 6.170731707317073, "grad_norm": 3.745959829816931, "learning_rate": 6.7549716986389146e-06, "loss": 0.6029, "step": 3036 }, { "epoch": 6.1727642276422765, "grad_norm": 4.594577473070062, "learning_rate": 6.7487452763519115e-06, "loss": 0.5599, "step": 3037 }, { "epoch": 6.17479674796748, "grad_norm": 5.462075359527265, "learning_rate": 6.7425202631910014e-06, "loss": 0.4862, "step": 3038 }, { "epoch": 6.176829268292683, "grad_norm": 4.240750063967756, "learning_rate": 6.736296661854162e-06, "loss": 0.6064, "step": 3039 }, { "epoch": 6.178861788617886, "grad_norm": 4.525270874761123, "learning_rate": 6.730074475038766e-06, "loss": 0.5263, "step": 3040 }, { "epoch": 6.180894308943089, "grad_norm": 5.205416620227145, "learning_rate": 6.7238537054415695e-06, "loss": 0.5665, "step": 3041 }, { "epoch": 6.182926829268292, "grad_norm": 3.860926374628463, "learning_rate": 6.7176343557587154e-06, "loss": 0.5361, "step": 3042 }, { "epoch": 6.184959349593496, "grad_norm": 4.244561964157803, "learning_rate": 6.711416428685736e-06, "loss": 0.3901, "step": 3043 }, { "epoch": 6.186991869918699, "grad_norm": 3.875882548518198, "learning_rate": 6.7051999269175405e-06, "loss": 0.5165, "step": 3044 }, { "epoch": 6.189024390243903, "grad_norm": 3.9007591929792778, "learning_rate": 6.69898485314842e-06, "loss": 0.5114, "step": 3045 }, { "epoch": 6.191056910569106, "grad_norm": 3.9461802388186515, "learning_rate": 6.6927712100720495e-06, "loss": 0.4036, "step": 3046 }, { "epoch": 6.193089430894309, "grad_norm": 4.220654329166311, "learning_rate": 6.686559000381485e-06, "loss": 0.5824, "step": 3047 }, { "epoch": 6.195121951219512, "grad_norm": 3.853777858818972, "learning_rate": 6.680348226769162e-06, "loss": 0.443, "step": 3048 }, { "epoch": 6.1971544715447155, "grad_norm": 3.6944602086337968, "learning_rate": 6.67413889192689e-06, "loss": 0.3744, "step": 3049 }, { "epoch": 6.199186991869919, "grad_norm": 4.135813256500782, "learning_rate": 6.667930998545855e-06, "loss": 0.5398, "step": 3050 }, { "epoch": 6.201219512195122, "grad_norm": 3.686961601441303, "learning_rate": 6.661724549316619e-06, "loss": 0.3693, "step": 3051 }, { "epoch": 6.203252032520325, "grad_norm": 3.850921948575544, "learning_rate": 6.655519546929121e-06, "loss": 0.4603, "step": 3052 }, { "epoch": 6.205284552845528, "grad_norm": 4.367244995488525, "learning_rate": 6.649315994072669e-06, "loss": 0.5838, "step": 3053 }, { "epoch": 6.2073170731707314, "grad_norm": 4.151382272941495, "learning_rate": 6.643113893435949e-06, "loss": 0.4229, "step": 3054 }, { "epoch": 6.209349593495935, "grad_norm": 4.4531025519264675, "learning_rate": 6.636913247707008e-06, "loss": 0.5325, "step": 3055 }, { "epoch": 6.211382113821138, "grad_norm": 4.212643025095264, "learning_rate": 6.630714059573267e-06, "loss": 0.5524, "step": 3056 }, { "epoch": 6.213414634146342, "grad_norm": 3.989791923124829, "learning_rate": 6.624516331721515e-06, "loss": 0.5878, "step": 3057 }, { "epoch": 6.215447154471545, "grad_norm": 4.076176744747355, "learning_rate": 6.6183200668379176e-06, "loss": 0.5544, "step": 3058 }, { "epoch": 6.217479674796748, "grad_norm": 4.484938431582354, "learning_rate": 6.612125267607988e-06, "loss": 0.555, "step": 3059 }, { "epoch": 6.219512195121951, "grad_norm": 4.258805106506017, "learning_rate": 6.6059319367166165e-06, "loss": 0.5268, "step": 3060 }, { "epoch": 6.221544715447155, "grad_norm": 4.127555964002789, "learning_rate": 6.599740076848053e-06, "loss": 0.5032, "step": 3061 }, { "epoch": 6.223577235772358, "grad_norm": 4.010009256323325, "learning_rate": 6.593549690685914e-06, "loss": 0.4669, "step": 3062 }, { "epoch": 6.225609756097561, "grad_norm": 3.841153001706688, "learning_rate": 6.587360780913177e-06, "loss": 0.3982, "step": 3063 }, { "epoch": 6.227642276422764, "grad_norm": 4.2681993593685, "learning_rate": 6.581173350212169e-06, "loss": 0.6182, "step": 3064 }, { "epoch": 6.229674796747967, "grad_norm": 3.97173328963945, "learning_rate": 6.574987401264588e-06, "loss": 0.3734, "step": 3065 }, { "epoch": 6.2317073170731705, "grad_norm": 4.535001562122179, "learning_rate": 6.568802936751486e-06, "loss": 0.6086, "step": 3066 }, { "epoch": 6.233739837398374, "grad_norm": 4.377000081260027, "learning_rate": 6.56261995935327e-06, "loss": 0.5199, "step": 3067 }, { "epoch": 6.235772357723577, "grad_norm": 4.52297559457693, "learning_rate": 6.55643847174971e-06, "loss": 0.5143, "step": 3068 }, { "epoch": 6.237804878048781, "grad_norm": 3.7997375673661953, "learning_rate": 6.550258476619913e-06, "loss": 0.4414, "step": 3069 }, { "epoch": 6.239837398373984, "grad_norm": 5.0327806558419095, "learning_rate": 6.5440799766423575e-06, "loss": 0.6036, "step": 3070 }, { "epoch": 6.241869918699187, "grad_norm": 3.9426572879225503, "learning_rate": 6.537902974494868e-06, "loss": 0.5714, "step": 3071 }, { "epoch": 6.2439024390243905, "grad_norm": 4.50019393668127, "learning_rate": 6.531727472854617e-06, "loss": 0.4858, "step": 3072 }, { "epoch": 6.245934959349594, "grad_norm": 3.950238728084128, "learning_rate": 6.525553474398123e-06, "loss": 0.3908, "step": 3073 }, { "epoch": 6.247967479674797, "grad_norm": 4.545064636703887, "learning_rate": 6.5193809818012645e-06, "loss": 0.4885, "step": 3074 }, { "epoch": 6.25, "grad_norm": 4.5558212785880885, "learning_rate": 6.513209997739257e-06, "loss": 0.5168, "step": 3075 }, { "epoch": 6.252032520325203, "grad_norm": 4.046114439192485, "learning_rate": 6.507040524886672e-06, "loss": 0.4407, "step": 3076 }, { "epoch": 6.254065040650406, "grad_norm": 4.569508423488142, "learning_rate": 6.500872565917417e-06, "loss": 0.4488, "step": 3077 }, { "epoch": 6.2560975609756095, "grad_norm": 4.326674236915518, "learning_rate": 6.494706123504744e-06, "loss": 0.5277, "step": 3078 }, { "epoch": 6.258130081300813, "grad_norm": 4.200279853989193, "learning_rate": 6.488541200321252e-06, "loss": 0.5372, "step": 3079 }, { "epoch": 6.260162601626016, "grad_norm": 3.988756813172458, "learning_rate": 6.482377799038882e-06, "loss": 0.5111, "step": 3080 }, { "epoch": 6.262195121951219, "grad_norm": 5.032829751478483, "learning_rate": 6.476215922328912e-06, "loss": 0.5686, "step": 3081 }, { "epoch": 6.264227642276423, "grad_norm": 5.210176894666242, "learning_rate": 6.470055572861963e-06, "loss": 0.5568, "step": 3082 }, { "epoch": 6.266260162601626, "grad_norm": 4.465044171683558, "learning_rate": 6.4638967533079876e-06, "loss": 0.4727, "step": 3083 }, { "epoch": 6.2682926829268295, "grad_norm": 4.278414730937865, "learning_rate": 6.45773946633628e-06, "loss": 0.4764, "step": 3084 }, { "epoch": 6.270325203252033, "grad_norm": 4.671400543247926, "learning_rate": 6.4515837146154704e-06, "loss": 0.5698, "step": 3085 }, { "epoch": 6.272357723577236, "grad_norm": 4.239733997047258, "learning_rate": 6.445429500813527e-06, "loss": 0.4604, "step": 3086 }, { "epoch": 6.274390243902439, "grad_norm": 4.694685588234972, "learning_rate": 6.439276827597739e-06, "loss": 0.4491, "step": 3087 }, { "epoch": 6.276422764227642, "grad_norm": 5.235592262698229, "learning_rate": 6.4331256976347434e-06, "loss": 0.4443, "step": 3088 }, { "epoch": 6.278455284552845, "grad_norm": 4.694678297565083, "learning_rate": 6.426976113590494e-06, "loss": 0.5656, "step": 3089 }, { "epoch": 6.280487804878049, "grad_norm": 4.252718363598622, "learning_rate": 6.4208280781302855e-06, "loss": 0.5579, "step": 3090 }, { "epoch": 6.282520325203252, "grad_norm": 4.2025580078934635, "learning_rate": 6.414681593918744e-06, "loss": 0.552, "step": 3091 }, { "epoch": 6.284552845528455, "grad_norm": 4.331546756990268, "learning_rate": 6.408536663619803e-06, "loss": 0.4985, "step": 3092 }, { "epoch": 6.286585365853658, "grad_norm": 5.35036778917872, "learning_rate": 6.402393289896747e-06, "loss": 0.7322, "step": 3093 }, { "epoch": 6.288617886178862, "grad_norm": 4.374388152674622, "learning_rate": 6.39625147541217e-06, "loss": 0.4877, "step": 3094 }, { "epoch": 6.290650406504065, "grad_norm": 4.01378505699353, "learning_rate": 6.390111222827997e-06, "loss": 0.4828, "step": 3095 }, { "epoch": 6.2926829268292686, "grad_norm": 4.07850899135185, "learning_rate": 6.383972534805478e-06, "loss": 0.6036, "step": 3096 }, { "epoch": 6.294715447154472, "grad_norm": 4.79098397950834, "learning_rate": 6.3778354140051755e-06, "loss": 0.4987, "step": 3097 }, { "epoch": 6.296747967479675, "grad_norm": 4.434604919152861, "learning_rate": 6.371699863086982e-06, "loss": 0.5785, "step": 3098 }, { "epoch": 6.298780487804878, "grad_norm": 3.746183372813677, "learning_rate": 6.365565884710104e-06, "loss": 0.3846, "step": 3099 }, { "epoch": 6.300813008130081, "grad_norm": 4.05361370593282, "learning_rate": 6.359433481533074e-06, "loss": 0.388, "step": 3100 }, { "epoch": 6.3028455284552845, "grad_norm": 4.112745000813897, "learning_rate": 6.353302656213728e-06, "loss": 0.513, "step": 3101 }, { "epoch": 6.304878048780488, "grad_norm": 4.046436408717923, "learning_rate": 6.347173411409231e-06, "loss": 0.4683, "step": 3102 }, { "epoch": 6.306910569105691, "grad_norm": 4.478353356653015, "learning_rate": 6.34104574977606e-06, "loss": 0.562, "step": 3103 }, { "epoch": 6.308943089430894, "grad_norm": 3.9839587972950765, "learning_rate": 6.3349196739700024e-06, "loss": 0.5081, "step": 3104 }, { "epoch": 6.310975609756097, "grad_norm": 4.696779855135111, "learning_rate": 6.328795186646164e-06, "loss": 0.6812, "step": 3105 }, { "epoch": 6.3130081300813, "grad_norm": 4.299537605098109, "learning_rate": 6.322672290458952e-06, "loss": 0.404, "step": 3106 }, { "epoch": 6.315040650406504, "grad_norm": 4.837080552374186, "learning_rate": 6.316550988062094e-06, "loss": 0.5682, "step": 3107 }, { "epoch": 6.317073170731708, "grad_norm": 4.046696143272297, "learning_rate": 6.310431282108622e-06, "loss": 0.3958, "step": 3108 }, { "epoch": 6.319105691056911, "grad_norm": 4.1975104144092255, "learning_rate": 6.30431317525088e-06, "loss": 0.4379, "step": 3109 }, { "epoch": 6.321138211382114, "grad_norm": 4.021300515189193, "learning_rate": 6.298196670140519e-06, "loss": 0.4372, "step": 3110 }, { "epoch": 6.323170731707317, "grad_norm": 4.178554338735089, "learning_rate": 6.292081769428482e-06, "loss": 0.4767, "step": 3111 }, { "epoch": 6.32520325203252, "grad_norm": 4.912980458916152, "learning_rate": 6.2859684757650365e-06, "loss": 0.6028, "step": 3112 }, { "epoch": 6.3272357723577235, "grad_norm": 4.708164925144803, "learning_rate": 6.279856791799741e-06, "loss": 0.6781, "step": 3113 }, { "epoch": 6.329268292682927, "grad_norm": 4.679022422923712, "learning_rate": 6.273746720181464e-06, "loss": 0.4786, "step": 3114 }, { "epoch": 6.33130081300813, "grad_norm": 5.6450755032849464, "learning_rate": 6.267638263558372e-06, "loss": 0.5673, "step": 3115 }, { "epoch": 6.333333333333333, "grad_norm": 4.749780863247734, "learning_rate": 6.261531424577923e-06, "loss": 0.4187, "step": 3116 }, { "epoch": 6.335365853658536, "grad_norm": 4.194320456305172, "learning_rate": 6.255426205886885e-06, "loss": 0.6591, "step": 3117 }, { "epoch": 6.33739837398374, "grad_norm": 4.677253273976444, "learning_rate": 6.249322610131324e-06, "loss": 0.4485, "step": 3118 }, { "epoch": 6.3394308943089435, "grad_norm": 4.7525221816931165, "learning_rate": 6.2432206399566e-06, "loss": 0.5486, "step": 3119 }, { "epoch": 6.341463414634147, "grad_norm": 3.9177576238547296, "learning_rate": 6.2371202980073596e-06, "loss": 0.4512, "step": 3120 }, { "epoch": 6.34349593495935, "grad_norm": 3.9660140086323334, "learning_rate": 6.231021586927559e-06, "loss": 0.5521, "step": 3121 }, { "epoch": 6.345528455284553, "grad_norm": 3.789856723380731, "learning_rate": 6.224924509360435e-06, "loss": 0.3516, "step": 3122 }, { "epoch": 6.347560975609756, "grad_norm": 4.521227128052981, "learning_rate": 6.218829067948524e-06, "loss": 0.6347, "step": 3123 }, { "epoch": 6.349593495934959, "grad_norm": 4.226377642162734, "learning_rate": 6.212735265333655e-06, "loss": 0.4555, "step": 3124 }, { "epoch": 6.3516260162601625, "grad_norm": 3.682146914775004, "learning_rate": 6.206643104156933e-06, "loss": 0.5015, "step": 3125 }, { "epoch": 6.353658536585366, "grad_norm": 3.794988875739594, "learning_rate": 6.20055258705877e-06, "loss": 0.5581, "step": 3126 }, { "epoch": 6.355691056910569, "grad_norm": 4.64961461283522, "learning_rate": 6.194463716678851e-06, "loss": 0.6366, "step": 3127 }, { "epoch": 6.357723577235772, "grad_norm": 3.89890376712068, "learning_rate": 6.188376495656156e-06, "loss": 0.546, "step": 3128 }, { "epoch": 6.359756097560975, "grad_norm": 4.26844806061673, "learning_rate": 6.182290926628949e-06, "loss": 0.4931, "step": 3129 }, { "epoch": 6.361788617886178, "grad_norm": 4.920409643274296, "learning_rate": 6.176207012234769e-06, "loss": 0.5785, "step": 3130 }, { "epoch": 6.3638211382113825, "grad_norm": 3.7203403923599003, "learning_rate": 6.170124755110451e-06, "loss": 0.4926, "step": 3131 }, { "epoch": 6.365853658536586, "grad_norm": 5.200165933792716, "learning_rate": 6.164044157892102e-06, "loss": 0.5678, "step": 3132 }, { "epoch": 6.367886178861789, "grad_norm": 4.931192716087, "learning_rate": 6.157965223215119e-06, "loss": 0.4677, "step": 3133 }, { "epoch": 6.369918699186992, "grad_norm": 4.033975648233533, "learning_rate": 6.151887953714164e-06, "loss": 0.3663, "step": 3134 }, { "epoch": 6.371951219512195, "grad_norm": 4.050635181373645, "learning_rate": 6.145812352023193e-06, "loss": 0.4151, "step": 3135 }, { "epoch": 6.373983739837398, "grad_norm": 3.9508813071211573, "learning_rate": 6.13973842077543e-06, "loss": 0.483, "step": 3136 }, { "epoch": 6.376016260162602, "grad_norm": 4.017640736771595, "learning_rate": 6.133666162603373e-06, "loss": 0.4624, "step": 3137 }, { "epoch": 6.378048780487805, "grad_norm": 4.628053291036126, "learning_rate": 6.12759558013881e-06, "loss": 0.5076, "step": 3138 }, { "epoch": 6.380081300813008, "grad_norm": 4.14866861800336, "learning_rate": 6.121526676012779e-06, "loss": 0.5043, "step": 3139 }, { "epoch": 6.382113821138211, "grad_norm": 3.4065835366001695, "learning_rate": 6.11545945285561e-06, "loss": 0.6257, "step": 3140 }, { "epoch": 6.384146341463414, "grad_norm": 4.1989479509922445, "learning_rate": 6.1093939132968975e-06, "loss": 0.5794, "step": 3141 }, { "epoch": 6.3861788617886175, "grad_norm": 3.8617074131969997, "learning_rate": 6.1033300599655044e-06, "loss": 0.4922, "step": 3142 }, { "epoch": 6.388211382113822, "grad_norm": 4.0171696542588515, "learning_rate": 6.097267895489569e-06, "loss": 0.505, "step": 3143 }, { "epoch": 6.390243902439025, "grad_norm": 4.084722059892983, "learning_rate": 6.091207422496489e-06, "loss": 0.5624, "step": 3144 }, { "epoch": 6.392276422764228, "grad_norm": 4.2096335791570105, "learning_rate": 6.085148643612935e-06, "loss": 0.4546, "step": 3145 }, { "epoch": 6.394308943089431, "grad_norm": 4.274159093851593, "learning_rate": 6.0790915614648425e-06, "loss": 0.4369, "step": 3146 }, { "epoch": 6.396341463414634, "grad_norm": 4.234890206713012, "learning_rate": 6.073036178677414e-06, "loss": 0.5592, "step": 3147 }, { "epoch": 6.3983739837398375, "grad_norm": 4.317083687561405, "learning_rate": 6.066982497875109e-06, "loss": 0.4721, "step": 3148 }, { "epoch": 6.400406504065041, "grad_norm": 4.605330107110506, "learning_rate": 6.06093052168165e-06, "loss": 0.3928, "step": 3149 }, { "epoch": 6.402439024390244, "grad_norm": 4.48619585765561, "learning_rate": 6.054880252720028e-06, "loss": 0.5783, "step": 3150 }, { "epoch": 6.404471544715447, "grad_norm": 4.552686450879551, "learning_rate": 6.048831693612489e-06, "loss": 0.5641, "step": 3151 }, { "epoch": 6.40650406504065, "grad_norm": 4.930599440481154, "learning_rate": 6.042784846980542e-06, "loss": 0.401, "step": 3152 }, { "epoch": 6.408536585365853, "grad_norm": 4.2046207171748495, "learning_rate": 6.036739715444943e-06, "loss": 0.5976, "step": 3153 }, { "epoch": 6.4105691056910565, "grad_norm": 3.805123421706209, "learning_rate": 6.030696301625716e-06, "loss": 0.5849, "step": 3154 }, { "epoch": 6.41260162601626, "grad_norm": 4.868688296101649, "learning_rate": 6.024654608142136e-06, "loss": 0.6306, "step": 3155 }, { "epoch": 6.414634146341464, "grad_norm": 4.306668237452839, "learning_rate": 6.018614637612733e-06, "loss": 0.5296, "step": 3156 }, { "epoch": 6.416666666666667, "grad_norm": 4.847461751406629, "learning_rate": 6.012576392655295e-06, "loss": 0.4999, "step": 3157 }, { "epoch": 6.41869918699187, "grad_norm": 4.268479908479692, "learning_rate": 6.006539875886848e-06, "loss": 0.5128, "step": 3158 }, { "epoch": 6.420731707317073, "grad_norm": 3.9260279061867234, "learning_rate": 6.000505089923686e-06, "loss": 0.5135, "step": 3159 }, { "epoch": 6.4227642276422765, "grad_norm": 5.04459555206957, "learning_rate": 5.99447203738134e-06, "loss": 0.5154, "step": 3160 }, { "epoch": 6.42479674796748, "grad_norm": 4.628475099364789, "learning_rate": 5.9884407208745975e-06, "loss": 0.4672, "step": 3161 }, { "epoch": 6.426829268292683, "grad_norm": 4.230572938154425, "learning_rate": 5.982411143017495e-06, "loss": 0.5188, "step": 3162 }, { "epoch": 6.428861788617886, "grad_norm": 4.806608714008848, "learning_rate": 5.976383306423302e-06, "loss": 0.4523, "step": 3163 }, { "epoch": 6.430894308943089, "grad_norm": 4.1926188145377585, "learning_rate": 5.9703572137045495e-06, "loss": 0.6183, "step": 3164 }, { "epoch": 6.432926829268292, "grad_norm": 4.75998574015112, "learning_rate": 5.964332867473004e-06, "loss": 0.6591, "step": 3165 }, { "epoch": 6.434959349593496, "grad_norm": 4.547441984220376, "learning_rate": 5.95831027033968e-06, "loss": 0.5005, "step": 3166 }, { "epoch": 6.4369918699187, "grad_norm": 4.29479663910378, "learning_rate": 5.9522894249148256e-06, "loss": 0.6125, "step": 3167 }, { "epoch": 6.439024390243903, "grad_norm": 3.946681251006244, "learning_rate": 5.946270333807937e-06, "loss": 0.5668, "step": 3168 }, { "epoch": 6.441056910569106, "grad_norm": 4.64481308376088, "learning_rate": 5.940252999627751e-06, "loss": 0.5407, "step": 3169 }, { "epoch": 6.443089430894309, "grad_norm": 5.813114736153323, "learning_rate": 5.934237424982236e-06, "loss": 0.5869, "step": 3170 }, { "epoch": 6.445121951219512, "grad_norm": 4.8679246064026955, "learning_rate": 5.928223612478607e-06, "loss": 0.4863, "step": 3171 }, { "epoch": 6.4471544715447155, "grad_norm": 4.376722551439104, "learning_rate": 5.922211564723302e-06, "loss": 0.5007, "step": 3172 }, { "epoch": 6.449186991869919, "grad_norm": 4.198998385096173, "learning_rate": 5.916201284322007e-06, "loss": 0.5126, "step": 3173 }, { "epoch": 6.451219512195122, "grad_norm": 4.735324730634318, "learning_rate": 5.910192773879639e-06, "loss": 0.4871, "step": 3174 }, { "epoch": 6.453252032520325, "grad_norm": 3.8920374226564842, "learning_rate": 5.904186036000343e-06, "loss": 0.5774, "step": 3175 }, { "epoch": 6.455284552845528, "grad_norm": 4.133590991197745, "learning_rate": 5.898181073287504e-06, "loss": 0.4656, "step": 3176 }, { "epoch": 6.4573170731707314, "grad_norm": 4.805201148164741, "learning_rate": 5.892177888343724e-06, "loss": 0.4938, "step": 3177 }, { "epoch": 6.459349593495935, "grad_norm": 4.32661167060832, "learning_rate": 5.886176483770848e-06, "loss": 0.434, "step": 3178 }, { "epoch": 6.461382113821138, "grad_norm": 4.125992123554225, "learning_rate": 5.880176862169946e-06, "loss": 0.4992, "step": 3179 }, { "epoch": 6.463414634146342, "grad_norm": 4.194838756450082, "learning_rate": 5.87417902614131e-06, "loss": 0.5344, "step": 3180 }, { "epoch": 6.465447154471545, "grad_norm": 5.1934249962513634, "learning_rate": 5.868182978284465e-06, "loss": 0.6093, "step": 3181 }, { "epoch": 6.467479674796748, "grad_norm": 3.9240680346267087, "learning_rate": 5.862188721198152e-06, "loss": 0.4942, "step": 3182 }, { "epoch": 6.469512195121951, "grad_norm": 4.279243769415668, "learning_rate": 5.856196257480347e-06, "loss": 0.4201, "step": 3183 }, { "epoch": 6.471544715447155, "grad_norm": 4.539463063854267, "learning_rate": 5.850205589728239e-06, "loss": 0.5107, "step": 3184 }, { "epoch": 6.473577235772358, "grad_norm": 4.701221907514324, "learning_rate": 5.844216720538252e-06, "loss": 0.5037, "step": 3185 }, { "epoch": 6.475609756097561, "grad_norm": 3.748732276877588, "learning_rate": 5.838229652506013e-06, "loss": 0.5397, "step": 3186 }, { "epoch": 6.477642276422764, "grad_norm": 3.907522973071197, "learning_rate": 5.832244388226373e-06, "loss": 0.5088, "step": 3187 }, { "epoch": 6.479674796747967, "grad_norm": 4.357096435398361, "learning_rate": 5.826260930293417e-06, "loss": 0.4456, "step": 3188 }, { "epoch": 6.4817073170731705, "grad_norm": 4.15897570968112, "learning_rate": 5.820279281300427e-06, "loss": 0.6126, "step": 3189 }, { "epoch": 6.483739837398374, "grad_norm": 4.821174294178655, "learning_rate": 5.814299443839911e-06, "loss": 0.4083, "step": 3190 }, { "epoch": 6.485772357723577, "grad_norm": 4.201566543193239, "learning_rate": 5.808321420503591e-06, "loss": 0.5019, "step": 3191 }, { "epoch": 6.487804878048781, "grad_norm": 3.874500518409369, "learning_rate": 5.802345213882396e-06, "loss": 0.5874, "step": 3192 }, { "epoch": 6.489837398373984, "grad_norm": 5.143932539245318, "learning_rate": 5.796370826566481e-06, "loss": 0.4491, "step": 3193 }, { "epoch": 6.491869918699187, "grad_norm": 4.52426345663266, "learning_rate": 5.790398261145203e-06, "loss": 0.5961, "step": 3194 }, { "epoch": 6.4939024390243905, "grad_norm": 4.457734609606392, "learning_rate": 5.7844275202071274e-06, "loss": 0.4699, "step": 3195 }, { "epoch": 6.495934959349594, "grad_norm": 4.89173181702346, "learning_rate": 5.778458606340037e-06, "loss": 0.5741, "step": 3196 }, { "epoch": 6.497967479674797, "grad_norm": 4.787683882260498, "learning_rate": 5.772491522130912e-06, "loss": 0.6521, "step": 3197 }, { "epoch": 6.5, "grad_norm": 5.229813747711314, "learning_rate": 5.766526270165955e-06, "loss": 0.5124, "step": 3198 }, { "epoch": 6.502032520325203, "grad_norm": 4.040388411319542, "learning_rate": 5.76056285303056e-06, "loss": 0.4418, "step": 3199 }, { "epoch": 6.504065040650406, "grad_norm": 4.262729624419289, "learning_rate": 5.754601273309333e-06, "loss": 0.5659, "step": 3200 }, { "epoch": 6.5060975609756095, "grad_norm": 4.157399251643136, "learning_rate": 5.748641533586079e-06, "loss": 0.5208, "step": 3201 }, { "epoch": 6.508130081300813, "grad_norm": 4.593293519629904, "learning_rate": 5.742683636443808e-06, "loss": 0.5842, "step": 3202 }, { "epoch": 6.510162601626016, "grad_norm": 4.338866068156644, "learning_rate": 5.736727584464736e-06, "loss": 0.4162, "step": 3203 }, { "epoch": 6.512195121951219, "grad_norm": 3.979639454656739, "learning_rate": 5.730773380230276e-06, "loss": 0.4779, "step": 3204 }, { "epoch": 6.514227642276423, "grad_norm": 4.309418140786239, "learning_rate": 5.7248210263210344e-06, "loss": 0.4466, "step": 3205 }, { "epoch": 6.516260162601626, "grad_norm": 4.3754110319073725, "learning_rate": 5.7188705253168235e-06, "loss": 0.5287, "step": 3206 }, { "epoch": 6.5182926829268295, "grad_norm": 4.444211371024574, "learning_rate": 5.712921879796645e-06, "loss": 0.4565, "step": 3207 }, { "epoch": 6.520325203252033, "grad_norm": 3.8085750266291454, "learning_rate": 5.70697509233871e-06, "loss": 0.2988, "step": 3208 }, { "epoch": 6.522357723577236, "grad_norm": 3.961490075575646, "learning_rate": 5.701030165520404e-06, "loss": 0.4601, "step": 3209 }, { "epoch": 6.524390243902439, "grad_norm": 5.401147428888839, "learning_rate": 5.6950871019183265e-06, "loss": 0.541, "step": 3210 }, { "epoch": 6.526422764227642, "grad_norm": 3.699876449081611, "learning_rate": 5.689145904108257e-06, "loss": 0.4091, "step": 3211 }, { "epoch": 6.528455284552845, "grad_norm": 4.44812895044382, "learning_rate": 5.683206574665165e-06, "loss": 0.5542, "step": 3212 }, { "epoch": 6.530487804878049, "grad_norm": 4.913521853806264, "learning_rate": 5.6772691161632275e-06, "loss": 0.5676, "step": 3213 }, { "epoch": 6.532520325203252, "grad_norm": 3.636028226907274, "learning_rate": 5.671333531175782e-06, "loss": 0.5221, "step": 3214 }, { "epoch": 6.534552845528455, "grad_norm": 4.272391675255547, "learning_rate": 5.665399822275383e-06, "loss": 0.5231, "step": 3215 }, { "epoch": 6.536585365853659, "grad_norm": 4.265909873095504, "learning_rate": 5.6594679920337514e-06, "loss": 0.3877, "step": 3216 }, { "epoch": 6.538617886178862, "grad_norm": 4.2185751065073775, "learning_rate": 5.653538043021802e-06, "loss": 0.5719, "step": 3217 }, { "epoch": 6.540650406504065, "grad_norm": 4.370842140842092, "learning_rate": 5.647609977809642e-06, "loss": 0.366, "step": 3218 }, { "epoch": 6.5426829268292686, "grad_norm": 4.389175181565978, "learning_rate": 5.6416837989665396e-06, "loss": 0.6318, "step": 3219 }, { "epoch": 6.544715447154472, "grad_norm": 5.143579349922602, "learning_rate": 5.635759509060969e-06, "loss": 0.6515, "step": 3220 }, { "epoch": 6.546747967479675, "grad_norm": 3.8877198720503143, "learning_rate": 5.6298371106605764e-06, "loss": 0.5303, "step": 3221 }, { "epoch": 6.548780487804878, "grad_norm": 3.911071129014838, "learning_rate": 5.6239166063321826e-06, "loss": 0.4532, "step": 3222 }, { "epoch": 6.550813008130081, "grad_norm": 4.128701267814484, "learning_rate": 5.617997998641803e-06, "loss": 0.3508, "step": 3223 }, { "epoch": 6.5528455284552845, "grad_norm": 3.9718202512361525, "learning_rate": 5.612081290154607e-06, "loss": 0.5531, "step": 3224 }, { "epoch": 6.554878048780488, "grad_norm": 4.224490962734024, "learning_rate": 5.606166483434965e-06, "loss": 0.5864, "step": 3225 }, { "epoch": 6.556910569105691, "grad_norm": 4.410023029052403, "learning_rate": 5.600253581046411e-06, "loss": 0.4724, "step": 3226 }, { "epoch": 6.558943089430894, "grad_norm": 4.356811398536568, "learning_rate": 5.594342585551655e-06, "loss": 0.4527, "step": 3227 }, { "epoch": 6.560975609756097, "grad_norm": 4.828225842726849, "learning_rate": 5.58843349951258e-06, "loss": 0.6198, "step": 3228 }, { "epoch": 6.5630081300813, "grad_norm": 4.46676165311679, "learning_rate": 5.582526325490241e-06, "loss": 0.5584, "step": 3229 }, { "epoch": 6.565040650406504, "grad_norm": 4.732200781095536, "learning_rate": 5.576621066044871e-06, "loss": 0.5284, "step": 3230 }, { "epoch": 6.567073170731708, "grad_norm": 4.89353813325015, "learning_rate": 5.570717723735866e-06, "loss": 0.5055, "step": 3231 }, { "epoch": 6.569105691056911, "grad_norm": 4.897306690147238, "learning_rate": 5.564816301121792e-06, "loss": 0.5367, "step": 3232 }, { "epoch": 6.571138211382114, "grad_norm": 4.110625973653487, "learning_rate": 5.558916800760386e-06, "loss": 0.5643, "step": 3233 }, { "epoch": 6.573170731707317, "grad_norm": 4.590933068488552, "learning_rate": 5.5530192252085435e-06, "loss": 0.4835, "step": 3234 }, { "epoch": 6.57520325203252, "grad_norm": 4.972003343213995, "learning_rate": 5.547123577022342e-06, "loss": 0.5201, "step": 3235 }, { "epoch": 6.5772357723577235, "grad_norm": 4.057544218575521, "learning_rate": 5.541229858757011e-06, "loss": 0.5454, "step": 3236 }, { "epoch": 6.579268292682927, "grad_norm": 4.430497458229481, "learning_rate": 5.535338072966943e-06, "loss": 0.6162, "step": 3237 }, { "epoch": 6.58130081300813, "grad_norm": 4.42351897841466, "learning_rate": 5.5294482222057e-06, "loss": 0.4957, "step": 3238 }, { "epoch": 6.583333333333333, "grad_norm": 3.935346114875388, "learning_rate": 5.523560309025997e-06, "loss": 0.5598, "step": 3239 }, { "epoch": 6.585365853658536, "grad_norm": 3.9008153913169936, "learning_rate": 5.517674335979721e-06, "loss": 0.5374, "step": 3240 }, { "epoch": 6.58739837398374, "grad_norm": 4.071012110746756, "learning_rate": 5.511790305617908e-06, "loss": 0.4609, "step": 3241 }, { "epoch": 6.5894308943089435, "grad_norm": 3.998909583317174, "learning_rate": 5.505908220490755e-06, "loss": 0.49, "step": 3242 }, { "epoch": 6.591463414634147, "grad_norm": 4.318887891818455, "learning_rate": 5.500028083147616e-06, "loss": 0.6864, "step": 3243 }, { "epoch": 6.59349593495935, "grad_norm": 4.793724393313551, "learning_rate": 5.494149896136998e-06, "loss": 0.4348, "step": 3244 }, { "epoch": 6.595528455284553, "grad_norm": 4.023477836953532, "learning_rate": 5.488273662006571e-06, "loss": 0.6523, "step": 3245 }, { "epoch": 6.597560975609756, "grad_norm": 4.132706163599626, "learning_rate": 5.482399383303151e-06, "loss": 0.4249, "step": 3246 }, { "epoch": 6.599593495934959, "grad_norm": 4.715319336851772, "learning_rate": 5.47652706257271e-06, "loss": 0.5609, "step": 3247 }, { "epoch": 6.6016260162601625, "grad_norm": 4.720557772398548, "learning_rate": 5.470656702360367e-06, "loss": 0.5138, "step": 3248 }, { "epoch": 6.603658536585366, "grad_norm": 3.7942060127017894, "learning_rate": 5.464788305210394e-06, "loss": 0.5155, "step": 3249 }, { "epoch": 6.605691056910569, "grad_norm": 4.284256374237948, "learning_rate": 5.458921873666216e-06, "loss": 0.532, "step": 3250 }, { "epoch": 6.607723577235772, "grad_norm": 4.046910714525459, "learning_rate": 5.453057410270404e-06, "loss": 0.6088, "step": 3251 }, { "epoch": 6.609756097560975, "grad_norm": 4.341377904209863, "learning_rate": 5.447194917564671e-06, "loss": 0.4401, "step": 3252 }, { "epoch": 6.611788617886178, "grad_norm": 4.370043480639196, "learning_rate": 5.441334398089882e-06, "loss": 0.4053, "step": 3253 }, { "epoch": 6.613821138211382, "grad_norm": 4.122757164108228, "learning_rate": 5.435475854386037e-06, "loss": 0.5074, "step": 3254 }, { "epoch": 6.615853658536586, "grad_norm": 4.64354495180048, "learning_rate": 5.429619288992303e-06, "loss": 0.5665, "step": 3255 }, { "epoch": 6.617886178861789, "grad_norm": 4.3622307655964905, "learning_rate": 5.423764704446954e-06, "loss": 0.5692, "step": 3256 }, { "epoch": 6.619918699186992, "grad_norm": 4.096972073062764, "learning_rate": 5.41791210328744e-06, "loss": 0.4867, "step": 3257 }, { "epoch": 6.621951219512195, "grad_norm": 4.380143535355617, "learning_rate": 5.412061488050327e-06, "loss": 0.5358, "step": 3258 }, { "epoch": 6.623983739837398, "grad_norm": 4.384509256126914, "learning_rate": 5.4062128612713385e-06, "loss": 0.45, "step": 3259 }, { "epoch": 6.626016260162602, "grad_norm": 4.917922754486003, "learning_rate": 5.400366225485326e-06, "loss": 0.5304, "step": 3260 }, { "epoch": 6.628048780487805, "grad_norm": 3.7919733051012763, "learning_rate": 5.3945215832262686e-06, "loss": 0.587, "step": 3261 }, { "epoch": 6.630081300813008, "grad_norm": 4.0367737363618525, "learning_rate": 5.388678937027305e-06, "loss": 0.491, "step": 3262 }, { "epoch": 6.632113821138211, "grad_norm": 4.3878118622035664, "learning_rate": 5.382838289420686e-06, "loss": 0.5882, "step": 3263 }, { "epoch": 6.634146341463414, "grad_norm": 4.695120502009004, "learning_rate": 5.376999642937817e-06, "loss": 0.3966, "step": 3264 }, { "epoch": 6.636178861788618, "grad_norm": 4.785175062479087, "learning_rate": 5.371163000109223e-06, "loss": 0.539, "step": 3265 }, { "epoch": 6.638211382113822, "grad_norm": 4.198051200858448, "learning_rate": 5.3653283634645524e-06, "loss": 0.5212, "step": 3266 }, { "epoch": 6.640243902439025, "grad_norm": 3.6581725399142164, "learning_rate": 5.359495735532607e-06, "loss": 0.46, "step": 3267 }, { "epoch": 6.642276422764228, "grad_norm": 3.961240459363973, "learning_rate": 5.353665118841296e-06, "loss": 0.4716, "step": 3268 }, { "epoch": 6.644308943089431, "grad_norm": 5.00304594512976, "learning_rate": 5.347836515917675e-06, "loss": 0.5989, "step": 3269 }, { "epoch": 6.646341463414634, "grad_norm": 4.231575529554706, "learning_rate": 5.3420099292879156e-06, "loss": 0.5604, "step": 3270 }, { "epoch": 6.6483739837398375, "grad_norm": 3.977005952474678, "learning_rate": 5.336185361477317e-06, "loss": 0.5938, "step": 3271 }, { "epoch": 6.650406504065041, "grad_norm": 4.0165630375691315, "learning_rate": 5.330362815010306e-06, "loss": 0.4376, "step": 3272 }, { "epoch": 6.652439024390244, "grad_norm": 4.591421785925192, "learning_rate": 5.3245422924104265e-06, "loss": 0.4168, "step": 3273 }, { "epoch": 6.654471544715447, "grad_norm": 5.369408372791297, "learning_rate": 5.318723796200365e-06, "loss": 0.6806, "step": 3274 }, { "epoch": 6.65650406504065, "grad_norm": 4.130569658771325, "learning_rate": 5.312907328901898e-06, "loss": 0.5945, "step": 3275 }, { "epoch": 6.658536585365853, "grad_norm": 5.1860517243564415, "learning_rate": 5.307092893035951e-06, "loss": 0.471, "step": 3276 }, { "epoch": 6.6605691056910565, "grad_norm": 3.9611068409208188, "learning_rate": 5.3012804911225555e-06, "loss": 0.5428, "step": 3277 }, { "epoch": 6.66260162601626, "grad_norm": 4.326507820299858, "learning_rate": 5.2954701256808615e-06, "loss": 0.4629, "step": 3278 }, { "epoch": 6.664634146341464, "grad_norm": 4.06534333681788, "learning_rate": 5.289661799229151e-06, "loss": 0.5123, "step": 3279 }, { "epoch": 6.666666666666667, "grad_norm": 4.152145702023607, "learning_rate": 5.2838555142847925e-06, "loss": 0.4667, "step": 3280 }, { "epoch": 6.66869918699187, "grad_norm": 3.969789909839315, "learning_rate": 5.2780512733643e-06, "loss": 0.5712, "step": 3281 }, { "epoch": 6.670731707317073, "grad_norm": 4.231850381634571, "learning_rate": 5.2722490789832854e-06, "loss": 0.4196, "step": 3282 }, { "epoch": 6.6727642276422765, "grad_norm": 4.625782314153474, "learning_rate": 5.266448933656474e-06, "loss": 0.5768, "step": 3283 }, { "epoch": 6.67479674796748, "grad_norm": 4.366399067698886, "learning_rate": 5.260650839897719e-06, "loss": 0.4659, "step": 3284 }, { "epoch": 6.676829268292683, "grad_norm": 3.7929198795022008, "learning_rate": 5.254854800219953e-06, "loss": 0.5093, "step": 3285 }, { "epoch": 6.678861788617886, "grad_norm": 4.233671057036396, "learning_rate": 5.249060817135252e-06, "loss": 0.6555, "step": 3286 }, { "epoch": 6.680894308943089, "grad_norm": 4.46580700320509, "learning_rate": 5.24326889315478e-06, "loss": 0.472, "step": 3287 }, { "epoch": 6.682926829268292, "grad_norm": 4.347653165504464, "learning_rate": 5.237479030788817e-06, "loss": 0.549, "step": 3288 }, { "epoch": 6.684959349593496, "grad_norm": 4.526076364093325, "learning_rate": 5.231691232546744e-06, "loss": 0.463, "step": 3289 }, { "epoch": 6.6869918699187, "grad_norm": 5.0614508136426295, "learning_rate": 5.225905500937047e-06, "loss": 0.6121, "step": 3290 }, { "epoch": 6.689024390243903, "grad_norm": 4.500253479152613, "learning_rate": 5.220121838467327e-06, "loss": 0.5007, "step": 3291 }, { "epoch": 6.691056910569106, "grad_norm": 4.1012077508806355, "learning_rate": 5.214340247644278e-06, "loss": 0.456, "step": 3292 }, { "epoch": 6.693089430894309, "grad_norm": 4.977295937230501, "learning_rate": 5.2085607309737e-06, "loss": 0.5082, "step": 3293 }, { "epoch": 6.695121951219512, "grad_norm": 4.276875444885724, "learning_rate": 5.20278329096049e-06, "loss": 0.571, "step": 3294 }, { "epoch": 6.6971544715447155, "grad_norm": 4.505428246706879, "learning_rate": 5.1970079301086465e-06, "loss": 0.6364, "step": 3295 }, { "epoch": 6.699186991869919, "grad_norm": 3.5255266040082387, "learning_rate": 5.191234650921273e-06, "loss": 0.4977, "step": 3296 }, { "epoch": 6.701219512195122, "grad_norm": 4.839173399992452, "learning_rate": 5.185463455900565e-06, "loss": 0.5553, "step": 3297 }, { "epoch": 6.703252032520325, "grad_norm": 4.081028213797582, "learning_rate": 5.179694347547816e-06, "loss": 0.4986, "step": 3298 }, { "epoch": 6.705284552845528, "grad_norm": 4.062217858977164, "learning_rate": 5.1739273283634124e-06, "loss": 0.5238, "step": 3299 }, { "epoch": 6.7073170731707314, "grad_norm": 3.920744474189046, "learning_rate": 5.168162400846835e-06, "loss": 0.5944, "step": 3300 }, { "epoch": 6.709349593495935, "grad_norm": 3.9096257926802127, "learning_rate": 5.1623995674966685e-06, "loss": 0.5368, "step": 3301 }, { "epoch": 6.711382113821138, "grad_norm": 4.876282874807634, "learning_rate": 5.156638830810579e-06, "loss": 0.5461, "step": 3302 }, { "epoch": 6.713414634146341, "grad_norm": 5.382460533825724, "learning_rate": 5.1508801932853245e-06, "loss": 0.5206, "step": 3303 }, { "epoch": 6.715447154471545, "grad_norm": 4.700063826208891, "learning_rate": 5.145123657416759e-06, "loss": 0.4826, "step": 3304 }, { "epoch": 6.717479674796748, "grad_norm": 5.388498038149773, "learning_rate": 5.139369225699815e-06, "loss": 0.6109, "step": 3305 }, { "epoch": 6.719512195121951, "grad_norm": 4.960241307612662, "learning_rate": 5.133616900628531e-06, "loss": 0.5006, "step": 3306 }, { "epoch": 6.721544715447155, "grad_norm": 4.038390583644165, "learning_rate": 5.127866684696017e-06, "loss": 0.4973, "step": 3307 }, { "epoch": 6.723577235772358, "grad_norm": 4.470010459544859, "learning_rate": 5.122118580394473e-06, "loss": 0.5903, "step": 3308 }, { "epoch": 6.725609756097561, "grad_norm": 4.1124808147624865, "learning_rate": 5.116372590215186e-06, "loss": 0.5547, "step": 3309 }, { "epoch": 6.727642276422764, "grad_norm": 4.6307289269386045, "learning_rate": 5.110628716648519e-06, "loss": 0.6012, "step": 3310 }, { "epoch": 6.729674796747967, "grad_norm": 4.337589078451899, "learning_rate": 5.104886962183934e-06, "loss": 0.5505, "step": 3311 }, { "epoch": 6.7317073170731705, "grad_norm": 4.390692130859257, "learning_rate": 5.099147329309959e-06, "loss": 0.5277, "step": 3312 }, { "epoch": 6.733739837398374, "grad_norm": 5.486252449656362, "learning_rate": 5.093409820514208e-06, "loss": 0.7163, "step": 3313 }, { "epoch": 6.735772357723577, "grad_norm": 4.833820008218865, "learning_rate": 5.087674438283375e-06, "loss": 0.6102, "step": 3314 }, { "epoch": 6.737804878048781, "grad_norm": 4.829862057336588, "learning_rate": 5.081941185103227e-06, "loss": 0.5716, "step": 3315 }, { "epoch": 6.739837398373984, "grad_norm": 4.776451702248643, "learning_rate": 5.076210063458622e-06, "loss": 0.4178, "step": 3316 }, { "epoch": 6.741869918699187, "grad_norm": 4.278934238396806, "learning_rate": 5.07048107583348e-06, "loss": 0.4424, "step": 3317 }, { "epoch": 6.7439024390243905, "grad_norm": 4.199099277888852, "learning_rate": 5.064754224710801e-06, "loss": 0.4954, "step": 3318 }, { "epoch": 6.745934959349594, "grad_norm": 3.897637378861527, "learning_rate": 5.059029512572661e-06, "loss": 0.4046, "step": 3319 }, { "epoch": 6.747967479674797, "grad_norm": 4.366727354802484, "learning_rate": 5.0533069419002e-06, "loss": 0.6384, "step": 3320 }, { "epoch": 6.75, "grad_norm": 4.086166851427506, "learning_rate": 5.04758651517365e-06, "loss": 0.6153, "step": 3321 }, { "epoch": 6.752032520325203, "grad_norm": 4.3616558991294525, "learning_rate": 5.041868234872287e-06, "loss": 0.4398, "step": 3322 }, { "epoch": 6.754065040650406, "grad_norm": 4.7716650526241935, "learning_rate": 5.0361521034744805e-06, "loss": 0.4409, "step": 3323 }, { "epoch": 6.7560975609756095, "grad_norm": 4.3725215779554025, "learning_rate": 5.030438123457655e-06, "loss": 0.517, "step": 3324 }, { "epoch": 6.758130081300813, "grad_norm": 4.2137948880748795, "learning_rate": 5.024726297298302e-06, "loss": 0.4732, "step": 3325 }, { "epoch": 6.760162601626016, "grad_norm": 4.980696248867833, "learning_rate": 5.0190166274719956e-06, "loss": 0.5317, "step": 3326 }, { "epoch": 6.762195121951219, "grad_norm": 4.63608974551669, "learning_rate": 5.013309116453348e-06, "loss": 0.564, "step": 3327 }, { "epoch": 6.764227642276423, "grad_norm": 4.473544169440162, "learning_rate": 5.007603766716063e-06, "loss": 0.5612, "step": 3328 }, { "epoch": 6.766260162601626, "grad_norm": 4.12648514890971, "learning_rate": 5.0019005807328905e-06, "loss": 0.4376, "step": 3329 }, { "epoch": 6.7682926829268295, "grad_norm": 4.012774855635074, "learning_rate": 4.996199560975646e-06, "loss": 0.4749, "step": 3330 }, { "epoch": 6.770325203252033, "grad_norm": 4.63375504065524, "learning_rate": 4.990500709915219e-06, "loss": 0.425, "step": 3331 }, { "epoch": 6.772357723577236, "grad_norm": 4.881742611640375, "learning_rate": 4.984804030021533e-06, "loss": 0.3975, "step": 3332 }, { "epoch": 6.774390243902439, "grad_norm": 4.443442095528215, "learning_rate": 4.979109523763597e-06, "loss": 0.5646, "step": 3333 }, { "epoch": 6.776422764227642, "grad_norm": 4.270600971589364, "learning_rate": 4.973417193609458e-06, "loss": 0.4784, "step": 3334 }, { "epoch": 6.778455284552845, "grad_norm": 4.70880075882931, "learning_rate": 4.967727042026241e-06, "loss": 0.5667, "step": 3335 }, { "epoch": 6.780487804878049, "grad_norm": 4.192790736839031, "learning_rate": 4.962039071480102e-06, "loss": 0.4515, "step": 3336 }, { "epoch": 6.782520325203252, "grad_norm": 4.557800341122823, "learning_rate": 4.9563532844362635e-06, "loss": 0.5776, "step": 3337 }, { "epoch": 6.784552845528455, "grad_norm": 4.493659578390984, "learning_rate": 4.9506696833590125e-06, "loss": 0.3961, "step": 3338 }, { "epoch": 6.786585365853659, "grad_norm": 4.029598461152843, "learning_rate": 4.944988270711668e-06, "loss": 0.4833, "step": 3339 }, { "epoch": 6.788617886178862, "grad_norm": 5.006220909341044, "learning_rate": 4.939309048956622e-06, "loss": 0.6623, "step": 3340 }, { "epoch": 6.790650406504065, "grad_norm": 4.958292572367828, "learning_rate": 4.9336320205552955e-06, "loss": 0.5803, "step": 3341 }, { "epoch": 6.7926829268292686, "grad_norm": 4.386033093365411, "learning_rate": 4.927957187968168e-06, "loss": 0.5792, "step": 3342 }, { "epoch": 6.794715447154472, "grad_norm": 4.514104836670973, "learning_rate": 4.922284553654776e-06, "loss": 0.6554, "step": 3343 }, { "epoch": 6.796747967479675, "grad_norm": 4.452189067796799, "learning_rate": 4.9166141200736885e-06, "loss": 0.4806, "step": 3344 }, { "epoch": 6.798780487804878, "grad_norm": 4.834106638131673, "learning_rate": 4.910945889682539e-06, "loss": 0.555, "step": 3345 }, { "epoch": 6.800813008130081, "grad_norm": 4.808533089417015, "learning_rate": 4.905279864937984e-06, "loss": 0.6964, "step": 3346 }, { "epoch": 6.8028455284552845, "grad_norm": 3.5327187139575726, "learning_rate": 4.899616048295733e-06, "loss": 0.5391, "step": 3347 }, { "epoch": 6.804878048780488, "grad_norm": 5.086595789322731, "learning_rate": 4.89395444221055e-06, "loss": 0.4517, "step": 3348 }, { "epoch": 6.806910569105691, "grad_norm": 4.179733699739046, "learning_rate": 4.888295049136226e-06, "loss": 0.5799, "step": 3349 }, { "epoch": 6.808943089430894, "grad_norm": 4.271442318056784, "learning_rate": 4.8826378715255995e-06, "loss": 0.5842, "step": 3350 }, { "epoch": 6.810975609756097, "grad_norm": 4.3791279292255165, "learning_rate": 4.8769829118305425e-06, "loss": 0.4142, "step": 3351 }, { "epoch": 6.8130081300813, "grad_norm": 4.34690100095291, "learning_rate": 4.871330172501979e-06, "loss": 0.4968, "step": 3352 }, { "epoch": 6.815040650406504, "grad_norm": 4.960967922136584, "learning_rate": 4.865679655989859e-06, "loss": 0.5275, "step": 3353 }, { "epoch": 6.817073170731708, "grad_norm": 4.4172812671643324, "learning_rate": 4.860031364743172e-06, "loss": 0.6692, "step": 3354 }, { "epoch": 6.819105691056911, "grad_norm": 4.091837116079371, "learning_rate": 4.854385301209943e-06, "loss": 0.4629, "step": 3355 }, { "epoch": 6.821138211382114, "grad_norm": 4.071889136717852, "learning_rate": 4.848741467837228e-06, "loss": 0.5397, "step": 3356 }, { "epoch": 6.823170731707317, "grad_norm": 4.57429986566943, "learning_rate": 4.8430998670711295e-06, "loss": 0.6073, "step": 3357 }, { "epoch": 6.82520325203252, "grad_norm": 4.477610810938147, "learning_rate": 4.837460501356767e-06, "loss": 0.4794, "step": 3358 }, { "epoch": 6.8272357723577235, "grad_norm": 4.934421497211445, "learning_rate": 4.831823373138301e-06, "loss": 0.5129, "step": 3359 }, { "epoch": 6.829268292682927, "grad_norm": 4.611333771868542, "learning_rate": 4.826188484858918e-06, "loss": 0.5154, "step": 3360 }, { "epoch": 6.83130081300813, "grad_norm": 4.7169755428439775, "learning_rate": 4.820555838960829e-06, "loss": 0.4707, "step": 3361 }, { "epoch": 6.833333333333333, "grad_norm": 4.8215047797053145, "learning_rate": 4.814925437885286e-06, "loss": 0.6422, "step": 3362 }, { "epoch": 6.835365853658536, "grad_norm": 4.955211332856066, "learning_rate": 4.809297284072559e-06, "loss": 0.6245, "step": 3363 }, { "epoch": 6.83739837398374, "grad_norm": 4.106912387032861, "learning_rate": 4.803671379961945e-06, "loss": 0.4627, "step": 3364 }, { "epoch": 6.8394308943089435, "grad_norm": 5.627572666406062, "learning_rate": 4.798047727991767e-06, "loss": 0.6242, "step": 3365 }, { "epoch": 6.841463414634147, "grad_norm": 4.519515436689513, "learning_rate": 4.792426330599367e-06, "loss": 0.601, "step": 3366 }, { "epoch": 6.84349593495935, "grad_norm": 4.680893294014503, "learning_rate": 4.786807190221122e-06, "loss": 0.4558, "step": 3367 }, { "epoch": 6.845528455284553, "grad_norm": 4.116172070443859, "learning_rate": 4.781190309292421e-06, "loss": 0.4424, "step": 3368 }, { "epoch": 6.847560975609756, "grad_norm": 4.317147672308874, "learning_rate": 4.775575690247674e-06, "loss": 0.6556, "step": 3369 }, { "epoch": 6.849593495934959, "grad_norm": 4.5776951397938745, "learning_rate": 4.769963335520314e-06, "loss": 0.6074, "step": 3370 }, { "epoch": 6.8516260162601625, "grad_norm": 4.192747839756457, "learning_rate": 4.764353247542788e-06, "loss": 0.5556, "step": 3371 }, { "epoch": 6.853658536585366, "grad_norm": 4.611204448948097, "learning_rate": 4.758745428746569e-06, "loss": 0.5017, "step": 3372 }, { "epoch": 6.855691056910569, "grad_norm": 4.952974643694028, "learning_rate": 4.75313988156214e-06, "loss": 0.4114, "step": 3373 }, { "epoch": 6.857723577235772, "grad_norm": 3.72284925481222, "learning_rate": 4.747536608418999e-06, "loss": 0.5738, "step": 3374 }, { "epoch": 6.859756097560975, "grad_norm": 4.973398429077903, "learning_rate": 4.741935611745662e-06, "loss": 0.5321, "step": 3375 }, { "epoch": 6.861788617886178, "grad_norm": 4.582360251265856, "learning_rate": 4.736336893969652e-06, "loss": 0.5224, "step": 3376 }, { "epoch": 6.863821138211382, "grad_norm": 5.157103030829743, "learning_rate": 4.730740457517515e-06, "loss": 0.5663, "step": 3377 }, { "epoch": 6.865853658536586, "grad_norm": 5.309581762154507, "learning_rate": 4.725146304814802e-06, "loss": 0.5698, "step": 3378 }, { "epoch": 6.867886178861789, "grad_norm": 4.684897287117988, "learning_rate": 4.71955443828607e-06, "loss": 0.5774, "step": 3379 }, { "epoch": 6.869918699186992, "grad_norm": 4.262462459464358, "learning_rate": 4.7139648603548925e-06, "loss": 0.5434, "step": 3380 }, { "epoch": 6.871951219512195, "grad_norm": 4.38051978429355, "learning_rate": 4.708377573443844e-06, "loss": 0.5725, "step": 3381 }, { "epoch": 6.873983739837398, "grad_norm": 4.637405959315348, "learning_rate": 4.702792579974521e-06, "loss": 0.5963, "step": 3382 }, { "epoch": 6.876016260162602, "grad_norm": 3.987686930421438, "learning_rate": 4.697209882367499e-06, "loss": 0.5359, "step": 3383 }, { "epoch": 6.878048780487805, "grad_norm": 4.564255176504791, "learning_rate": 4.691629483042387e-06, "loss": 0.7646, "step": 3384 }, { "epoch": 6.880081300813008, "grad_norm": 4.213614390451372, "learning_rate": 4.686051384417783e-06, "loss": 0.5311, "step": 3385 }, { "epoch": 6.882113821138211, "grad_norm": 4.861369866422705, "learning_rate": 4.680475588911283e-06, "loss": 0.5979, "step": 3386 }, { "epoch": 6.884146341463414, "grad_norm": 4.783972512302271, "learning_rate": 4.674902098939507e-06, "loss": 0.4348, "step": 3387 }, { "epoch": 6.886178861788618, "grad_norm": 4.128832650549739, "learning_rate": 4.669330916918043e-06, "loss": 0.4829, "step": 3388 }, { "epoch": 6.888211382113822, "grad_norm": 4.261784851111209, "learning_rate": 4.663762045261511e-06, "loss": 0.5701, "step": 3389 }, { "epoch": 6.890243902439025, "grad_norm": 5.111406143622183, "learning_rate": 4.658195486383509e-06, "loss": 0.5374, "step": 3390 }, { "epoch": 6.892276422764228, "grad_norm": 4.281032415046977, "learning_rate": 4.652631242696637e-06, "loss": 0.532, "step": 3391 }, { "epoch": 6.894308943089431, "grad_norm": 4.817040064237154, "learning_rate": 4.647069316612502e-06, "loss": 0.4614, "step": 3392 }, { "epoch": 6.896341463414634, "grad_norm": 3.9714307862031686, "learning_rate": 4.641509710541686e-06, "loss": 0.4548, "step": 3393 }, { "epoch": 6.8983739837398375, "grad_norm": 4.242661756274824, "learning_rate": 4.6359524268937885e-06, "loss": 0.4337, "step": 3394 }, { "epoch": 6.900406504065041, "grad_norm": 4.3256796797620884, "learning_rate": 4.630397468077385e-06, "loss": 0.5628, "step": 3395 }, { "epoch": 6.902439024390244, "grad_norm": 3.8440945678636345, "learning_rate": 4.624844836500052e-06, "loss": 0.4888, "step": 3396 }, { "epoch": 6.904471544715447, "grad_norm": 4.461407209208128, "learning_rate": 4.619294534568355e-06, "loss": 0.5179, "step": 3397 }, { "epoch": 6.90650406504065, "grad_norm": 4.898980123829806, "learning_rate": 4.613746564687846e-06, "loss": 0.4664, "step": 3398 }, { "epoch": 6.908536585365853, "grad_norm": 5.157309868838986, "learning_rate": 4.608200929263078e-06, "loss": 0.609, "step": 3399 }, { "epoch": 6.9105691056910565, "grad_norm": 4.174682959522855, "learning_rate": 4.60265763069758e-06, "loss": 0.4741, "step": 3400 }, { "epoch": 6.91260162601626, "grad_norm": 4.60471844775255, "learning_rate": 4.597116671393873e-06, "loss": 0.5163, "step": 3401 }, { "epoch": 6.914634146341464, "grad_norm": 4.957331585090218, "learning_rate": 4.591578053753464e-06, "loss": 0.5838, "step": 3402 }, { "epoch": 6.916666666666667, "grad_norm": 4.722934808602006, "learning_rate": 4.586041780176841e-06, "loss": 0.5387, "step": 3403 }, { "epoch": 6.91869918699187, "grad_norm": 4.210733900997508, "learning_rate": 4.580507853063487e-06, "loss": 0.6772, "step": 3404 }, { "epoch": 6.920731707317073, "grad_norm": 4.872354575254526, "learning_rate": 4.5749762748118586e-06, "loss": 0.472, "step": 3405 }, { "epoch": 6.9227642276422765, "grad_norm": 4.239308216174616, "learning_rate": 4.569447047819395e-06, "loss": 0.3733, "step": 3406 }, { "epoch": 6.92479674796748, "grad_norm": 4.684113715476844, "learning_rate": 4.563920174482519e-06, "loss": 0.4777, "step": 3407 }, { "epoch": 6.926829268292683, "grad_norm": 4.441844656499735, "learning_rate": 4.5583956571966295e-06, "loss": 0.4798, "step": 3408 }, { "epoch": 6.928861788617886, "grad_norm": 3.908848659315213, "learning_rate": 4.552873498356113e-06, "loss": 0.4987, "step": 3409 }, { "epoch": 6.930894308943089, "grad_norm": 4.446338500276345, "learning_rate": 4.547353700354326e-06, "loss": 0.5182, "step": 3410 }, { "epoch": 6.932926829268292, "grad_norm": 4.122249001961883, "learning_rate": 4.5418362655836015e-06, "loss": 0.4414, "step": 3411 }, { "epoch": 6.934959349593496, "grad_norm": 4.483761547752582, "learning_rate": 4.5363211964352524e-06, "loss": 0.5481, "step": 3412 }, { "epoch": 6.9369918699187, "grad_norm": 4.836147545937433, "learning_rate": 4.53080849529956e-06, "loss": 0.4928, "step": 3413 }, { "epoch": 6.939024390243903, "grad_norm": 3.983527170529198, "learning_rate": 4.525298164565791e-06, "loss": 0.4651, "step": 3414 }, { "epoch": 6.941056910569106, "grad_norm": 4.214630611331943, "learning_rate": 4.519790206622175e-06, "loss": 0.5614, "step": 3415 }, { "epoch": 6.943089430894309, "grad_norm": 4.249321287812276, "learning_rate": 4.514284623855915e-06, "loss": 0.5958, "step": 3416 }, { "epoch": 6.945121951219512, "grad_norm": 4.279439229479154, "learning_rate": 4.508781418653183e-06, "loss": 0.5266, "step": 3417 }, { "epoch": 6.9471544715447155, "grad_norm": 4.507304073894861, "learning_rate": 4.503280593399123e-06, "loss": 0.4536, "step": 3418 }, { "epoch": 6.949186991869919, "grad_norm": 4.152471548986522, "learning_rate": 4.497782150477852e-06, "loss": 0.4705, "step": 3419 }, { "epoch": 6.951219512195122, "grad_norm": 3.8135973029704653, "learning_rate": 4.4922860922724466e-06, "loss": 0.5149, "step": 3420 }, { "epoch": 6.953252032520325, "grad_norm": 4.853272649290583, "learning_rate": 4.486792421164954e-06, "loss": 0.5351, "step": 3421 }, { "epoch": 6.955284552845528, "grad_norm": 4.858140953078048, "learning_rate": 4.481301139536385e-06, "loss": 0.625, "step": 3422 }, { "epoch": 6.9573170731707314, "grad_norm": 5.054258350457952, "learning_rate": 4.475812249766711e-06, "loss": 0.6134, "step": 3423 }, { "epoch": 6.959349593495935, "grad_norm": 4.978322342232898, "learning_rate": 4.470325754234881e-06, "loss": 0.5449, "step": 3424 }, { "epoch": 6.961382113821138, "grad_norm": 3.6875224229580894, "learning_rate": 4.464841655318791e-06, "loss": 0.3997, "step": 3425 }, { "epoch": 6.963414634146341, "grad_norm": 4.175771549146209, "learning_rate": 4.459359955395305e-06, "loss": 0.6665, "step": 3426 }, { "epoch": 6.965447154471545, "grad_norm": 4.5066497562677705, "learning_rate": 4.453880656840241e-06, "loss": 0.5149, "step": 3427 }, { "epoch": 6.967479674796748, "grad_norm": 3.9241686644780818, "learning_rate": 4.448403762028391e-06, "loss": 0.4566, "step": 3428 }, { "epoch": 6.969512195121951, "grad_norm": 4.189152099831439, "learning_rate": 4.442929273333495e-06, "loss": 0.502, "step": 3429 }, { "epoch": 6.971544715447155, "grad_norm": 4.136909044956409, "learning_rate": 4.4374571931282394e-06, "loss": 0.5292, "step": 3430 }, { "epoch": 6.973577235772358, "grad_norm": 4.307225542515006, "learning_rate": 4.431987523784288e-06, "loss": 0.5235, "step": 3431 }, { "epoch": 6.975609756097561, "grad_norm": 3.9328144933632876, "learning_rate": 4.426520267672244e-06, "loss": 0.5254, "step": 3432 }, { "epoch": 6.977642276422764, "grad_norm": 4.638681769544583, "learning_rate": 4.421055427161677e-06, "loss": 0.4764, "step": 3433 }, { "epoch": 6.979674796747967, "grad_norm": 4.333046682185973, "learning_rate": 4.415593004621104e-06, "loss": 0.4425, "step": 3434 }, { "epoch": 6.9817073170731705, "grad_norm": 4.271009422991603, "learning_rate": 4.4101330024179815e-06, "loss": 0.5195, "step": 3435 }, { "epoch": 6.983739837398374, "grad_norm": 4.758640579689906, "learning_rate": 4.40467542291874e-06, "loss": 0.6739, "step": 3436 }, { "epoch": 6.985772357723577, "grad_norm": 4.244831458415048, "learning_rate": 4.399220268488741e-06, "loss": 0.6143, "step": 3437 }, { "epoch": 6.987804878048781, "grad_norm": 4.385109599436834, "learning_rate": 4.39376754149231e-06, "loss": 0.3859, "step": 3438 }, { "epoch": 6.989837398373984, "grad_norm": 4.672382227789252, "learning_rate": 4.3883172442927095e-06, "loss": 0.594, "step": 3439 }, { "epoch": 6.991869918699187, "grad_norm": 4.0485207508975405, "learning_rate": 4.382869379252152e-06, "loss": 0.4985, "step": 3440 }, { "epoch": 6.9939024390243905, "grad_norm": 4.8282632069704965, "learning_rate": 4.377423948731797e-06, "loss": 0.4636, "step": 3441 }, { "epoch": 6.995934959349594, "grad_norm": 4.078079311684172, "learning_rate": 4.371980955091745e-06, "loss": 0.4924, "step": 3442 }, { "epoch": 6.997967479674797, "grad_norm": 3.9584278912187765, "learning_rate": 4.366540400691054e-06, "loss": 0.5769, "step": 3443 }, { "epoch": 7.0, "grad_norm": 4.658235798628794, "learning_rate": 4.361102287887698e-06, "loss": 0.443, "step": 3444 }, { "epoch": 7.0, "eval_loss": 1.0022398233413696, "eval_runtime": 132.2771, "eval_samples_per_second": 7.439, "eval_steps_per_second": 0.93, "step": 3444 }, { "epoch": 7.002032520325203, "grad_norm": 4.570626877450353, "learning_rate": 4.355666619038622e-06, "loss": 0.5489, "step": 3445 }, { "epoch": 7.004065040650406, "grad_norm": 3.9123246115356576, "learning_rate": 4.350233396499696e-06, "loss": 0.502, "step": 3446 }, { "epoch": 7.0060975609756095, "grad_norm": 4.371743322220585, "learning_rate": 4.344802622625727e-06, "loss": 0.5057, "step": 3447 }, { "epoch": 7.008130081300813, "grad_norm": 4.163538350363716, "learning_rate": 4.339374299770477e-06, "loss": 0.5141, "step": 3448 }, { "epoch": 7.010162601626016, "grad_norm": 3.712014866272812, "learning_rate": 4.333948430286622e-06, "loss": 0.3717, "step": 3449 }, { "epoch": 7.012195121951219, "grad_norm": 4.692812960695796, "learning_rate": 4.3285250165257985e-06, "loss": 0.3807, "step": 3450 }, { "epoch": 7.014227642276423, "grad_norm": 3.9091430536671328, "learning_rate": 4.3231040608385635e-06, "loss": 0.5365, "step": 3451 }, { "epoch": 7.016260162601626, "grad_norm": 3.940076747213321, "learning_rate": 4.31768556557441e-06, "loss": 0.5011, "step": 3452 }, { "epoch": 7.0182926829268295, "grad_norm": 4.79628595514563, "learning_rate": 4.312269533081777e-06, "loss": 0.5256, "step": 3453 }, { "epoch": 7.020325203252033, "grad_norm": 4.0683691671319595, "learning_rate": 4.306855965708016e-06, "loss": 0.3892, "step": 3454 }, { "epoch": 7.022357723577236, "grad_norm": 4.76419566406756, "learning_rate": 4.301444865799428e-06, "loss": 0.4726, "step": 3455 }, { "epoch": 7.024390243902439, "grad_norm": 4.461358458693788, "learning_rate": 4.296036235701235e-06, "loss": 0.5054, "step": 3456 }, { "epoch": 7.026422764227642, "grad_norm": 5.0968391438115015, "learning_rate": 4.290630077757591e-06, "loss": 0.5579, "step": 3457 }, { "epoch": 7.028455284552845, "grad_norm": 4.999244859290969, "learning_rate": 4.285226394311579e-06, "loss": 0.3827, "step": 3458 }, { "epoch": 7.030487804878049, "grad_norm": 4.156191591835996, "learning_rate": 4.279825187705206e-06, "loss": 0.3978, "step": 3459 }, { "epoch": 7.032520325203252, "grad_norm": 4.036030556413671, "learning_rate": 4.274426460279412e-06, "loss": 0.3606, "step": 3460 }, { "epoch": 7.034552845528455, "grad_norm": 4.669091751796393, "learning_rate": 4.2690302143740614e-06, "loss": 0.5175, "step": 3461 }, { "epoch": 7.036585365853658, "grad_norm": 4.3626080332175174, "learning_rate": 4.263636452327937e-06, "loss": 0.3951, "step": 3462 }, { "epoch": 7.038617886178862, "grad_norm": 5.257319756505435, "learning_rate": 4.25824517647875e-06, "loss": 0.5402, "step": 3463 }, { "epoch": 7.040650406504065, "grad_norm": 4.937737714130852, "learning_rate": 4.252856389163128e-06, "loss": 0.388, "step": 3464 }, { "epoch": 7.0426829268292686, "grad_norm": 5.045291336380246, "learning_rate": 4.247470092716635e-06, "loss": 0.5159, "step": 3465 }, { "epoch": 7.044715447154472, "grad_norm": 4.516016883531905, "learning_rate": 4.242086289473738e-06, "loss": 0.4498, "step": 3466 }, { "epoch": 7.046747967479675, "grad_norm": 4.383807604739719, "learning_rate": 4.236704981767833e-06, "loss": 0.4309, "step": 3467 }, { "epoch": 7.048780487804878, "grad_norm": 4.815751705460665, "learning_rate": 4.231326171931231e-06, "loss": 0.5109, "step": 3468 }, { "epoch": 7.050813008130081, "grad_norm": 4.032654390787229, "learning_rate": 4.225949862295161e-06, "loss": 0.3898, "step": 3469 }, { "epoch": 7.0528455284552845, "grad_norm": 5.527439824880064, "learning_rate": 4.2205760551897734e-06, "loss": 0.556, "step": 3470 }, { "epoch": 7.054878048780488, "grad_norm": 4.578174479315058, "learning_rate": 4.215204752944127e-06, "loss": 0.5164, "step": 3471 }, { "epoch": 7.056910569105691, "grad_norm": 4.530577503543605, "learning_rate": 4.209835957886196e-06, "loss": 0.4696, "step": 3472 }, { "epoch": 7.058943089430894, "grad_norm": 4.4942628918600604, "learning_rate": 4.2044696723428716e-06, "loss": 0.4921, "step": 3473 }, { "epoch": 7.060975609756097, "grad_norm": 4.417440348425542, "learning_rate": 4.1991058986399505e-06, "loss": 0.3965, "step": 3474 }, { "epoch": 7.063008130081301, "grad_norm": 4.675412833636336, "learning_rate": 4.193744639102152e-06, "loss": 0.4809, "step": 3475 }, { "epoch": 7.065040650406504, "grad_norm": 3.8509428116455364, "learning_rate": 4.188385896053098e-06, "loss": 0.4195, "step": 3476 }, { "epoch": 7.067073170731708, "grad_norm": 5.2337637480927075, "learning_rate": 4.1830296718153185e-06, "loss": 0.5379, "step": 3477 }, { "epoch": 7.069105691056911, "grad_norm": 4.669103075694707, "learning_rate": 4.1776759687102565e-06, "loss": 0.4742, "step": 3478 }, { "epoch": 7.071138211382114, "grad_norm": 4.423075645054703, "learning_rate": 4.172324789058254e-06, "loss": 0.4964, "step": 3479 }, { "epoch": 7.073170731707317, "grad_norm": 4.313237351101876, "learning_rate": 4.166976135178575e-06, "loss": 0.4247, "step": 3480 }, { "epoch": 7.07520325203252, "grad_norm": 4.900393978239432, "learning_rate": 4.161630009389374e-06, "loss": 0.4511, "step": 3481 }, { "epoch": 7.0772357723577235, "grad_norm": 5.080729101593962, "learning_rate": 4.156286414007716e-06, "loss": 0.3376, "step": 3482 }, { "epoch": 7.079268292682927, "grad_norm": 3.8249691472537997, "learning_rate": 4.150945351349568e-06, "loss": 0.4507, "step": 3483 }, { "epoch": 7.08130081300813, "grad_norm": 4.402304755239338, "learning_rate": 4.1456068237297964e-06, "loss": 0.429, "step": 3484 }, { "epoch": 7.083333333333333, "grad_norm": 3.944480747135148, "learning_rate": 4.140270833462178e-06, "loss": 0.4582, "step": 3485 }, { "epoch": 7.085365853658536, "grad_norm": 4.708188164648544, "learning_rate": 4.1349373828593804e-06, "loss": 0.4394, "step": 3486 }, { "epoch": 7.08739837398374, "grad_norm": 4.498223636527411, "learning_rate": 4.129606474232977e-06, "loss": 0.5145, "step": 3487 }, { "epoch": 7.0894308943089435, "grad_norm": 4.8284442982604086, "learning_rate": 4.124278109893432e-06, "loss": 0.4634, "step": 3488 }, { "epoch": 7.091463414634147, "grad_norm": 5.165137370457802, "learning_rate": 4.118952292150112e-06, "loss": 0.5616, "step": 3489 }, { "epoch": 7.09349593495935, "grad_norm": 4.360252289222028, "learning_rate": 4.113629023311288e-06, "loss": 0.4089, "step": 3490 }, { "epoch": 7.095528455284553, "grad_norm": 5.538891719693715, "learning_rate": 4.108308305684104e-06, "loss": 0.4296, "step": 3491 }, { "epoch": 7.097560975609756, "grad_norm": 4.827364046680363, "learning_rate": 4.10299014157462e-06, "loss": 0.5482, "step": 3492 }, { "epoch": 7.099593495934959, "grad_norm": 5.027919829321984, "learning_rate": 4.097674533287781e-06, "loss": 0.5374, "step": 3493 }, { "epoch": 7.1016260162601625, "grad_norm": 4.78305031791949, "learning_rate": 4.09236148312742e-06, "loss": 0.5666, "step": 3494 }, { "epoch": 7.103658536585366, "grad_norm": 4.214150112968785, "learning_rate": 4.087050993396276e-06, "loss": 0.4706, "step": 3495 }, { "epoch": 7.105691056910569, "grad_norm": 4.243727444035603, "learning_rate": 4.0817430663959536e-06, "loss": 0.428, "step": 3496 }, { "epoch": 7.107723577235772, "grad_norm": 4.6621031003129785, "learning_rate": 4.076437704426972e-06, "loss": 0.4514, "step": 3497 }, { "epoch": 7.109756097560975, "grad_norm": 4.504016824262071, "learning_rate": 4.071134909788723e-06, "loss": 0.384, "step": 3498 }, { "epoch": 7.111788617886178, "grad_norm": 5.13591063527437, "learning_rate": 4.065834684779489e-06, "loss": 0.4323, "step": 3499 }, { "epoch": 7.1138211382113825, "grad_norm": 4.59260136853026, "learning_rate": 4.06053703169645e-06, "loss": 0.6258, "step": 3500 }, { "epoch": 7.115853658536586, "grad_norm": 4.755049042979774, "learning_rate": 4.0552419528356466e-06, "loss": 0.508, "step": 3501 }, { "epoch": 7.117886178861789, "grad_norm": 4.254419452441994, "learning_rate": 4.04994945049203e-06, "loss": 0.4191, "step": 3502 }, { "epoch": 7.119918699186992, "grad_norm": 5.294166477284896, "learning_rate": 4.04465952695942e-06, "loss": 0.4138, "step": 3503 }, { "epoch": 7.121951219512195, "grad_norm": 4.453654370077619, "learning_rate": 4.039372184530521e-06, "loss": 0.4741, "step": 3504 }, { "epoch": 7.123983739837398, "grad_norm": 4.806378100044864, "learning_rate": 4.03408742549692e-06, "loss": 0.4875, "step": 3505 }, { "epoch": 7.126016260162602, "grad_norm": 4.811531042465347, "learning_rate": 4.028805252149081e-06, "loss": 0.5008, "step": 3506 }, { "epoch": 7.128048780487805, "grad_norm": 4.535314812868245, "learning_rate": 4.023525666776357e-06, "loss": 0.4537, "step": 3507 }, { "epoch": 7.130081300813008, "grad_norm": 4.211040677099213, "learning_rate": 4.0182486716669656e-06, "loss": 0.4119, "step": 3508 }, { "epoch": 7.132113821138211, "grad_norm": 4.239452762510259, "learning_rate": 4.01297426910802e-06, "loss": 0.4953, "step": 3509 }, { "epoch": 7.134146341463414, "grad_norm": 5.684594946363081, "learning_rate": 4.007702461385487e-06, "loss": 0.4614, "step": 3510 }, { "epoch": 7.1361788617886175, "grad_norm": 5.1259335327638444, "learning_rate": 4.002433250784223e-06, "loss": 0.488, "step": 3511 }, { "epoch": 7.138211382113822, "grad_norm": 4.071280142219383, "learning_rate": 3.9971666395879605e-06, "loss": 0.507, "step": 3512 }, { "epoch": 7.140243902439025, "grad_norm": 4.3929378344776495, "learning_rate": 3.991902630079295e-06, "loss": 0.4358, "step": 3513 }, { "epoch": 7.142276422764228, "grad_norm": 4.403256293753873, "learning_rate": 3.986641224539711e-06, "loss": 0.4714, "step": 3514 }, { "epoch": 7.144308943089431, "grad_norm": 5.046799329277943, "learning_rate": 3.981382425249543e-06, "loss": 0.4422, "step": 3515 }, { "epoch": 7.146341463414634, "grad_norm": 4.644935217373697, "learning_rate": 3.9761262344880096e-06, "loss": 0.4754, "step": 3516 }, { "epoch": 7.1483739837398375, "grad_norm": 5.175179243027899, "learning_rate": 3.9708726545331984e-06, "loss": 0.5056, "step": 3517 }, { "epoch": 7.150406504065041, "grad_norm": 4.3837964642281095, "learning_rate": 3.965621687662063e-06, "loss": 0.5403, "step": 3518 }, { "epoch": 7.152439024390244, "grad_norm": 5.199291724036894, "learning_rate": 3.960373336150424e-06, "loss": 0.4203, "step": 3519 }, { "epoch": 7.154471544715447, "grad_norm": 4.063686849980109, "learning_rate": 3.9551276022729644e-06, "loss": 0.3686, "step": 3520 }, { "epoch": 7.15650406504065, "grad_norm": 4.0579715971278425, "learning_rate": 3.949884488303245e-06, "loss": 0.5979, "step": 3521 }, { "epoch": 7.158536585365853, "grad_norm": 4.442933602564555, "learning_rate": 3.944643996513681e-06, "loss": 0.4927, "step": 3522 }, { "epoch": 7.1605691056910565, "grad_norm": 5.110236423648602, "learning_rate": 3.939406129175552e-06, "loss": 0.529, "step": 3523 }, { "epoch": 7.16260162601626, "grad_norm": 4.649919751558217, "learning_rate": 3.9341708885590034e-06, "loss": 0.3925, "step": 3524 }, { "epoch": 7.164634146341464, "grad_norm": 4.739492046184995, "learning_rate": 3.928938276933036e-06, "loss": 0.4708, "step": 3525 }, { "epoch": 7.166666666666667, "grad_norm": 4.26436514517467, "learning_rate": 3.923708296565521e-06, "loss": 0.4, "step": 3526 }, { "epoch": 7.16869918699187, "grad_norm": 4.293389328982522, "learning_rate": 3.918480949723183e-06, "loss": 0.3377, "step": 3527 }, { "epoch": 7.170731707317073, "grad_norm": 4.636068104943018, "learning_rate": 3.913256238671607e-06, "loss": 0.2811, "step": 3528 }, { "epoch": 7.1727642276422765, "grad_norm": 5.0242436332891325, "learning_rate": 3.908034165675231e-06, "loss": 0.5469, "step": 3529 }, { "epoch": 7.17479674796748, "grad_norm": 3.7462345569970106, "learning_rate": 3.902814732997351e-06, "loss": 0.5003, "step": 3530 }, { "epoch": 7.176829268292683, "grad_norm": 4.502029937834923, "learning_rate": 3.89759794290013e-06, "loss": 0.446, "step": 3531 }, { "epoch": 7.178861788617886, "grad_norm": 3.9906826402039006, "learning_rate": 3.89238379764457e-06, "loss": 0.4507, "step": 3532 }, { "epoch": 7.180894308943089, "grad_norm": 4.383845621183463, "learning_rate": 3.887172299490536e-06, "loss": 0.4507, "step": 3533 }, { "epoch": 7.182926829268292, "grad_norm": 4.377767927014211, "learning_rate": 3.881963450696742e-06, "loss": 0.3837, "step": 3534 }, { "epoch": 7.184959349593496, "grad_norm": 5.361136090511597, "learning_rate": 3.876757253520751e-06, "loss": 0.512, "step": 3535 }, { "epoch": 7.186991869918699, "grad_norm": 5.575062641664432, "learning_rate": 3.871553710218988e-06, "loss": 0.5473, "step": 3536 }, { "epoch": 7.189024390243903, "grad_norm": 4.3049213587327095, "learning_rate": 3.866352823046718e-06, "loss": 0.39, "step": 3537 }, { "epoch": 7.191056910569106, "grad_norm": 4.483643601110612, "learning_rate": 3.861154594258054e-06, "loss": 0.407, "step": 3538 }, { "epoch": 7.193089430894309, "grad_norm": 4.7624885715251795, "learning_rate": 3.855959026105963e-06, "loss": 0.3767, "step": 3539 }, { "epoch": 7.195121951219512, "grad_norm": 5.029894190031121, "learning_rate": 3.850766120842252e-06, "loss": 0.481, "step": 3540 }, { "epoch": 7.1971544715447155, "grad_norm": 4.5383341680995075, "learning_rate": 3.845575880717582e-06, "loss": 0.4418, "step": 3541 }, { "epoch": 7.199186991869919, "grad_norm": 4.919446237300954, "learning_rate": 3.840388307981455e-06, "loss": 0.5739, "step": 3542 }, { "epoch": 7.201219512195122, "grad_norm": 5.331424188431185, "learning_rate": 3.835203404882214e-06, "loss": 0.5676, "step": 3543 }, { "epoch": 7.203252032520325, "grad_norm": 5.248861256797132, "learning_rate": 3.830021173667048e-06, "loss": 0.5194, "step": 3544 }, { "epoch": 7.205284552845528, "grad_norm": 4.749334294604093, "learning_rate": 3.824841616581985e-06, "loss": 0.4998, "step": 3545 }, { "epoch": 7.2073170731707314, "grad_norm": 4.896548470848824, "learning_rate": 3.8196647358719015e-06, "loss": 0.422, "step": 3546 }, { "epoch": 7.209349593495935, "grad_norm": 4.531801948395949, "learning_rate": 3.8144905337805073e-06, "loss": 0.4367, "step": 3547 }, { "epoch": 7.211382113821138, "grad_norm": 4.977547812118935, "learning_rate": 3.809319012550352e-06, "loss": 0.4734, "step": 3548 }, { "epoch": 7.213414634146342, "grad_norm": 5.165649129533219, "learning_rate": 3.8041501744228247e-06, "loss": 0.4819, "step": 3549 }, { "epoch": 7.215447154471545, "grad_norm": 4.8147048178674074, "learning_rate": 3.798984021638147e-06, "loss": 0.5091, "step": 3550 }, { "epoch": 7.217479674796748, "grad_norm": 5.276954251627964, "learning_rate": 3.7938205564353923e-06, "loss": 0.4612, "step": 3551 }, { "epoch": 7.219512195121951, "grad_norm": 4.686847508350206, "learning_rate": 3.788659781052444e-06, "loss": 0.4053, "step": 3552 }, { "epoch": 7.221544715447155, "grad_norm": 4.5353228051736725, "learning_rate": 3.783501697726042e-06, "loss": 0.5517, "step": 3553 }, { "epoch": 7.223577235772358, "grad_norm": 4.218677649184901, "learning_rate": 3.7783463086917495e-06, "loss": 0.4433, "step": 3554 }, { "epoch": 7.225609756097561, "grad_norm": 4.417783195281609, "learning_rate": 3.773193616183959e-06, "loss": 0.468, "step": 3555 }, { "epoch": 7.227642276422764, "grad_norm": 5.151855981016904, "learning_rate": 3.7680436224359084e-06, "loss": 0.4497, "step": 3556 }, { "epoch": 7.229674796747967, "grad_norm": 5.321906200811783, "learning_rate": 3.762896329679643e-06, "loss": 0.5883, "step": 3557 }, { "epoch": 7.2317073170731705, "grad_norm": 5.029211288792432, "learning_rate": 3.7577517401460608e-06, "loss": 0.5107, "step": 3558 }, { "epoch": 7.233739837398374, "grad_norm": 4.630136508034595, "learning_rate": 3.752609856064874e-06, "loss": 0.4733, "step": 3559 }, { "epoch": 7.235772357723577, "grad_norm": 4.2215320936158145, "learning_rate": 3.747470679664624e-06, "loss": 0.4694, "step": 3560 }, { "epoch": 7.237804878048781, "grad_norm": 4.493180959858634, "learning_rate": 3.74233421317269e-06, "loss": 0.4468, "step": 3561 }, { "epoch": 7.239837398373984, "grad_norm": 5.162734254702832, "learning_rate": 3.7372004588152554e-06, "loss": 0.5745, "step": 3562 }, { "epoch": 7.241869918699187, "grad_norm": 4.618870007673496, "learning_rate": 3.732069418817349e-06, "loss": 0.4199, "step": 3563 }, { "epoch": 7.2439024390243905, "grad_norm": 4.478022770132583, "learning_rate": 3.7269410954028107e-06, "loss": 0.3777, "step": 3564 }, { "epoch": 7.245934959349594, "grad_norm": 4.55026733815407, "learning_rate": 3.7218154907943095e-06, "loss": 0.5432, "step": 3565 }, { "epoch": 7.247967479674797, "grad_norm": 4.508339176214145, "learning_rate": 3.7166926072133302e-06, "loss": 0.4638, "step": 3566 }, { "epoch": 7.25, "grad_norm": 4.5655504320959395, "learning_rate": 3.7115724468801807e-06, "loss": 0.5565, "step": 3567 }, { "epoch": 7.252032520325203, "grad_norm": 4.564874370888226, "learning_rate": 3.706455012013994e-06, "loss": 0.3859, "step": 3568 }, { "epoch": 7.254065040650406, "grad_norm": 4.634748344231804, "learning_rate": 3.7013403048327144e-06, "loss": 0.4425, "step": 3569 }, { "epoch": 7.2560975609756095, "grad_norm": 5.250840236071907, "learning_rate": 3.6962283275531085e-06, "loss": 0.5276, "step": 3570 }, { "epoch": 7.258130081300813, "grad_norm": 4.959599192172502, "learning_rate": 3.6911190823907563e-06, "loss": 0.5256, "step": 3571 }, { "epoch": 7.260162601626016, "grad_norm": 4.637339520370645, "learning_rate": 3.6860125715600513e-06, "loss": 0.5544, "step": 3572 }, { "epoch": 7.262195121951219, "grad_norm": 5.8173674303583764, "learning_rate": 3.680908797274214e-06, "loss": 0.5424, "step": 3573 }, { "epoch": 7.264227642276423, "grad_norm": 4.506093771910051, "learning_rate": 3.675807761745268e-06, "loss": 0.3725, "step": 3574 }, { "epoch": 7.266260162601626, "grad_norm": 4.69228958755581, "learning_rate": 3.6707094671840514e-06, "loss": 0.4193, "step": 3575 }, { "epoch": 7.2682926829268295, "grad_norm": 4.654507264983753, "learning_rate": 3.665613915800217e-06, "loss": 0.4937, "step": 3576 }, { "epoch": 7.270325203252033, "grad_norm": 5.7629796736177745, "learning_rate": 3.660521109802223e-06, "loss": 0.5496, "step": 3577 }, { "epoch": 7.272357723577236, "grad_norm": 4.62565022680904, "learning_rate": 3.65543105139735e-06, "loss": 0.4952, "step": 3578 }, { "epoch": 7.274390243902439, "grad_norm": 4.3808484937976235, "learning_rate": 3.6503437427916743e-06, "loss": 0.4943, "step": 3579 }, { "epoch": 7.276422764227642, "grad_norm": 4.068999979774335, "learning_rate": 3.6452591861900886e-06, "loss": 0.4519, "step": 3580 }, { "epoch": 7.278455284552845, "grad_norm": 4.339488037498146, "learning_rate": 3.640177383796288e-06, "loss": 0.4544, "step": 3581 }, { "epoch": 7.280487804878049, "grad_norm": 4.268551679153814, "learning_rate": 3.6350983378127746e-06, "loss": 0.4223, "step": 3582 }, { "epoch": 7.282520325203252, "grad_norm": 3.8236034405938395, "learning_rate": 3.6300220504408634e-06, "loss": 0.5577, "step": 3583 }, { "epoch": 7.284552845528455, "grad_norm": 5.121166877795666, "learning_rate": 3.6249485238806637e-06, "loss": 0.3949, "step": 3584 }, { "epoch": 7.286585365853658, "grad_norm": 4.71538065253276, "learning_rate": 3.619877760331093e-06, "loss": 0.5569, "step": 3585 }, { "epoch": 7.288617886178862, "grad_norm": 4.9337566899455405, "learning_rate": 3.6148097619898713e-06, "loss": 0.434, "step": 3586 }, { "epoch": 7.290650406504065, "grad_norm": 5.090468374610648, "learning_rate": 3.609744531053515e-06, "loss": 0.4861, "step": 3587 }, { "epoch": 7.2926829268292686, "grad_norm": 5.263624121548225, "learning_rate": 3.6046820697173514e-06, "loss": 0.4232, "step": 3588 }, { "epoch": 7.294715447154472, "grad_norm": 4.224663207932418, "learning_rate": 3.5996223801755002e-06, "loss": 0.436, "step": 3589 }, { "epoch": 7.296747967479675, "grad_norm": 5.236458572801527, "learning_rate": 3.5945654646208806e-06, "loss": 0.3897, "step": 3590 }, { "epoch": 7.298780487804878, "grad_norm": 4.385964532794754, "learning_rate": 3.5895113252452084e-06, "loss": 0.4025, "step": 3591 }, { "epoch": 7.300813008130081, "grad_norm": 4.062279340583674, "learning_rate": 3.5844599642389965e-06, "loss": 0.4458, "step": 3592 }, { "epoch": 7.3028455284552845, "grad_norm": 6.153863043634337, "learning_rate": 3.5794113837915602e-06, "loss": 0.5855, "step": 3593 }, { "epoch": 7.304878048780488, "grad_norm": 5.000087082938412, "learning_rate": 3.574365586091002e-06, "loss": 0.5321, "step": 3594 }, { "epoch": 7.306910569105691, "grad_norm": 5.025460283540934, "learning_rate": 3.56932257332422e-06, "loss": 0.4014, "step": 3595 }, { "epoch": 7.308943089430894, "grad_norm": 5.370659337187138, "learning_rate": 3.564282347676903e-06, "loss": 0.4291, "step": 3596 }, { "epoch": 7.310975609756097, "grad_norm": 5.6556813955246845, "learning_rate": 3.5592449113335414e-06, "loss": 0.4799, "step": 3597 }, { "epoch": 7.3130081300813, "grad_norm": 4.60937625906389, "learning_rate": 3.5542102664774115e-06, "loss": 0.5325, "step": 3598 }, { "epoch": 7.315040650406504, "grad_norm": 5.43455876790116, "learning_rate": 3.549178415290566e-06, "loss": 0.446, "step": 3599 }, { "epoch": 7.317073170731708, "grad_norm": 4.98492605785906, "learning_rate": 3.54414935995387e-06, "loss": 0.4289, "step": 3600 }, { "epoch": 7.319105691056911, "grad_norm": 4.471374764131094, "learning_rate": 3.5391231026469607e-06, "loss": 0.4241, "step": 3601 }, { "epoch": 7.321138211382114, "grad_norm": 4.223615412846455, "learning_rate": 3.5340996455482735e-06, "loss": 0.4788, "step": 3602 }, { "epoch": 7.323170731707317, "grad_norm": 4.110055429020748, "learning_rate": 3.5290789908350254e-06, "loss": 0.4905, "step": 3603 }, { "epoch": 7.32520325203252, "grad_norm": 4.82765880944976, "learning_rate": 3.524061140683206e-06, "loss": 0.3692, "step": 3604 }, { "epoch": 7.3272357723577235, "grad_norm": 5.272815096751536, "learning_rate": 3.519046097267613e-06, "loss": 0.5216, "step": 3605 }, { "epoch": 7.329268292682927, "grad_norm": 4.875078292443692, "learning_rate": 3.51403386276181e-06, "loss": 0.5723, "step": 3606 }, { "epoch": 7.33130081300813, "grad_norm": 5.034072817995457, "learning_rate": 3.5090244393381546e-06, "loss": 0.5118, "step": 3607 }, { "epoch": 7.333333333333333, "grad_norm": 5.232457554834072, "learning_rate": 3.5040178291677816e-06, "loss": 0.5968, "step": 3608 }, { "epoch": 7.335365853658536, "grad_norm": 4.3457540478980405, "learning_rate": 3.499014034420595e-06, "loss": 0.4213, "step": 3609 }, { "epoch": 7.33739837398374, "grad_norm": 4.971971063718951, "learning_rate": 3.4940130572652985e-06, "loss": 0.3925, "step": 3610 }, { "epoch": 7.3394308943089435, "grad_norm": 4.977849170774981, "learning_rate": 3.48901489986936e-06, "loss": 0.4686, "step": 3611 }, { "epoch": 7.341463414634147, "grad_norm": 4.743605238352682, "learning_rate": 3.4840195643990383e-06, "loss": 0.4187, "step": 3612 }, { "epoch": 7.34349593495935, "grad_norm": 4.154811014502775, "learning_rate": 3.4790270530193516e-06, "loss": 0.3542, "step": 3613 }, { "epoch": 7.345528455284553, "grad_norm": 4.083429428585595, "learning_rate": 3.474037367894111e-06, "loss": 0.5437, "step": 3614 }, { "epoch": 7.347560975609756, "grad_norm": 4.868302251970648, "learning_rate": 3.469050511185893e-06, "loss": 0.465, "step": 3615 }, { "epoch": 7.349593495934959, "grad_norm": 5.296623066933855, "learning_rate": 3.464066485056048e-06, "loss": 0.5391, "step": 3616 }, { "epoch": 7.3516260162601625, "grad_norm": 4.489168566121903, "learning_rate": 3.4590852916647122e-06, "loss": 0.443, "step": 3617 }, { "epoch": 7.353658536585366, "grad_norm": 5.150483455567241, "learning_rate": 3.454106933170771e-06, "loss": 0.474, "step": 3618 }, { "epoch": 7.355691056910569, "grad_norm": 5.818269590723202, "learning_rate": 3.449131411731904e-06, "loss": 0.48, "step": 3619 }, { "epoch": 7.357723577235772, "grad_norm": 4.494650135527557, "learning_rate": 3.444158729504549e-06, "loss": 0.5267, "step": 3620 }, { "epoch": 7.359756097560975, "grad_norm": 4.736372573548626, "learning_rate": 3.439188888643912e-06, "loss": 0.4877, "step": 3621 }, { "epoch": 7.361788617886178, "grad_norm": 4.701452225026714, "learning_rate": 3.4342218913039806e-06, "loss": 0.4505, "step": 3622 }, { "epoch": 7.3638211382113825, "grad_norm": 5.089827090235151, "learning_rate": 3.4292577396374893e-06, "loss": 0.5054, "step": 3623 }, { "epoch": 7.365853658536586, "grad_norm": 4.4659252858916485, "learning_rate": 3.4242964357959597e-06, "loss": 0.5729, "step": 3624 }, { "epoch": 7.367886178861789, "grad_norm": 5.46234864806171, "learning_rate": 3.419337981929667e-06, "loss": 0.3543, "step": 3625 }, { "epoch": 7.369918699186992, "grad_norm": 4.873325947629953, "learning_rate": 3.4143823801876507e-06, "loss": 0.422, "step": 3626 }, { "epoch": 7.371951219512195, "grad_norm": 5.354078173205154, "learning_rate": 3.409429632717729e-06, "loss": 0.5839, "step": 3627 }, { "epoch": 7.373983739837398, "grad_norm": 4.224611282784434, "learning_rate": 3.4044797416664564e-06, "loss": 0.4772, "step": 3628 }, { "epoch": 7.376016260162602, "grad_norm": 5.722773676189501, "learning_rate": 3.3995327091791765e-06, "loss": 0.5671, "step": 3629 }, { "epoch": 7.378048780487805, "grad_norm": 5.260345530254824, "learning_rate": 3.3945885373999786e-06, "loss": 0.4582, "step": 3630 }, { "epoch": 7.380081300813008, "grad_norm": 4.44825349461301, "learning_rate": 3.389647228471715e-06, "loss": 0.5596, "step": 3631 }, { "epoch": 7.382113821138211, "grad_norm": 4.7790146136779885, "learning_rate": 3.3847087845359996e-06, "loss": 0.6158, "step": 3632 }, { "epoch": 7.384146341463414, "grad_norm": 5.0659530385641105, "learning_rate": 3.379773207733199e-06, "loss": 0.4587, "step": 3633 }, { "epoch": 7.3861788617886175, "grad_norm": 4.737603826145709, "learning_rate": 3.3748405002024475e-06, "loss": 0.4733, "step": 3634 }, { "epoch": 7.388211382113822, "grad_norm": 4.564487098770683, "learning_rate": 3.369910664081627e-06, "loss": 0.4684, "step": 3635 }, { "epoch": 7.390243902439025, "grad_norm": 4.570951289162464, "learning_rate": 3.364983701507376e-06, "loss": 0.3877, "step": 3636 }, { "epoch": 7.392276422764228, "grad_norm": 5.456230975625752, "learning_rate": 3.360059614615091e-06, "loss": 0.414, "step": 3637 }, { "epoch": 7.394308943089431, "grad_norm": 3.765889663322952, "learning_rate": 3.355138405538916e-06, "loss": 0.4294, "step": 3638 }, { "epoch": 7.396341463414634, "grad_norm": 4.7517388561997445, "learning_rate": 3.3502200764117576e-06, "loss": 0.3703, "step": 3639 }, { "epoch": 7.3983739837398375, "grad_norm": 4.272584432749044, "learning_rate": 3.3453046293652657e-06, "loss": 0.3252, "step": 3640 }, { "epoch": 7.400406504065041, "grad_norm": 4.814866618192605, "learning_rate": 3.340392066529845e-06, "loss": 0.4179, "step": 3641 }, { "epoch": 7.402439024390244, "grad_norm": 4.607088179315851, "learning_rate": 3.335482390034648e-06, "loss": 0.3601, "step": 3642 }, { "epoch": 7.404471544715447, "grad_norm": 5.337056853501016, "learning_rate": 3.330575602007573e-06, "loss": 0.4662, "step": 3643 }, { "epoch": 7.40650406504065, "grad_norm": 5.220962332821225, "learning_rate": 3.3256717045752794e-06, "loss": 0.5371, "step": 3644 }, { "epoch": 7.408536585365853, "grad_norm": 4.8227689581555655, "learning_rate": 3.3207706998631606e-06, "loss": 0.4566, "step": 3645 }, { "epoch": 7.4105691056910565, "grad_norm": 4.552403637503241, "learning_rate": 3.31587258999536e-06, "loss": 0.55, "step": 3646 }, { "epoch": 7.41260162601626, "grad_norm": 4.662980958666271, "learning_rate": 3.3109773770947683e-06, "loss": 0.5272, "step": 3647 }, { "epoch": 7.414634146341464, "grad_norm": 4.908220235046397, "learning_rate": 3.3060850632830167e-06, "loss": 0.4669, "step": 3648 }, { "epoch": 7.416666666666667, "grad_norm": 4.5094641115357215, "learning_rate": 3.3011956506804876e-06, "loss": 0.5206, "step": 3649 }, { "epoch": 7.41869918699187, "grad_norm": 6.025065224615541, "learning_rate": 3.2963091414062977e-06, "loss": 0.4634, "step": 3650 }, { "epoch": 7.420731707317073, "grad_norm": 4.342252690371979, "learning_rate": 3.2914255375783098e-06, "loss": 0.4744, "step": 3651 }, { "epoch": 7.4227642276422765, "grad_norm": 5.145966485853846, "learning_rate": 3.286544841313126e-06, "loss": 0.426, "step": 3652 }, { "epoch": 7.42479674796748, "grad_norm": 4.555643026390303, "learning_rate": 3.2816670547260842e-06, "loss": 0.4331, "step": 3653 }, { "epoch": 7.426829268292683, "grad_norm": 6.142064686531574, "learning_rate": 3.276792179931274e-06, "loss": 0.457, "step": 3654 }, { "epoch": 7.428861788617886, "grad_norm": 4.468978428403116, "learning_rate": 3.271920219041509e-06, "loss": 0.4666, "step": 3655 }, { "epoch": 7.430894308943089, "grad_norm": 4.92535823991805, "learning_rate": 3.2670511741683475e-06, "loss": 0.3755, "step": 3656 }, { "epoch": 7.432926829268292, "grad_norm": 4.731102444214811, "learning_rate": 3.2621850474220807e-06, "loss": 0.3686, "step": 3657 }, { "epoch": 7.434959349593496, "grad_norm": 5.010656338355946, "learning_rate": 3.2573218409117337e-06, "loss": 0.4985, "step": 3658 }, { "epoch": 7.4369918699187, "grad_norm": 4.861497265462308, "learning_rate": 3.252461556745077e-06, "loss": 0.602, "step": 3659 }, { "epoch": 7.439024390243903, "grad_norm": 5.438139251441004, "learning_rate": 3.2476041970285945e-06, "loss": 0.5246, "step": 3660 }, { "epoch": 7.441056910569106, "grad_norm": 5.26480836846014, "learning_rate": 3.242749763867523e-06, "loss": 0.5066, "step": 3661 }, { "epoch": 7.443089430894309, "grad_norm": 5.5276497259515525, "learning_rate": 3.2378982593658192e-06, "loss": 0.4861, "step": 3662 }, { "epoch": 7.445121951219512, "grad_norm": 4.594320626121325, "learning_rate": 3.2330496856261707e-06, "loss": 0.3779, "step": 3663 }, { "epoch": 7.4471544715447155, "grad_norm": 4.5071034535978205, "learning_rate": 3.2282040447500063e-06, "loss": 0.5218, "step": 3664 }, { "epoch": 7.449186991869919, "grad_norm": 4.85745250684188, "learning_rate": 3.223361338837462e-06, "loss": 0.5093, "step": 3665 }, { "epoch": 7.451219512195122, "grad_norm": 5.1320134879384645, "learning_rate": 3.218521569987424e-06, "loss": 0.5903, "step": 3666 }, { "epoch": 7.453252032520325, "grad_norm": 4.087245303272715, "learning_rate": 3.213684740297495e-06, "loss": 0.4314, "step": 3667 }, { "epoch": 7.455284552845528, "grad_norm": 4.191986659076628, "learning_rate": 3.208850851863998e-06, "loss": 0.4504, "step": 3668 }, { "epoch": 7.4573170731707314, "grad_norm": 4.331351776988704, "learning_rate": 3.204019906782002e-06, "loss": 0.4802, "step": 3669 }, { "epoch": 7.459349593495935, "grad_norm": 4.880372080401179, "learning_rate": 3.1991919071452714e-06, "loss": 0.6107, "step": 3670 }, { "epoch": 7.461382113821138, "grad_norm": 4.450555892433645, "learning_rate": 3.1943668550463204e-06, "loss": 0.4661, "step": 3671 }, { "epoch": 7.463414634146342, "grad_norm": 4.489550541439964, "learning_rate": 3.189544752576369e-06, "loss": 0.4797, "step": 3672 }, { "epoch": 7.465447154471545, "grad_norm": 4.702681278355428, "learning_rate": 3.1847256018253656e-06, "loss": 0.4733, "step": 3673 }, { "epoch": 7.467479674796748, "grad_norm": 5.095479556393842, "learning_rate": 3.1799094048819778e-06, "loss": 0.4123, "step": 3674 }, { "epoch": 7.469512195121951, "grad_norm": 5.286250440699437, "learning_rate": 3.175096163833591e-06, "loss": 0.5728, "step": 3675 }, { "epoch": 7.471544715447155, "grad_norm": 4.450802371049586, "learning_rate": 3.1702858807663175e-06, "loss": 0.5577, "step": 3676 }, { "epoch": 7.473577235772358, "grad_norm": 4.2457950729599405, "learning_rate": 3.1654785577649737e-06, "loss": 0.3938, "step": 3677 }, { "epoch": 7.475609756097561, "grad_norm": 4.1391308763591095, "learning_rate": 3.160674196913114e-06, "loss": 0.392, "step": 3678 }, { "epoch": 7.477642276422764, "grad_norm": 4.958186005085288, "learning_rate": 3.155872800292983e-06, "loss": 0.4532, "step": 3679 }, { "epoch": 7.479674796747967, "grad_norm": 5.488195755764732, "learning_rate": 3.151074369985556e-06, "loss": 0.4556, "step": 3680 }, { "epoch": 7.4817073170731705, "grad_norm": 4.1338329793692585, "learning_rate": 3.1462789080705256e-06, "loss": 0.4555, "step": 3681 }, { "epoch": 7.483739837398374, "grad_norm": 4.564235867335125, "learning_rate": 3.141486416626287e-06, "loss": 0.4873, "step": 3682 }, { "epoch": 7.485772357723577, "grad_norm": 4.759919595205715, "learning_rate": 3.136696897729963e-06, "loss": 0.5345, "step": 3683 }, { "epoch": 7.487804878048781, "grad_norm": 4.787850656725748, "learning_rate": 3.131910353457369e-06, "loss": 0.4212, "step": 3684 }, { "epoch": 7.489837398373984, "grad_norm": 4.286219093610994, "learning_rate": 3.1271267858830413e-06, "loss": 0.4087, "step": 3685 }, { "epoch": 7.491869918699187, "grad_norm": 3.8948098822275643, "learning_rate": 3.1223461970802305e-06, "loss": 0.4552, "step": 3686 }, { "epoch": 7.4939024390243905, "grad_norm": 4.145188233153887, "learning_rate": 3.1175685891208863e-06, "loss": 0.5409, "step": 3687 }, { "epoch": 7.495934959349594, "grad_norm": 4.908608455564596, "learning_rate": 3.112793964075681e-06, "loss": 0.4727, "step": 3688 }, { "epoch": 7.497967479674797, "grad_norm": 3.9941465351439858, "learning_rate": 3.1080223240139708e-06, "loss": 0.4771, "step": 3689 }, { "epoch": 7.5, "grad_norm": 4.391651239692008, "learning_rate": 3.103253671003842e-06, "loss": 0.4688, "step": 3690 }, { "epoch": 7.502032520325203, "grad_norm": 4.935626284605809, "learning_rate": 3.0984880071120738e-06, "loss": 0.5316, "step": 3691 }, { "epoch": 7.504065040650406, "grad_norm": 4.636854965412212, "learning_rate": 3.0937253344041507e-06, "loss": 0.4605, "step": 3692 }, { "epoch": 7.5060975609756095, "grad_norm": 5.2405393948832515, "learning_rate": 3.0889656549442627e-06, "loss": 0.4289, "step": 3693 }, { "epoch": 7.508130081300813, "grad_norm": 5.8172131316021645, "learning_rate": 3.0842089707953006e-06, "loss": 0.5316, "step": 3694 }, { "epoch": 7.510162601626016, "grad_norm": 4.768913846931271, "learning_rate": 3.079455284018863e-06, "loss": 0.5756, "step": 3695 }, { "epoch": 7.512195121951219, "grad_norm": 4.456194366356688, "learning_rate": 3.074704596675242e-06, "loss": 0.6032, "step": 3696 }, { "epoch": 7.514227642276423, "grad_norm": 4.331522260829037, "learning_rate": 3.0699569108234338e-06, "loss": 0.4457, "step": 3697 }, { "epoch": 7.516260162601626, "grad_norm": 5.6580352628622474, "learning_rate": 3.0652122285211317e-06, "loss": 0.4292, "step": 3698 }, { "epoch": 7.5182926829268295, "grad_norm": 5.212418255808058, "learning_rate": 3.0604705518247257e-06, "loss": 0.4705, "step": 3699 }, { "epoch": 7.520325203252033, "grad_norm": 4.122706083537725, "learning_rate": 3.055731882789311e-06, "loss": 0.5655, "step": 3700 }, { "epoch": 7.522357723577236, "grad_norm": 5.026435122806168, "learning_rate": 3.0509962234686706e-06, "loss": 0.3786, "step": 3701 }, { "epoch": 7.524390243902439, "grad_norm": 6.174830789975938, "learning_rate": 3.046263575915288e-06, "loss": 0.3505, "step": 3702 }, { "epoch": 7.526422764227642, "grad_norm": 4.753166463213646, "learning_rate": 3.041533942180338e-06, "loss": 0.4218, "step": 3703 }, { "epoch": 7.528455284552845, "grad_norm": 4.922373935001802, "learning_rate": 3.0368073243136874e-06, "loss": 0.4914, "step": 3704 }, { "epoch": 7.530487804878049, "grad_norm": 5.126939367804718, "learning_rate": 3.0320837243639078e-06, "loss": 0.6351, "step": 3705 }, { "epoch": 7.532520325203252, "grad_norm": 5.6857155324431, "learning_rate": 3.027363144378249e-06, "loss": 0.5047, "step": 3706 }, { "epoch": 7.534552845528455, "grad_norm": 5.3314478975251225, "learning_rate": 3.0226455864026583e-06, "loss": 0.4517, "step": 3707 }, { "epoch": 7.536585365853659, "grad_norm": 4.447331053753311, "learning_rate": 3.0179310524817707e-06, "loss": 0.4489, "step": 3708 }, { "epoch": 7.538617886178862, "grad_norm": 5.4802924458617825, "learning_rate": 3.013219544658912e-06, "loss": 0.405, "step": 3709 }, { "epoch": 7.540650406504065, "grad_norm": 4.912472061724118, "learning_rate": 3.0085110649760996e-06, "loss": 0.4754, "step": 3710 }, { "epoch": 7.5426829268292686, "grad_norm": 5.375728751927224, "learning_rate": 3.0038056154740336e-06, "loss": 0.5507, "step": 3711 }, { "epoch": 7.544715447154472, "grad_norm": 4.429691343967992, "learning_rate": 2.9991031981921026e-06, "loss": 0.4982, "step": 3712 }, { "epoch": 7.546747967479675, "grad_norm": 4.710359697168867, "learning_rate": 2.9944038151683796e-06, "loss": 0.4128, "step": 3713 }, { "epoch": 7.548780487804878, "grad_norm": 4.627040066859626, "learning_rate": 2.989707468439623e-06, "loss": 0.4305, "step": 3714 }, { "epoch": 7.550813008130081, "grad_norm": 4.699411549772119, "learning_rate": 2.9850141600412818e-06, "loss": 0.4664, "step": 3715 }, { "epoch": 7.5528455284552845, "grad_norm": 4.448777892088852, "learning_rate": 2.9803238920074784e-06, "loss": 0.4747, "step": 3716 }, { "epoch": 7.554878048780488, "grad_norm": 5.196038121411549, "learning_rate": 2.975636666371022e-06, "loss": 0.4534, "step": 3717 }, { "epoch": 7.556910569105691, "grad_norm": 5.10011662818138, "learning_rate": 2.970952485163402e-06, "loss": 0.4416, "step": 3718 }, { "epoch": 7.558943089430894, "grad_norm": 4.8453165044527875, "learning_rate": 2.9662713504147877e-06, "loss": 0.4773, "step": 3719 }, { "epoch": 7.560975609756097, "grad_norm": 4.797867983520828, "learning_rate": 2.961593264154038e-06, "loss": 0.3324, "step": 3720 }, { "epoch": 7.5630081300813, "grad_norm": 5.1087276596256705, "learning_rate": 2.9569182284086685e-06, "loss": 0.4549, "step": 3721 }, { "epoch": 7.565040650406504, "grad_norm": 4.507012828586158, "learning_rate": 2.9522462452048962e-06, "loss": 0.525, "step": 3722 }, { "epoch": 7.567073170731708, "grad_norm": 4.929816618388189, "learning_rate": 2.9475773165676013e-06, "loss": 0.4464, "step": 3723 }, { "epoch": 7.569105691056911, "grad_norm": 5.046611764387076, "learning_rate": 2.9429114445203423e-06, "loss": 0.5353, "step": 3724 }, { "epoch": 7.571138211382114, "grad_norm": 4.784211100134764, "learning_rate": 2.938248631085362e-06, "loss": 0.453, "step": 3725 }, { "epoch": 7.573170731707317, "grad_norm": 4.863380300232617, "learning_rate": 2.93358887828356e-06, "loss": 0.4659, "step": 3726 }, { "epoch": 7.57520325203252, "grad_norm": 5.85315984691058, "learning_rate": 2.9289321881345257e-06, "loss": 0.5215, "step": 3727 }, { "epoch": 7.5772357723577235, "grad_norm": 5.104530410781777, "learning_rate": 2.924278562656514e-06, "loss": 0.5264, "step": 3728 }, { "epoch": 7.579268292682927, "grad_norm": 4.790751321783992, "learning_rate": 2.9196280038664482e-06, "loss": 0.6053, "step": 3729 }, { "epoch": 7.58130081300813, "grad_norm": 4.9592448226881665, "learning_rate": 2.914980513779937e-06, "loss": 0.5606, "step": 3730 }, { "epoch": 7.583333333333333, "grad_norm": 5.061149230284006, "learning_rate": 2.9103360944112347e-06, "loss": 0.5241, "step": 3731 }, { "epoch": 7.585365853658536, "grad_norm": 4.195228206853583, "learning_rate": 2.90569474777329e-06, "loss": 0.5206, "step": 3732 }, { "epoch": 7.58739837398374, "grad_norm": 4.469722627436474, "learning_rate": 2.901056475877704e-06, "loss": 0.4083, "step": 3733 }, { "epoch": 7.5894308943089435, "grad_norm": 5.283478887910522, "learning_rate": 2.896421280734748e-06, "loss": 0.4353, "step": 3734 }, { "epoch": 7.591463414634147, "grad_norm": 4.70235622802823, "learning_rate": 2.8917891643533704e-06, "loss": 0.5091, "step": 3735 }, { "epoch": 7.59349593495935, "grad_norm": 5.22305453397846, "learning_rate": 2.8871601287411634e-06, "loss": 0.532, "step": 3736 }, { "epoch": 7.595528455284553, "grad_norm": 4.564425875523574, "learning_rate": 2.8825341759044066e-06, "loss": 0.4698, "step": 3737 }, { "epoch": 7.597560975609756, "grad_norm": 4.379800402557346, "learning_rate": 2.8779113078480312e-06, "loss": 0.5691, "step": 3738 }, { "epoch": 7.599593495934959, "grad_norm": 5.048381888442893, "learning_rate": 2.8732915265756343e-06, "loss": 0.4459, "step": 3739 }, { "epoch": 7.6016260162601625, "grad_norm": 4.929906532728674, "learning_rate": 2.8686748340894744e-06, "loss": 0.3932, "step": 3740 }, { "epoch": 7.603658536585366, "grad_norm": 5.428997665298903, "learning_rate": 2.8640612323904693e-06, "loss": 0.5361, "step": 3741 }, { "epoch": 7.605691056910569, "grad_norm": 4.241709553279804, "learning_rate": 2.8594507234782056e-06, "loss": 0.4534, "step": 3742 }, { "epoch": 7.607723577235772, "grad_norm": 4.472997378635855, "learning_rate": 2.8548433093509207e-06, "loss": 0.4538, "step": 3743 }, { "epoch": 7.609756097560975, "grad_norm": 4.372396141357866, "learning_rate": 2.850238992005514e-06, "loss": 0.4635, "step": 3744 }, { "epoch": 7.611788617886178, "grad_norm": 5.227702594148673, "learning_rate": 2.845637773437541e-06, "loss": 0.5151, "step": 3745 }, { "epoch": 7.613821138211382, "grad_norm": 5.22507988102395, "learning_rate": 2.8410396556412145e-06, "loss": 0.4995, "step": 3746 }, { "epoch": 7.615853658536586, "grad_norm": 4.951961772715589, "learning_rate": 2.8364446406094082e-06, "loss": 0.6049, "step": 3747 }, { "epoch": 7.617886178861789, "grad_norm": 4.276998047187996, "learning_rate": 2.8318527303336465e-06, "loss": 0.5632, "step": 3748 }, { "epoch": 7.619918699186992, "grad_norm": 5.18547429355749, "learning_rate": 2.827263926804108e-06, "loss": 0.5517, "step": 3749 }, { "epoch": 7.621951219512195, "grad_norm": 5.214551476545565, "learning_rate": 2.8226782320096256e-06, "loss": 0.6782, "step": 3750 }, { "epoch": 7.623983739837398, "grad_norm": 5.2154605680424195, "learning_rate": 2.818095647937682e-06, "loss": 0.5421, "step": 3751 }, { "epoch": 7.626016260162602, "grad_norm": 5.010585376015617, "learning_rate": 2.81351617657442e-06, "loss": 0.4198, "step": 3752 }, { "epoch": 7.628048780487805, "grad_norm": 4.1276496549344674, "learning_rate": 2.808939819904627e-06, "loss": 0.5206, "step": 3753 }, { "epoch": 7.630081300813008, "grad_norm": 4.233289835900628, "learning_rate": 2.8043665799117393e-06, "loss": 0.555, "step": 3754 }, { "epoch": 7.632113821138211, "grad_norm": 4.714441092494784, "learning_rate": 2.799796458577845e-06, "loss": 0.4595, "step": 3755 }, { "epoch": 7.634146341463414, "grad_norm": 4.198834900228163, "learning_rate": 2.795229457883678e-06, "loss": 0.4975, "step": 3756 }, { "epoch": 7.636178861788618, "grad_norm": 5.159424424088631, "learning_rate": 2.7906655798086268e-06, "loss": 0.603, "step": 3757 }, { "epoch": 7.638211382113822, "grad_norm": 4.193649041532838, "learning_rate": 2.7861048263307188e-06, "loss": 0.4268, "step": 3758 }, { "epoch": 7.640243902439025, "grad_norm": 4.454552184214266, "learning_rate": 2.7815471994266295e-06, "loss": 0.4209, "step": 3759 }, { "epoch": 7.642276422764228, "grad_norm": 4.547948471008033, "learning_rate": 2.7769927010716814e-06, "loss": 0.4288, "step": 3760 }, { "epoch": 7.644308943089431, "grad_norm": 3.849725210772251, "learning_rate": 2.772441333239834e-06, "loss": 0.3554, "step": 3761 }, { "epoch": 7.646341463414634, "grad_norm": 5.4319576895837285, "learning_rate": 2.767893097903702e-06, "loss": 0.4988, "step": 3762 }, { "epoch": 7.6483739837398375, "grad_norm": 4.783721813038007, "learning_rate": 2.7633479970345333e-06, "loss": 0.4069, "step": 3763 }, { "epoch": 7.650406504065041, "grad_norm": 5.524616901920686, "learning_rate": 2.7588060326022205e-06, "loss": 0.5717, "step": 3764 }, { "epoch": 7.652439024390244, "grad_norm": 4.30454843369827, "learning_rate": 2.754267206575294e-06, "loss": 0.4176, "step": 3765 }, { "epoch": 7.654471544715447, "grad_norm": 4.700748130592763, "learning_rate": 2.7497315209209252e-06, "loss": 0.5448, "step": 3766 }, { "epoch": 7.65650406504065, "grad_norm": 5.088180767295921, "learning_rate": 2.7451989776049327e-06, "loss": 0.7093, "step": 3767 }, { "epoch": 7.658536585365853, "grad_norm": 4.765416438447765, "learning_rate": 2.740669578591755e-06, "loss": 0.4578, "step": 3768 }, { "epoch": 7.6605691056910565, "grad_norm": 4.686919757840828, "learning_rate": 2.736143325844487e-06, "loss": 0.454, "step": 3769 }, { "epoch": 7.66260162601626, "grad_norm": 4.127581587875808, "learning_rate": 2.7316202213248446e-06, "loss": 0.5267, "step": 3770 }, { "epoch": 7.664634146341464, "grad_norm": 4.556135125390092, "learning_rate": 2.7271002669931944e-06, "loss": 0.5139, "step": 3771 }, { "epoch": 7.666666666666667, "grad_norm": 5.347580018927086, "learning_rate": 2.7225834648085282e-06, "loss": 0.4461, "step": 3772 }, { "epoch": 7.66869918699187, "grad_norm": 4.969909814705541, "learning_rate": 2.7180698167284636e-06, "loss": 0.5475, "step": 3773 }, { "epoch": 7.670731707317073, "grad_norm": 5.631979675908869, "learning_rate": 2.7135593247092707e-06, "loss": 0.4535, "step": 3774 }, { "epoch": 7.6727642276422765, "grad_norm": 5.221806681033006, "learning_rate": 2.7090519907058344e-06, "loss": 0.4581, "step": 3775 }, { "epoch": 7.67479674796748, "grad_norm": 4.614132203652862, "learning_rate": 2.7045478166716843e-06, "loss": 0.4741, "step": 3776 }, { "epoch": 7.676829268292683, "grad_norm": 5.500336408506733, "learning_rate": 2.7000468045589743e-06, "loss": 0.4808, "step": 3777 }, { "epoch": 7.678861788617886, "grad_norm": 5.093313916889725, "learning_rate": 2.69554895631848e-06, "loss": 0.4384, "step": 3778 }, { "epoch": 7.680894308943089, "grad_norm": 4.605304898751118, "learning_rate": 2.6910542738996206e-06, "loss": 0.6727, "step": 3779 }, { "epoch": 7.682926829268292, "grad_norm": 5.028468380154885, "learning_rate": 2.6865627592504295e-06, "loss": 0.4201, "step": 3780 }, { "epoch": 7.684959349593496, "grad_norm": 3.927706328403866, "learning_rate": 2.6820744143175826e-06, "loss": 0.3509, "step": 3781 }, { "epoch": 7.6869918699187, "grad_norm": 5.169100782875247, "learning_rate": 2.6775892410463677e-06, "loss": 0.5485, "step": 3782 }, { "epoch": 7.689024390243903, "grad_norm": 5.092366893319107, "learning_rate": 2.673107241380706e-06, "loss": 0.3768, "step": 3783 }, { "epoch": 7.691056910569106, "grad_norm": 4.53322802583634, "learning_rate": 2.668628417263137e-06, "loss": 0.5173, "step": 3784 }, { "epoch": 7.693089430894309, "grad_norm": 4.600290427727572, "learning_rate": 2.664152770634828e-06, "loss": 0.4665, "step": 3785 }, { "epoch": 7.695121951219512, "grad_norm": 4.164791348969317, "learning_rate": 2.6596803034355756e-06, "loss": 0.2891, "step": 3786 }, { "epoch": 7.6971544715447155, "grad_norm": 4.954295602477142, "learning_rate": 2.6552110176037804e-06, "loss": 0.5471, "step": 3787 }, { "epoch": 7.699186991869919, "grad_norm": 4.234838350601568, "learning_rate": 2.6507449150764852e-06, "loss": 0.4258, "step": 3788 }, { "epoch": 7.701219512195122, "grad_norm": 4.492456308326234, "learning_rate": 2.646281997789337e-06, "loss": 0.5393, "step": 3789 }, { "epoch": 7.703252032520325, "grad_norm": 3.90528902916101, "learning_rate": 2.6418222676766093e-06, "loss": 0.379, "step": 3790 }, { "epoch": 7.705284552845528, "grad_norm": 4.254710148546163, "learning_rate": 2.6373657266712007e-06, "loss": 0.4343, "step": 3791 }, { "epoch": 7.7073170731707314, "grad_norm": 4.355238120905955, "learning_rate": 2.632912376704607e-06, "loss": 0.4551, "step": 3792 }, { "epoch": 7.709349593495935, "grad_norm": 3.971636844607251, "learning_rate": 2.6284622197069652e-06, "loss": 0.4109, "step": 3793 }, { "epoch": 7.711382113821138, "grad_norm": 4.732251139725413, "learning_rate": 2.6240152576070143e-06, "loss": 0.5896, "step": 3794 }, { "epoch": 7.713414634146341, "grad_norm": 5.703842019309404, "learning_rate": 2.619571492332108e-06, "loss": 0.5103, "step": 3795 }, { "epoch": 7.715447154471545, "grad_norm": 4.453665924871902, "learning_rate": 2.615130925808228e-06, "loss": 0.4579, "step": 3796 }, { "epoch": 7.717479674796748, "grad_norm": 4.406549343455073, "learning_rate": 2.6106935599599482e-06, "loss": 0.6048, "step": 3797 }, { "epoch": 7.719512195121951, "grad_norm": 5.134596017780205, "learning_rate": 2.6062593967104756e-06, "loss": 0.4183, "step": 3798 }, { "epoch": 7.721544715447155, "grad_norm": 4.180727872219198, "learning_rate": 2.601828437981617e-06, "loss": 0.4372, "step": 3799 }, { "epoch": 7.723577235772358, "grad_norm": 4.260117849100918, "learning_rate": 2.597400685693795e-06, "loss": 0.5217, "step": 3800 }, { "epoch": 7.725609756097561, "grad_norm": 4.438341135970303, "learning_rate": 2.5929761417660402e-06, "loss": 0.4855, "step": 3801 }, { "epoch": 7.727642276422764, "grad_norm": 4.805654637068411, "learning_rate": 2.5885548081159927e-06, "loss": 0.4461, "step": 3802 }, { "epoch": 7.729674796747967, "grad_norm": 4.659121797486674, "learning_rate": 2.584136686659907e-06, "loss": 0.594, "step": 3803 }, { "epoch": 7.7317073170731705, "grad_norm": 5.32281906885249, "learning_rate": 2.5797217793126373e-06, "loss": 0.4809, "step": 3804 }, { "epoch": 7.733739837398374, "grad_norm": 4.619562123294261, "learning_rate": 2.57531008798765e-06, "loss": 0.4896, "step": 3805 }, { "epoch": 7.735772357723577, "grad_norm": 4.291309532679024, "learning_rate": 2.5709016145970146e-06, "loss": 0.4998, "step": 3806 }, { "epoch": 7.737804878048781, "grad_norm": 4.8722557733875025, "learning_rate": 2.566496361051405e-06, "loss": 0.4229, "step": 3807 }, { "epoch": 7.739837398373984, "grad_norm": 4.208208023682546, "learning_rate": 2.5620943292601074e-06, "loss": 0.4161, "step": 3808 }, { "epoch": 7.741869918699187, "grad_norm": 5.182797036733125, "learning_rate": 2.557695521131004e-06, "loss": 0.5301, "step": 3809 }, { "epoch": 7.7439024390243905, "grad_norm": 4.554723051583676, "learning_rate": 2.55329993857058e-06, "loss": 0.2923, "step": 3810 }, { "epoch": 7.745934959349594, "grad_norm": 4.388536748873854, "learning_rate": 2.5489075834839272e-06, "loss": 0.4437, "step": 3811 }, { "epoch": 7.747967479674797, "grad_norm": 4.20213187160745, "learning_rate": 2.5445184577747305e-06, "loss": 0.4131, "step": 3812 }, { "epoch": 7.75, "grad_norm": 4.499205090408137, "learning_rate": 2.5401325633452876e-06, "loss": 0.6037, "step": 3813 }, { "epoch": 7.752032520325203, "grad_norm": 4.763872513052832, "learning_rate": 2.535749902096486e-06, "loss": 0.5629, "step": 3814 }, { "epoch": 7.754065040650406, "grad_norm": 4.625912363552164, "learning_rate": 2.5313704759278147e-06, "loss": 0.4354, "step": 3815 }, { "epoch": 7.7560975609756095, "grad_norm": 4.547824000156418, "learning_rate": 2.52699428673736e-06, "loss": 0.4454, "step": 3816 }, { "epoch": 7.758130081300813, "grad_norm": 4.535673991954115, "learning_rate": 2.522621336421802e-06, "loss": 0.3484, "step": 3817 }, { "epoch": 7.760162601626016, "grad_norm": 4.111891359997489, "learning_rate": 2.5182516268764277e-06, "loss": 0.534, "step": 3818 }, { "epoch": 7.762195121951219, "grad_norm": 5.399624671244039, "learning_rate": 2.51388515999511e-06, "loss": 0.5993, "step": 3819 }, { "epoch": 7.764227642276423, "grad_norm": 4.5588907822407485, "learning_rate": 2.5095219376703183e-06, "loss": 0.4319, "step": 3820 }, { "epoch": 7.766260162601626, "grad_norm": 5.181408140599117, "learning_rate": 2.5051619617931165e-06, "loss": 0.4893, "step": 3821 }, { "epoch": 7.7682926829268295, "grad_norm": 4.665716620809846, "learning_rate": 2.5008052342531595e-06, "loss": 0.5023, "step": 3822 }, { "epoch": 7.770325203252033, "grad_norm": 5.2431025763151835, "learning_rate": 2.4964517569387015e-06, "loss": 0.4305, "step": 3823 }, { "epoch": 7.772357723577236, "grad_norm": 4.993983753138674, "learning_rate": 2.4921015317365794e-06, "loss": 0.5773, "step": 3824 }, { "epoch": 7.774390243902439, "grad_norm": 6.675974494937307, "learning_rate": 2.487754560532224e-06, "loss": 0.5683, "step": 3825 }, { "epoch": 7.776422764227642, "grad_norm": 5.752113561927763, "learning_rate": 2.4834108452096572e-06, "loss": 0.5037, "step": 3826 }, { "epoch": 7.778455284552845, "grad_norm": 4.548231900432309, "learning_rate": 2.4790703876514843e-06, "loss": 0.4037, "step": 3827 }, { "epoch": 7.780487804878049, "grad_norm": 3.9658623785574765, "learning_rate": 2.4747331897389103e-06, "loss": 0.404, "step": 3828 }, { "epoch": 7.782520325203252, "grad_norm": 4.523717721553568, "learning_rate": 2.4703992533517108e-06, "loss": 0.5037, "step": 3829 }, { "epoch": 7.784552845528455, "grad_norm": 5.028234685275456, "learning_rate": 2.466068580368265e-06, "loss": 0.5625, "step": 3830 }, { "epoch": 7.786585365853659, "grad_norm": 5.1084271259675536, "learning_rate": 2.4617411726655247e-06, "loss": 0.441, "step": 3831 }, { "epoch": 7.788617886178862, "grad_norm": 4.718621393988089, "learning_rate": 2.4574170321190305e-06, "loss": 0.5084, "step": 3832 }, { "epoch": 7.790650406504065, "grad_norm": 4.364829044783747, "learning_rate": 2.4530961606029146e-06, "loss": 0.7042, "step": 3833 }, { "epoch": 7.7926829268292686, "grad_norm": 4.888451938995284, "learning_rate": 2.448778559989876e-06, "loss": 0.3277, "step": 3834 }, { "epoch": 7.794715447154472, "grad_norm": 4.732245145533276, "learning_rate": 2.4444642321512148e-06, "loss": 0.4838, "step": 3835 }, { "epoch": 7.796747967479675, "grad_norm": 3.837943501487874, "learning_rate": 2.440153178956798e-06, "loss": 0.3372, "step": 3836 }, { "epoch": 7.798780487804878, "grad_norm": 5.0378457108590355, "learning_rate": 2.435845402275079e-06, "loss": 0.492, "step": 3837 }, { "epoch": 7.800813008130081, "grad_norm": 5.453242549515645, "learning_rate": 2.431540903973096e-06, "loss": 0.4101, "step": 3838 }, { "epoch": 7.8028455284552845, "grad_norm": 4.879436334102326, "learning_rate": 2.4272396859164537e-06, "loss": 0.5519, "step": 3839 }, { "epoch": 7.804878048780488, "grad_norm": 5.94120418116907, "learning_rate": 2.42294174996935e-06, "loss": 0.4092, "step": 3840 }, { "epoch": 7.806910569105691, "grad_norm": 4.9023982977920015, "learning_rate": 2.4186470979945485e-06, "loss": 0.5504, "step": 3841 }, { "epoch": 7.808943089430894, "grad_norm": 4.680883230843906, "learning_rate": 2.4143557318533926e-06, "loss": 0.4264, "step": 3842 }, { "epoch": 7.810975609756097, "grad_norm": 4.795228612766729, "learning_rate": 2.410067653405812e-06, "loss": 0.5371, "step": 3843 }, { "epoch": 7.8130081300813, "grad_norm": 4.176059264871699, "learning_rate": 2.40578286451029e-06, "loss": 0.3341, "step": 3844 }, { "epoch": 7.815040650406504, "grad_norm": 4.855975937354845, "learning_rate": 2.401501367023905e-06, "loss": 0.4669, "step": 3845 }, { "epoch": 7.817073170731708, "grad_norm": 4.479889853941703, "learning_rate": 2.397223162802297e-06, "loss": 0.458, "step": 3846 }, { "epoch": 7.819105691056911, "grad_norm": 4.423319028743715, "learning_rate": 2.3929482536996886e-06, "loss": 0.4341, "step": 3847 }, { "epoch": 7.821138211382114, "grad_norm": 4.688894316598821, "learning_rate": 2.38867664156886e-06, "loss": 0.5829, "step": 3848 }, { "epoch": 7.823170731707317, "grad_norm": 5.749044773617812, "learning_rate": 2.3844083282611696e-06, "loss": 0.4535, "step": 3849 }, { "epoch": 7.82520325203252, "grad_norm": 5.583200771177931, "learning_rate": 2.380143315626554e-06, "loss": 0.5782, "step": 3850 }, { "epoch": 7.8272357723577235, "grad_norm": 5.034294457532555, "learning_rate": 2.3758816055135057e-06, "loss": 0.443, "step": 3851 }, { "epoch": 7.829268292682927, "grad_norm": 6.856026323959198, "learning_rate": 2.3716231997691007e-06, "loss": 0.4143, "step": 3852 }, { "epoch": 7.83130081300813, "grad_norm": 4.380425662045393, "learning_rate": 2.367368100238965e-06, "loss": 0.4394, "step": 3853 }, { "epoch": 7.833333333333333, "grad_norm": 4.103782134393256, "learning_rate": 2.3631163087673027e-06, "loss": 0.5321, "step": 3854 }, { "epoch": 7.835365853658536, "grad_norm": 4.9217927377185555, "learning_rate": 2.358867827196886e-06, "loss": 0.3912, "step": 3855 }, { "epoch": 7.83739837398374, "grad_norm": 4.188839868727801, "learning_rate": 2.3546226573690444e-06, "loss": 0.5113, "step": 3856 }, { "epoch": 7.8394308943089435, "grad_norm": 5.2773984524355635, "learning_rate": 2.350380801123686e-06, "loss": 0.566, "step": 3857 }, { "epoch": 7.841463414634147, "grad_norm": 4.783463543789386, "learning_rate": 2.3461422602992646e-06, "loss": 0.4312, "step": 3858 }, { "epoch": 7.84349593495935, "grad_norm": 5.060716937466694, "learning_rate": 2.3419070367328047e-06, "loss": 0.4981, "step": 3859 }, { "epoch": 7.845528455284553, "grad_norm": 5.5068154294132485, "learning_rate": 2.3376751322599e-06, "loss": 0.4658, "step": 3860 }, { "epoch": 7.847560975609756, "grad_norm": 4.679926744497658, "learning_rate": 2.3334465487146983e-06, "loss": 0.5999, "step": 3861 }, { "epoch": 7.849593495934959, "grad_norm": 5.273715054843306, "learning_rate": 2.3292212879299104e-06, "loss": 0.4802, "step": 3862 }, { "epoch": 7.8516260162601625, "grad_norm": 5.397987001053626, "learning_rate": 2.324999351736801e-06, "loss": 0.5594, "step": 3863 }, { "epoch": 7.853658536585366, "grad_norm": 4.186644492467845, "learning_rate": 2.320780741965206e-06, "loss": 0.5631, "step": 3864 }, { "epoch": 7.855691056910569, "grad_norm": 5.605619372111786, "learning_rate": 2.3165654604435107e-06, "loss": 0.5811, "step": 3865 }, { "epoch": 7.857723577235772, "grad_norm": 4.828541626952495, "learning_rate": 2.312353508998658e-06, "loss": 0.4149, "step": 3866 }, { "epoch": 7.859756097560975, "grad_norm": 4.90805993540478, "learning_rate": 2.308144889456152e-06, "loss": 0.5849, "step": 3867 }, { "epoch": 7.861788617886178, "grad_norm": 4.573860312363479, "learning_rate": 2.3039396036400463e-06, "loss": 0.4247, "step": 3868 }, { "epoch": 7.863821138211382, "grad_norm": 4.701392787322737, "learning_rate": 2.2997376533729586e-06, "loss": 0.4548, "step": 3869 }, { "epoch": 7.865853658536586, "grad_norm": 5.284926771169445, "learning_rate": 2.2955390404760537e-06, "loss": 0.508, "step": 3870 }, { "epoch": 7.867886178861789, "grad_norm": 4.912389809857798, "learning_rate": 2.291343766769052e-06, "loss": 0.3765, "step": 3871 }, { "epoch": 7.869918699186992, "grad_norm": 5.03355275850466, "learning_rate": 2.287151834070226e-06, "loss": 0.4641, "step": 3872 }, { "epoch": 7.871951219512195, "grad_norm": 5.149381273098321, "learning_rate": 2.2829632441964e-06, "loss": 0.4564, "step": 3873 }, { "epoch": 7.873983739837398, "grad_norm": 4.575313206867425, "learning_rate": 2.278777998962954e-06, "loss": 0.4967, "step": 3874 }, { "epoch": 7.876016260162602, "grad_norm": 4.42820052598419, "learning_rate": 2.2745961001838134e-06, "loss": 0.6259, "step": 3875 }, { "epoch": 7.878048780487805, "grad_norm": 4.593268844416341, "learning_rate": 2.2704175496714552e-06, "loss": 0.5422, "step": 3876 }, { "epoch": 7.880081300813008, "grad_norm": 4.72772291633465, "learning_rate": 2.266242349236902e-06, "loss": 0.4385, "step": 3877 }, { "epoch": 7.882113821138211, "grad_norm": 4.497965921548958, "learning_rate": 2.262070500689728e-06, "loss": 0.4238, "step": 3878 }, { "epoch": 7.884146341463414, "grad_norm": 5.3507683885666335, "learning_rate": 2.2579020058380565e-06, "loss": 0.4727, "step": 3879 }, { "epoch": 7.886178861788618, "grad_norm": 4.84333722162912, "learning_rate": 2.2537368664885527e-06, "loss": 0.4421, "step": 3880 }, { "epoch": 7.888211382113822, "grad_norm": 5.54325595042027, "learning_rate": 2.2495750844464303e-06, "loss": 0.4965, "step": 3881 }, { "epoch": 7.890243902439025, "grad_norm": 5.335985263266367, "learning_rate": 2.245416661515446e-06, "loss": 0.4764, "step": 3882 }, { "epoch": 7.892276422764228, "grad_norm": 4.285462403554655, "learning_rate": 2.241261599497898e-06, "loss": 0.3738, "step": 3883 }, { "epoch": 7.894308943089431, "grad_norm": 4.579572076940086, "learning_rate": 2.2371099001946385e-06, "loss": 0.5959, "step": 3884 }, { "epoch": 7.896341463414634, "grad_norm": 5.10353793859756, "learning_rate": 2.232961565405052e-06, "loss": 0.4522, "step": 3885 }, { "epoch": 7.8983739837398375, "grad_norm": 4.286787480035963, "learning_rate": 2.228816596927068e-06, "loss": 0.389, "step": 3886 }, { "epoch": 7.900406504065041, "grad_norm": 5.183916439713678, "learning_rate": 2.2246749965571567e-06, "loss": 0.4997, "step": 3887 }, { "epoch": 7.902439024390244, "grad_norm": 5.126118455450605, "learning_rate": 2.2205367660903267e-06, "loss": 0.4737, "step": 3888 }, { "epoch": 7.904471544715447, "grad_norm": 4.56431730673819, "learning_rate": 2.2164019073201326e-06, "loss": 0.5009, "step": 3889 }, { "epoch": 7.90650406504065, "grad_norm": 4.430404787303161, "learning_rate": 2.2122704220386616e-06, "loss": 0.4584, "step": 3890 }, { "epoch": 7.908536585365853, "grad_norm": 4.797823564296694, "learning_rate": 2.2081423120365396e-06, "loss": 0.4026, "step": 3891 }, { "epoch": 7.9105691056910565, "grad_norm": 4.526270106275515, "learning_rate": 2.2040175791029305e-06, "loss": 0.4437, "step": 3892 }, { "epoch": 7.91260162601626, "grad_norm": 4.686106561972406, "learning_rate": 2.199896225025534e-06, "loss": 0.414, "step": 3893 }, { "epoch": 7.914634146341464, "grad_norm": 5.484214934892898, "learning_rate": 2.1957782515905914e-06, "loss": 0.3623, "step": 3894 }, { "epoch": 7.916666666666667, "grad_norm": 4.477269706407296, "learning_rate": 2.1916636605828624e-06, "loss": 0.5283, "step": 3895 }, { "epoch": 7.91869918699187, "grad_norm": 5.441185615030369, "learning_rate": 2.187552453785662e-06, "loss": 0.301, "step": 3896 }, { "epoch": 7.920731707317073, "grad_norm": 4.766672821044264, "learning_rate": 2.1834446329808225e-06, "loss": 0.5458, "step": 3897 }, { "epoch": 7.9227642276422765, "grad_norm": 4.849174656488165, "learning_rate": 2.179340199948714e-06, "loss": 0.5995, "step": 3898 }, { "epoch": 7.92479674796748, "grad_norm": 5.826115516490803, "learning_rate": 2.1752391564682463e-06, "loss": 0.438, "step": 3899 }, { "epoch": 7.926829268292683, "grad_norm": 6.2291878289945135, "learning_rate": 2.1711415043168395e-06, "loss": 0.5521, "step": 3900 }, { "epoch": 7.928861788617886, "grad_norm": 5.019732423147515, "learning_rate": 2.1670472452704673e-06, "loss": 0.4153, "step": 3901 }, { "epoch": 7.930894308943089, "grad_norm": 4.8550267893038175, "learning_rate": 2.162956381103618e-06, "loss": 0.4899, "step": 3902 }, { "epoch": 7.932926829268292, "grad_norm": 5.189439473486082, "learning_rate": 2.1588689135893104e-06, "loss": 0.4834, "step": 3903 }, { "epoch": 7.934959349593496, "grad_norm": 5.376457646645116, "learning_rate": 2.1547848444991025e-06, "loss": 0.4157, "step": 3904 }, { "epoch": 7.9369918699187, "grad_norm": 5.476067915679332, "learning_rate": 2.150704175603059e-06, "loss": 0.4435, "step": 3905 }, { "epoch": 7.939024390243903, "grad_norm": 3.8263384722429454, "learning_rate": 2.14662690866979e-06, "loss": 0.4427, "step": 3906 }, { "epoch": 7.941056910569106, "grad_norm": 6.310398374095678, "learning_rate": 2.142553045466421e-06, "loss": 0.4871, "step": 3907 }, { "epoch": 7.943089430894309, "grad_norm": 4.1237040575246215, "learning_rate": 2.138482587758605e-06, "loss": 0.3939, "step": 3908 }, { "epoch": 7.945121951219512, "grad_norm": 4.977326851223955, "learning_rate": 2.134415537310519e-06, "loss": 0.348, "step": 3909 }, { "epoch": 7.9471544715447155, "grad_norm": 4.194638140433369, "learning_rate": 2.130351895884859e-06, "loss": 0.46, "step": 3910 }, { "epoch": 7.949186991869919, "grad_norm": 5.201752838834405, "learning_rate": 2.1262916652428554e-06, "loss": 0.4495, "step": 3911 }, { "epoch": 7.951219512195122, "grad_norm": 5.0824069388639295, "learning_rate": 2.1222348471442477e-06, "loss": 0.4036, "step": 3912 }, { "epoch": 7.953252032520325, "grad_norm": 5.096300540857981, "learning_rate": 2.1181814433473026e-06, "loss": 0.4954, "step": 3913 }, { "epoch": 7.955284552845528, "grad_norm": 4.706910898710556, "learning_rate": 2.1141314556088033e-06, "loss": 0.5573, "step": 3914 }, { "epoch": 7.9573170731707314, "grad_norm": 4.420561603037966, "learning_rate": 2.110084885684055e-06, "loss": 0.5218, "step": 3915 }, { "epoch": 7.959349593495935, "grad_norm": 4.273050380796053, "learning_rate": 2.1060417353268845e-06, "loss": 0.564, "step": 3916 }, { "epoch": 7.961382113821138, "grad_norm": 4.582872814600341, "learning_rate": 2.1020020062896317e-06, "loss": 0.5159, "step": 3917 }, { "epoch": 7.963414634146341, "grad_norm": 6.3783245349504, "learning_rate": 2.0979657003231547e-06, "loss": 0.429, "step": 3918 }, { "epoch": 7.965447154471545, "grad_norm": 4.997711235955955, "learning_rate": 2.0939328191768293e-06, "loss": 0.4402, "step": 3919 }, { "epoch": 7.967479674796748, "grad_norm": 5.388904914003042, "learning_rate": 2.0899033645985423e-06, "loss": 0.4078, "step": 3920 }, { "epoch": 7.969512195121951, "grad_norm": 4.945336443270994, "learning_rate": 2.0858773383347063e-06, "loss": 0.4027, "step": 3921 }, { "epoch": 7.971544715447155, "grad_norm": 4.303040801151482, "learning_rate": 2.081854742130237e-06, "loss": 0.5691, "step": 3922 }, { "epoch": 7.973577235772358, "grad_norm": 5.197115116448573, "learning_rate": 2.0778355777285686e-06, "loss": 0.4308, "step": 3923 }, { "epoch": 7.975609756097561, "grad_norm": 4.897656805090381, "learning_rate": 2.073819846871646e-06, "loss": 0.4754, "step": 3924 }, { "epoch": 7.977642276422764, "grad_norm": 4.45683863844661, "learning_rate": 2.069807551299925e-06, "loss": 0.4547, "step": 3925 }, { "epoch": 7.979674796747967, "grad_norm": 4.7791510356459215, "learning_rate": 2.065798692752379e-06, "loss": 0.4662, "step": 3926 }, { "epoch": 7.9817073170731705, "grad_norm": 4.788456244975144, "learning_rate": 2.0617932729664846e-06, "loss": 0.3692, "step": 3927 }, { "epoch": 7.983739837398374, "grad_norm": 4.259034507200789, "learning_rate": 2.0577912936782317e-06, "loss": 0.4203, "step": 3928 }, { "epoch": 7.985772357723577, "grad_norm": 5.29351203400426, "learning_rate": 2.053792756622116e-06, "loss": 0.4664, "step": 3929 }, { "epoch": 7.987804878048781, "grad_norm": 4.7671641931599265, "learning_rate": 2.0497976635311413e-06, "loss": 0.4412, "step": 3930 }, { "epoch": 7.989837398373984, "grad_norm": 4.979495631612022, "learning_rate": 2.0458060161368252e-06, "loss": 0.4633, "step": 3931 }, { "epoch": 7.991869918699187, "grad_norm": 5.209914615847358, "learning_rate": 2.041817816169187e-06, "loss": 0.5265, "step": 3932 }, { "epoch": 7.9939024390243905, "grad_norm": 4.803751519677722, "learning_rate": 2.03783306535675e-06, "loss": 0.5186, "step": 3933 }, { "epoch": 7.995934959349594, "grad_norm": 5.237442282089653, "learning_rate": 2.0338517654265454e-06, "loss": 0.596, "step": 3934 }, { "epoch": 7.997967479674797, "grad_norm": 4.305388536229715, "learning_rate": 2.029873918104105e-06, "loss": 0.4365, "step": 3935 }, { "epoch": 8.0, "grad_norm": 5.451819473399103, "learning_rate": 2.025899525113474e-06, "loss": 0.5506, "step": 3936 }, { "epoch": 8.0, "eval_loss": 1.0416016578674316, "eval_runtime": 132.2497, "eval_samples_per_second": 7.44, "eval_steps_per_second": 0.93, "step": 3936 }, { "epoch": 8.002032520325203, "grad_norm": 4.589758536804154, "learning_rate": 2.0219285881771887e-06, "loss": 0.4704, "step": 3937 }, { "epoch": 8.004065040650406, "grad_norm": 4.195445874379309, "learning_rate": 2.0179611090162955e-06, "loss": 0.4159, "step": 3938 }, { "epoch": 8.00609756097561, "grad_norm": 4.89362667033615, "learning_rate": 2.013997089350336e-06, "loss": 0.4783, "step": 3939 }, { "epoch": 8.008130081300813, "grad_norm": 4.277475043630846, "learning_rate": 2.010036530897359e-06, "loss": 0.5139, "step": 3940 }, { "epoch": 8.010162601626016, "grad_norm": 4.7627473803969265, "learning_rate": 2.006079435373911e-06, "loss": 0.4875, "step": 3941 }, { "epoch": 8.012195121951219, "grad_norm": 4.527953953826826, "learning_rate": 2.0021258044950296e-06, "loss": 0.4187, "step": 3942 }, { "epoch": 8.014227642276422, "grad_norm": 5.169525555224727, "learning_rate": 1.9981756399742645e-06, "loss": 0.4247, "step": 3943 }, { "epoch": 8.016260162601625, "grad_norm": 4.836695546217096, "learning_rate": 1.9942289435236506e-06, "loss": 0.4615, "step": 3944 }, { "epoch": 8.018292682926829, "grad_norm": 3.989384361665082, "learning_rate": 1.9902857168537326e-06, "loss": 0.4962, "step": 3945 }, { "epoch": 8.020325203252032, "grad_norm": 4.345695006900319, "learning_rate": 1.9863459616735412e-06, "loss": 0.4902, "step": 3946 }, { "epoch": 8.022357723577235, "grad_norm": 4.531904977620209, "learning_rate": 1.9824096796906e-06, "loss": 0.3581, "step": 3947 }, { "epoch": 8.024390243902438, "grad_norm": 5.241967198019063, "learning_rate": 1.978476872610939e-06, "loss": 0.5473, "step": 3948 }, { "epoch": 8.026422764227643, "grad_norm": 4.594838740379607, "learning_rate": 1.9745475421390724e-06, "loss": 0.4498, "step": 3949 }, { "epoch": 8.028455284552846, "grad_norm": 4.623259805693507, "learning_rate": 1.970621689978014e-06, "loss": 0.4076, "step": 3950 }, { "epoch": 8.03048780487805, "grad_norm": 5.217269340461873, "learning_rate": 1.966699317829266e-06, "loss": 0.3807, "step": 3951 }, { "epoch": 8.032520325203253, "grad_norm": 4.613855976165618, "learning_rate": 1.962780427392823e-06, "loss": 0.4183, "step": 3952 }, { "epoch": 8.034552845528456, "grad_norm": 5.381205983168269, "learning_rate": 1.958865020367172e-06, "loss": 0.4034, "step": 3953 }, { "epoch": 8.036585365853659, "grad_norm": 5.337970809513579, "learning_rate": 1.9549530984492847e-06, "loss": 0.4845, "step": 3954 }, { "epoch": 8.038617886178862, "grad_norm": 5.1452602426454375, "learning_rate": 1.951044663334637e-06, "loss": 0.5262, "step": 3955 }, { "epoch": 8.040650406504065, "grad_norm": 4.300666165619188, "learning_rate": 1.9471397167171714e-06, "loss": 0.4293, "step": 3956 }, { "epoch": 8.042682926829269, "grad_norm": 4.163354531441558, "learning_rate": 1.9432382602893386e-06, "loss": 0.4574, "step": 3957 }, { "epoch": 8.044715447154472, "grad_norm": 5.053844139963833, "learning_rate": 1.939340295742066e-06, "loss": 0.4784, "step": 3958 }, { "epoch": 8.046747967479675, "grad_norm": 5.370052220378414, "learning_rate": 1.9354458247647678e-06, "loss": 0.5959, "step": 3959 }, { "epoch": 8.048780487804878, "grad_norm": 4.835995561770971, "learning_rate": 1.931554849045355e-06, "loss": 0.3995, "step": 3960 }, { "epoch": 8.050813008130081, "grad_norm": 4.6643797681122345, "learning_rate": 1.927667370270204e-06, "loss": 0.5438, "step": 3961 }, { "epoch": 8.052845528455284, "grad_norm": 4.454331376133608, "learning_rate": 1.9237833901241943e-06, "loss": 0.527, "step": 3962 }, { "epoch": 8.054878048780488, "grad_norm": 4.673864435403591, "learning_rate": 1.9199029102906786e-06, "loss": 0.4139, "step": 3963 }, { "epoch": 8.05691056910569, "grad_norm": 4.860754030164447, "learning_rate": 1.916025932451493e-06, "loss": 0.486, "step": 3964 }, { "epoch": 8.058943089430894, "grad_norm": 5.191329197279034, "learning_rate": 1.9121524582869665e-06, "loss": 0.442, "step": 3965 }, { "epoch": 8.060975609756097, "grad_norm": 4.914748404463738, "learning_rate": 1.908282489475891e-06, "loss": 0.4326, "step": 3966 }, { "epoch": 8.0630081300813, "grad_norm": 4.905104352142498, "learning_rate": 1.904416027695557e-06, "loss": 0.4267, "step": 3967 }, { "epoch": 8.065040650406504, "grad_norm": 4.306076578564235, "learning_rate": 1.9005530746217238e-06, "loss": 0.4138, "step": 3968 }, { "epoch": 8.067073170731707, "grad_norm": 3.73664804352769, "learning_rate": 1.8966936319286355e-06, "loss": 0.5004, "step": 3969 }, { "epoch": 8.06910569105691, "grad_norm": 4.762702813581556, "learning_rate": 1.892837701289011e-06, "loss": 0.4472, "step": 3970 }, { "epoch": 8.071138211382113, "grad_norm": 4.957221426199844, "learning_rate": 1.8889852843740474e-06, "loss": 0.5777, "step": 3971 }, { "epoch": 8.073170731707316, "grad_norm": 5.116679911269073, "learning_rate": 1.8851363828534253e-06, "loss": 0.4406, "step": 3972 }, { "epoch": 8.075203252032521, "grad_norm": 4.176825197409603, "learning_rate": 1.8812909983952953e-06, "loss": 0.3717, "step": 3973 }, { "epoch": 8.077235772357724, "grad_norm": 4.782112550361292, "learning_rate": 1.8774491326662837e-06, "loss": 0.419, "step": 3974 }, { "epoch": 8.079268292682928, "grad_norm": 4.293564473378306, "learning_rate": 1.8736107873314946e-06, "loss": 0.2874, "step": 3975 }, { "epoch": 8.08130081300813, "grad_norm": 5.024357525947435, "learning_rate": 1.869775964054501e-06, "loss": 0.5017, "step": 3976 }, { "epoch": 8.083333333333334, "grad_norm": 4.690663132240715, "learning_rate": 1.865944664497359e-06, "loss": 0.5083, "step": 3977 }, { "epoch": 8.085365853658537, "grad_norm": 5.263811491799118, "learning_rate": 1.86211689032059e-06, "loss": 0.3866, "step": 3978 }, { "epoch": 8.08739837398374, "grad_norm": 6.128278416988475, "learning_rate": 1.8582926431831895e-06, "loss": 0.3621, "step": 3979 }, { "epoch": 8.089430894308943, "grad_norm": 4.277600880091014, "learning_rate": 1.8544719247426224e-06, "loss": 0.5714, "step": 3980 }, { "epoch": 8.091463414634147, "grad_norm": 5.773994159961718, "learning_rate": 1.8506547366548255e-06, "loss": 0.4339, "step": 3981 }, { "epoch": 8.09349593495935, "grad_norm": 5.080729429505566, "learning_rate": 1.846841080574211e-06, "loss": 0.403, "step": 3982 }, { "epoch": 8.095528455284553, "grad_norm": 4.883738362285198, "learning_rate": 1.843030958153651e-06, "loss": 0.3126, "step": 3983 }, { "epoch": 8.097560975609756, "grad_norm": 5.360858982467936, "learning_rate": 1.8392243710444911e-06, "loss": 0.4568, "step": 3984 }, { "epoch": 8.09959349593496, "grad_norm": 4.764318853815818, "learning_rate": 1.8354213208965455e-06, "loss": 0.4922, "step": 3985 }, { "epoch": 8.101626016260163, "grad_norm": 4.736319738435438, "learning_rate": 1.8316218093580906e-06, "loss": 0.5452, "step": 3986 }, { "epoch": 8.103658536585366, "grad_norm": 4.08785984383511, "learning_rate": 1.8278258380758762e-06, "loss": 0.457, "step": 3987 }, { "epoch": 8.105691056910569, "grad_norm": 4.89089973981868, "learning_rate": 1.8240334086951117e-06, "loss": 0.4466, "step": 3988 }, { "epoch": 8.107723577235772, "grad_norm": 4.6958452067821845, "learning_rate": 1.8202445228594755e-06, "loss": 0.4034, "step": 3989 }, { "epoch": 8.109756097560975, "grad_norm": 5.693867664528515, "learning_rate": 1.8164591822111067e-06, "loss": 0.4977, "step": 3990 }, { "epoch": 8.111788617886178, "grad_norm": 4.974767847101418, "learning_rate": 1.812677388390608e-06, "loss": 0.6608, "step": 3991 }, { "epoch": 8.113821138211382, "grad_norm": 4.292127572483456, "learning_rate": 1.8088991430370506e-06, "loss": 0.429, "step": 3992 }, { "epoch": 8.115853658536585, "grad_norm": 4.7278443227242555, "learning_rate": 1.8051244477879614e-06, "loss": 0.3897, "step": 3993 }, { "epoch": 8.117886178861788, "grad_norm": 4.604185687167935, "learning_rate": 1.8013533042793308e-06, "loss": 0.4017, "step": 3994 }, { "epoch": 8.119918699186991, "grad_norm": 4.754653825891211, "learning_rate": 1.7975857141456098e-06, "loss": 0.4688, "step": 3995 }, { "epoch": 8.121951219512194, "grad_norm": 4.961510546131164, "learning_rate": 1.7938216790197071e-06, "loss": 0.406, "step": 3996 }, { "epoch": 8.123983739837398, "grad_norm": 5.246906075028827, "learning_rate": 1.7900612005329977e-06, "loss": 0.4538, "step": 3997 }, { "epoch": 8.126016260162602, "grad_norm": 4.671026933500926, "learning_rate": 1.7863042803153074e-06, "loss": 0.4166, "step": 3998 }, { "epoch": 8.128048780487806, "grad_norm": 4.865349668343228, "learning_rate": 1.7825509199949243e-06, "loss": 0.6003, "step": 3999 }, { "epoch": 8.130081300813009, "grad_norm": 4.922965288169278, "learning_rate": 1.77880112119859e-06, "loss": 0.4211, "step": 4000 }, { "epoch": 8.132113821138212, "grad_norm": 5.1457979437197965, "learning_rate": 1.7750548855515026e-06, "loss": 0.4905, "step": 4001 }, { "epoch": 8.134146341463415, "grad_norm": 4.154918066114621, "learning_rate": 1.7713122146773264e-06, "loss": 0.4127, "step": 4002 }, { "epoch": 8.136178861788618, "grad_norm": 5.110634706131806, "learning_rate": 1.767573110198162e-06, "loss": 0.4597, "step": 4003 }, { "epoch": 8.138211382113822, "grad_norm": 4.8023874064711976, "learning_rate": 1.7638375737345804e-06, "loss": 0.4177, "step": 4004 }, { "epoch": 8.140243902439025, "grad_norm": 4.948695423804456, "learning_rate": 1.7601056069055988e-06, "loss": 0.3869, "step": 4005 }, { "epoch": 8.142276422764228, "grad_norm": 5.456550594781418, "learning_rate": 1.756377211328687e-06, "loss": 0.4724, "step": 4006 }, { "epoch": 8.144308943089431, "grad_norm": 5.908975966166486, "learning_rate": 1.7526523886197745e-06, "loss": 0.4841, "step": 4007 }, { "epoch": 8.146341463414634, "grad_norm": 4.9353611104568555, "learning_rate": 1.7489311403932274e-06, "loss": 0.4201, "step": 4008 }, { "epoch": 8.148373983739837, "grad_norm": 4.456935591945545, "learning_rate": 1.7452134682618782e-06, "loss": 0.4324, "step": 4009 }, { "epoch": 8.15040650406504, "grad_norm": 4.627002769039866, "learning_rate": 1.7414993738370023e-06, "loss": 0.3754, "step": 4010 }, { "epoch": 8.152439024390244, "grad_norm": 5.887067062562263, "learning_rate": 1.737788858728321e-06, "loss": 0.4801, "step": 4011 }, { "epoch": 8.154471544715447, "grad_norm": 4.368047473040722, "learning_rate": 1.7340819245440166e-06, "loss": 0.464, "step": 4012 }, { "epoch": 8.15650406504065, "grad_norm": 4.751948641517524, "learning_rate": 1.7303785728907009e-06, "loss": 0.5079, "step": 4013 }, { "epoch": 8.158536585365853, "grad_norm": 5.048390971801749, "learning_rate": 1.7266788053734507e-06, "loss": 0.4105, "step": 4014 }, { "epoch": 8.160569105691057, "grad_norm": 4.740890374883191, "learning_rate": 1.722982623595779e-06, "loss": 0.4388, "step": 4015 }, { "epoch": 8.16260162601626, "grad_norm": 4.84326100815281, "learning_rate": 1.7192900291596493e-06, "loss": 0.476, "step": 4016 }, { "epoch": 8.164634146341463, "grad_norm": 4.426561418918474, "learning_rate": 1.7156010236654686e-06, "loss": 0.4624, "step": 4017 }, { "epoch": 8.166666666666666, "grad_norm": 5.646038484795159, "learning_rate": 1.7119156087120836e-06, "loss": 0.5519, "step": 4018 }, { "epoch": 8.16869918699187, "grad_norm": 4.752224282042025, "learning_rate": 1.7082337858967968e-06, "loss": 0.4771, "step": 4019 }, { "epoch": 8.170731707317072, "grad_norm": 4.754367802285472, "learning_rate": 1.7045555568153415e-06, "loss": 0.5408, "step": 4020 }, { "epoch": 8.172764227642276, "grad_norm": 4.7204128153072, "learning_rate": 1.700880923061906e-06, "loss": 0.4123, "step": 4021 }, { "epoch": 8.17479674796748, "grad_norm": 5.010283347845906, "learning_rate": 1.6972098862291054e-06, "loss": 0.3816, "step": 4022 }, { "epoch": 8.176829268292684, "grad_norm": 5.324888509564224, "learning_rate": 1.6935424479080042e-06, "loss": 0.544, "step": 4023 }, { "epoch": 8.178861788617887, "grad_norm": 4.928174692632644, "learning_rate": 1.6898786096881104e-06, "loss": 0.4686, "step": 4024 }, { "epoch": 8.18089430894309, "grad_norm": 5.4245769107877315, "learning_rate": 1.6862183731573644e-06, "loss": 0.4398, "step": 4025 }, { "epoch": 8.182926829268293, "grad_norm": 5.017231659517912, "learning_rate": 1.6825617399021555e-06, "loss": 0.4785, "step": 4026 }, { "epoch": 8.184959349593496, "grad_norm": 5.290755199225435, "learning_rate": 1.6789087115072966e-06, "loss": 0.4287, "step": 4027 }, { "epoch": 8.1869918699187, "grad_norm": 5.7104812126694195, "learning_rate": 1.6752592895560493e-06, "loss": 0.5562, "step": 4028 }, { "epoch": 8.189024390243903, "grad_norm": 4.834481239454836, "learning_rate": 1.6716134756301117e-06, "loss": 0.4127, "step": 4029 }, { "epoch": 8.191056910569106, "grad_norm": 5.251294766112467, "learning_rate": 1.667971271309613e-06, "loss": 0.5342, "step": 4030 }, { "epoch": 8.19308943089431, "grad_norm": 5.453295112787276, "learning_rate": 1.6643326781731238e-06, "loss": 0.4201, "step": 4031 }, { "epoch": 8.195121951219512, "grad_norm": 5.438418333834926, "learning_rate": 1.6606976977976408e-06, "loss": 0.5437, "step": 4032 }, { "epoch": 8.197154471544716, "grad_norm": 5.103630512131774, "learning_rate": 1.6570663317586067e-06, "loss": 0.6226, "step": 4033 }, { "epoch": 8.199186991869919, "grad_norm": 6.19922529076377, "learning_rate": 1.6534385816298903e-06, "loss": 0.4363, "step": 4034 }, { "epoch": 8.201219512195122, "grad_norm": 5.477521535357576, "learning_rate": 1.6498144489837932e-06, "loss": 0.3953, "step": 4035 }, { "epoch": 8.203252032520325, "grad_norm": 4.677890340206304, "learning_rate": 1.6461939353910494e-06, "loss": 0.3613, "step": 4036 }, { "epoch": 8.205284552845528, "grad_norm": 5.650956025052605, "learning_rate": 1.6425770424208253e-06, "loss": 0.4698, "step": 4037 }, { "epoch": 8.207317073170731, "grad_norm": 4.951350890877876, "learning_rate": 1.6389637716407225e-06, "loss": 0.4699, "step": 4038 }, { "epoch": 8.209349593495935, "grad_norm": 4.83973424953234, "learning_rate": 1.6353541246167658e-06, "loss": 0.5019, "step": 4039 }, { "epoch": 8.211382113821138, "grad_norm": 5.107321247767607, "learning_rate": 1.631748102913412e-06, "loss": 0.5632, "step": 4040 }, { "epoch": 8.213414634146341, "grad_norm": 5.697299809468287, "learning_rate": 1.6281457080935458e-06, "loss": 0.4, "step": 4041 }, { "epoch": 8.215447154471544, "grad_norm": 5.132162824738276, "learning_rate": 1.62454694171848e-06, "loss": 0.3884, "step": 4042 }, { "epoch": 8.217479674796747, "grad_norm": 4.507495798164821, "learning_rate": 1.62095180534796e-06, "loss": 0.3696, "step": 4043 }, { "epoch": 8.21951219512195, "grad_norm": 4.42547812266712, "learning_rate": 1.6173603005401505e-06, "loss": 0.337, "step": 4044 }, { "epoch": 8.221544715447154, "grad_norm": 5.1560854006109915, "learning_rate": 1.6137724288516466e-06, "loss": 0.5306, "step": 4045 }, { "epoch": 8.223577235772357, "grad_norm": 4.941627605723726, "learning_rate": 1.6101881918374673e-06, "loss": 0.4568, "step": 4046 }, { "epoch": 8.225609756097562, "grad_norm": 5.19835953266889, "learning_rate": 1.6066075910510525e-06, "loss": 0.3911, "step": 4047 }, { "epoch": 8.227642276422765, "grad_norm": 5.150811061059321, "learning_rate": 1.6030306280442764e-06, "loss": 0.3901, "step": 4048 }, { "epoch": 8.229674796747968, "grad_norm": 4.80034108924502, "learning_rate": 1.5994573043674277e-06, "loss": 0.3163, "step": 4049 }, { "epoch": 8.231707317073171, "grad_norm": 5.659416609245703, "learning_rate": 1.5958876215692209e-06, "loss": 0.5322, "step": 4050 }, { "epoch": 8.233739837398375, "grad_norm": 4.417758701806553, "learning_rate": 1.5923215811967908e-06, "loss": 0.4362, "step": 4051 }, { "epoch": 8.235772357723578, "grad_norm": 5.502653302010915, "learning_rate": 1.588759184795694e-06, "loss": 0.362, "step": 4052 }, { "epoch": 8.237804878048781, "grad_norm": 5.052919043425897, "learning_rate": 1.5852004339099114e-06, "loss": 0.4603, "step": 4053 }, { "epoch": 8.239837398373984, "grad_norm": 5.613594509745576, "learning_rate": 1.581645330081839e-06, "loss": 0.3915, "step": 4054 }, { "epoch": 8.241869918699187, "grad_norm": 5.043059012184215, "learning_rate": 1.5780938748522945e-06, "loss": 0.4809, "step": 4055 }, { "epoch": 8.24390243902439, "grad_norm": 6.400952058925732, "learning_rate": 1.574546069760514e-06, "loss": 0.4526, "step": 4056 }, { "epoch": 8.245934959349594, "grad_norm": 5.065123751495993, "learning_rate": 1.5710019163441493e-06, "loss": 0.4232, "step": 4057 }, { "epoch": 8.247967479674797, "grad_norm": 4.953456444610346, "learning_rate": 1.5674614161392753e-06, "loss": 0.5534, "step": 4058 }, { "epoch": 8.25, "grad_norm": 5.309957689227767, "learning_rate": 1.5639245706803796e-06, "loss": 0.5338, "step": 4059 }, { "epoch": 8.252032520325203, "grad_norm": 5.673496648968486, "learning_rate": 1.5603913815003634e-06, "loss": 0.5303, "step": 4060 }, { "epoch": 8.254065040650406, "grad_norm": 4.217806508608546, "learning_rate": 1.5568618501305487e-06, "loss": 0.3525, "step": 4061 }, { "epoch": 8.25609756097561, "grad_norm": 4.945505179908956, "learning_rate": 1.5533359781006664e-06, "loss": 0.5725, "step": 4062 }, { "epoch": 8.258130081300813, "grad_norm": 5.179673912428328, "learning_rate": 1.549813766938869e-06, "loss": 0.4786, "step": 4063 }, { "epoch": 8.260162601626016, "grad_norm": 5.662175235954827, "learning_rate": 1.5462952181717117e-06, "loss": 0.5111, "step": 4064 }, { "epoch": 8.262195121951219, "grad_norm": 5.038561900539951, "learning_rate": 1.5427803333241741e-06, "loss": 0.501, "step": 4065 }, { "epoch": 8.264227642276422, "grad_norm": 6.0519674911663754, "learning_rate": 1.5392691139196403e-06, "loss": 0.4229, "step": 4066 }, { "epoch": 8.266260162601625, "grad_norm": 4.484988898811348, "learning_rate": 1.535761561479905e-06, "loss": 0.2858, "step": 4067 }, { "epoch": 8.268292682926829, "grad_norm": 5.482068756670487, "learning_rate": 1.532257677525183e-06, "loss": 0.4788, "step": 4068 }, { "epoch": 8.270325203252032, "grad_norm": 4.306994586069906, "learning_rate": 1.5287574635740832e-06, "loss": 0.4894, "step": 4069 }, { "epoch": 8.272357723577235, "grad_norm": 5.709552516953369, "learning_rate": 1.5252609211436397e-06, "loss": 0.5168, "step": 4070 }, { "epoch": 8.274390243902438, "grad_norm": 5.769494362758992, "learning_rate": 1.5217680517492883e-06, "loss": 0.514, "step": 4071 }, { "epoch": 8.276422764227643, "grad_norm": 5.308279175933005, "learning_rate": 1.5182788569048689e-06, "loss": 0.496, "step": 4072 }, { "epoch": 8.278455284552846, "grad_norm": 6.2796191278751206, "learning_rate": 1.514793338122641e-06, "loss": 0.4653, "step": 4073 }, { "epoch": 8.28048780487805, "grad_norm": 5.46484111907019, "learning_rate": 1.5113114969132536e-06, "loss": 0.5203, "step": 4074 }, { "epoch": 8.282520325203253, "grad_norm": 4.496576144992112, "learning_rate": 1.5078333347857776e-06, "loss": 0.3992, "step": 4075 }, { "epoch": 8.284552845528456, "grad_norm": 4.74872834739719, "learning_rate": 1.5043588532476827e-06, "loss": 0.4521, "step": 4076 }, { "epoch": 8.286585365853659, "grad_norm": 5.30858879257781, "learning_rate": 1.5008880538048408e-06, "loss": 0.4992, "step": 4077 }, { "epoch": 8.288617886178862, "grad_norm": 5.3604361875705155, "learning_rate": 1.4974209379615335e-06, "loss": 0.484, "step": 4078 }, { "epoch": 8.290650406504065, "grad_norm": 4.911566966046322, "learning_rate": 1.4939575072204405e-06, "loss": 0.4021, "step": 4079 }, { "epoch": 8.292682926829269, "grad_norm": 4.528027843323043, "learning_rate": 1.49049776308265e-06, "loss": 0.426, "step": 4080 }, { "epoch": 8.294715447154472, "grad_norm": 4.2421077099871995, "learning_rate": 1.4870417070476485e-06, "loss": 0.4305, "step": 4081 }, { "epoch": 8.296747967479675, "grad_norm": 5.242348971665217, "learning_rate": 1.483589340613324e-06, "loss": 0.561, "step": 4082 }, { "epoch": 8.298780487804878, "grad_norm": 4.342920408561089, "learning_rate": 1.4801406652759687e-06, "loss": 0.4049, "step": 4083 }, { "epoch": 8.300813008130081, "grad_norm": 5.03448363480964, "learning_rate": 1.476695682530268e-06, "loss": 0.5326, "step": 4084 }, { "epoch": 8.302845528455284, "grad_norm": 4.0196913146273, "learning_rate": 1.4732543938693167e-06, "loss": 0.4056, "step": 4085 }, { "epoch": 8.304878048780488, "grad_norm": 4.79674005951828, "learning_rate": 1.4698168007846024e-06, "loss": 0.3294, "step": 4086 }, { "epoch": 8.30691056910569, "grad_norm": 5.37482942729978, "learning_rate": 1.4663829047660105e-06, "loss": 0.3407, "step": 4087 }, { "epoch": 8.308943089430894, "grad_norm": 4.510826519387661, "learning_rate": 1.4629527073018267e-06, "loss": 0.4292, "step": 4088 }, { "epoch": 8.310975609756097, "grad_norm": 4.660958196950948, "learning_rate": 1.459526209878731e-06, "loss": 0.4095, "step": 4089 }, { "epoch": 8.3130081300813, "grad_norm": 5.023528014329018, "learning_rate": 1.4561034139818043e-06, "loss": 0.4655, "step": 4090 }, { "epoch": 8.315040650406504, "grad_norm": 4.555835978837808, "learning_rate": 1.4526843210945186e-06, "loss": 0.5161, "step": 4091 }, { "epoch": 8.317073170731707, "grad_norm": 4.66961323185409, "learning_rate": 1.449268932698743e-06, "loss": 0.3926, "step": 4092 }, { "epoch": 8.31910569105691, "grad_norm": 5.234778496659939, "learning_rate": 1.4458572502747414e-06, "loss": 0.3123, "step": 4093 }, { "epoch": 8.321138211382113, "grad_norm": 5.409728709153953, "learning_rate": 1.4424492753011677e-06, "loss": 0.4409, "step": 4094 }, { "epoch": 8.323170731707316, "grad_norm": 4.983274883549267, "learning_rate": 1.4390450092550768e-06, "loss": 0.4413, "step": 4095 }, { "epoch": 8.32520325203252, "grad_norm": 5.455213876645116, "learning_rate": 1.4356444536119085e-06, "loss": 0.4513, "step": 4096 }, { "epoch": 8.327235772357724, "grad_norm": 4.7662473591650825, "learning_rate": 1.4322476098454986e-06, "loss": 0.3993, "step": 4097 }, { "epoch": 8.329268292682928, "grad_norm": 5.0121786391371055, "learning_rate": 1.4288544794280724e-06, "loss": 0.4258, "step": 4098 }, { "epoch": 8.33130081300813, "grad_norm": 5.878037429815469, "learning_rate": 1.4254650638302436e-06, "loss": 0.4405, "step": 4099 }, { "epoch": 8.333333333333334, "grad_norm": 5.085325407235772, "learning_rate": 1.422079364521024e-06, "loss": 0.3761, "step": 4100 }, { "epoch": 8.335365853658537, "grad_norm": 5.316538812888184, "learning_rate": 1.4186973829678075e-06, "loss": 0.5439, "step": 4101 }, { "epoch": 8.33739837398374, "grad_norm": 4.58013704985736, "learning_rate": 1.4153191206363758e-06, "loss": 0.3775, "step": 4102 }, { "epoch": 8.339430894308943, "grad_norm": 6.060688233069579, "learning_rate": 1.4119445789909048e-06, "loss": 0.3468, "step": 4103 }, { "epoch": 8.341463414634147, "grad_norm": 4.6028025263391354, "learning_rate": 1.4085737594939497e-06, "loss": 0.3484, "step": 4104 }, { "epoch": 8.34349593495935, "grad_norm": 4.798593611345974, "learning_rate": 1.405206663606462e-06, "loss": 0.5152, "step": 4105 }, { "epoch": 8.345528455284553, "grad_norm": 5.505263920027516, "learning_rate": 1.4018432927877735e-06, "loss": 0.4792, "step": 4106 }, { "epoch": 8.347560975609756, "grad_norm": 4.349596707152931, "learning_rate": 1.3984836484956011e-06, "loss": 0.501, "step": 4107 }, { "epoch": 8.34959349593496, "grad_norm": 4.905158777127327, "learning_rate": 1.3951277321860468e-06, "loss": 0.4647, "step": 4108 }, { "epoch": 8.351626016260163, "grad_norm": 5.304273633667205, "learning_rate": 1.3917755453136017e-06, "loss": 0.4882, "step": 4109 }, { "epoch": 8.353658536585366, "grad_norm": 4.965419444818217, "learning_rate": 1.388427089331138e-06, "loss": 0.3977, "step": 4110 }, { "epoch": 8.355691056910569, "grad_norm": 5.144989233803182, "learning_rate": 1.3850823656899015e-06, "loss": 0.4663, "step": 4111 }, { "epoch": 8.357723577235772, "grad_norm": 5.5349105782833234, "learning_rate": 1.381741375839537e-06, "loss": 0.4496, "step": 4112 }, { "epoch": 8.359756097560975, "grad_norm": 5.117765380509526, "learning_rate": 1.3784041212280562e-06, "loss": 0.426, "step": 4113 }, { "epoch": 8.361788617886178, "grad_norm": 5.709735222737947, "learning_rate": 1.3750706033018658e-06, "loss": 0.4179, "step": 4114 }, { "epoch": 8.363821138211382, "grad_norm": 4.979883181994664, "learning_rate": 1.3717408235057427e-06, "loss": 0.4781, "step": 4115 }, { "epoch": 8.365853658536585, "grad_norm": 6.025117520351054, "learning_rate": 1.3684147832828409e-06, "loss": 0.4759, "step": 4116 }, { "epoch": 8.367886178861788, "grad_norm": 5.11995207905486, "learning_rate": 1.3650924840747072e-06, "loss": 0.5122, "step": 4117 }, { "epoch": 8.369918699186991, "grad_norm": 4.384650891455989, "learning_rate": 1.3617739273212527e-06, "loss": 0.3651, "step": 4118 }, { "epoch": 8.371951219512194, "grad_norm": 6.130488952843612, "learning_rate": 1.3584591144607774e-06, "loss": 0.3891, "step": 4119 }, { "epoch": 8.373983739837398, "grad_norm": 5.34188897400407, "learning_rate": 1.355148046929956e-06, "loss": 0.5311, "step": 4120 }, { "epoch": 8.376016260162602, "grad_norm": 4.517612053870646, "learning_rate": 1.35184072616383e-06, "loss": 0.4705, "step": 4121 }, { "epoch": 8.378048780487806, "grad_norm": 5.309253210093655, "learning_rate": 1.3485371535958314e-06, "loss": 0.4948, "step": 4122 }, { "epoch": 8.380081300813009, "grad_norm": 4.608813546054089, "learning_rate": 1.345237330657758e-06, "loss": 0.3714, "step": 4123 }, { "epoch": 8.382113821138212, "grad_norm": 4.879722669687005, "learning_rate": 1.3419412587797908e-06, "loss": 0.3773, "step": 4124 }, { "epoch": 8.384146341463415, "grad_norm": 5.5282972314485646, "learning_rate": 1.3386489393904735e-06, "loss": 0.4015, "step": 4125 }, { "epoch": 8.386178861788618, "grad_norm": 4.633486193445139, "learning_rate": 1.3353603739167354e-06, "loss": 0.4506, "step": 4126 }, { "epoch": 8.388211382113822, "grad_norm": 4.864019701808656, "learning_rate": 1.332075563783871e-06, "loss": 0.3791, "step": 4127 }, { "epoch": 8.390243902439025, "grad_norm": 4.843043002684947, "learning_rate": 1.3287945104155487e-06, "loss": 0.3787, "step": 4128 }, { "epoch": 8.392276422764228, "grad_norm": 4.818102828443174, "learning_rate": 1.3255172152338147e-06, "loss": 0.4262, "step": 4129 }, { "epoch": 8.394308943089431, "grad_norm": 4.96326359588211, "learning_rate": 1.3222436796590731e-06, "loss": 0.4796, "step": 4130 }, { "epoch": 8.396341463414634, "grad_norm": 4.504863303869217, "learning_rate": 1.3189739051101135e-06, "loss": 0.3543, "step": 4131 }, { "epoch": 8.398373983739837, "grad_norm": 4.939342529433042, "learning_rate": 1.3157078930040856e-06, "loss": 0.3878, "step": 4132 }, { "epoch": 8.40040650406504, "grad_norm": 5.626244315512982, "learning_rate": 1.3124456447565104e-06, "loss": 0.4528, "step": 4133 }, { "epoch": 8.402439024390244, "grad_norm": 5.170539977112848, "learning_rate": 1.3091871617812856e-06, "loss": 0.4447, "step": 4134 }, { "epoch": 8.404471544715447, "grad_norm": 4.427945350127277, "learning_rate": 1.3059324454906596e-06, "loss": 0.4418, "step": 4135 }, { "epoch": 8.40650406504065, "grad_norm": 4.0053013828059125, "learning_rate": 1.3026814972952674e-06, "loss": 0.3095, "step": 4136 }, { "epoch": 8.408536585365853, "grad_norm": 4.748170891811588, "learning_rate": 1.2994343186040982e-06, "loss": 0.5202, "step": 4137 }, { "epoch": 8.410569105691057, "grad_norm": 5.139534092623926, "learning_rate": 1.2961909108245119e-06, "loss": 0.4877, "step": 4138 }, { "epoch": 8.41260162601626, "grad_norm": 5.124392566166113, "learning_rate": 1.2929512753622342e-06, "loss": 0.4594, "step": 4139 }, { "epoch": 8.414634146341463, "grad_norm": 5.0061997599712935, "learning_rate": 1.2897154136213542e-06, "loss": 0.4234, "step": 4140 }, { "epoch": 8.416666666666666, "grad_norm": 4.84154759008945, "learning_rate": 1.2864833270043286e-06, "loss": 0.4816, "step": 4141 }, { "epoch": 8.41869918699187, "grad_norm": 4.947507470154593, "learning_rate": 1.283255016911974e-06, "loss": 0.5527, "step": 4142 }, { "epoch": 8.420731707317072, "grad_norm": 5.494297610466342, "learning_rate": 1.280030484743473e-06, "loss": 0.4197, "step": 4143 }, { "epoch": 8.422764227642276, "grad_norm": 4.386484196814274, "learning_rate": 1.2768097318963701e-06, "loss": 0.3752, "step": 4144 }, { "epoch": 8.42479674796748, "grad_norm": 5.643136185464088, "learning_rate": 1.2735927597665687e-06, "loss": 0.5116, "step": 4145 }, { "epoch": 8.426829268292684, "grad_norm": 5.137194804677528, "learning_rate": 1.2703795697483412e-06, "loss": 0.4825, "step": 4146 }, { "epoch": 8.428861788617887, "grad_norm": 4.816273090569064, "learning_rate": 1.2671701632343137e-06, "loss": 0.5528, "step": 4147 }, { "epoch": 8.43089430894309, "grad_norm": 5.392889983145439, "learning_rate": 1.2639645416154744e-06, "loss": 0.4127, "step": 4148 }, { "epoch": 8.432926829268293, "grad_norm": 4.5800355996151545, "learning_rate": 1.2607627062811733e-06, "loss": 0.4215, "step": 4149 }, { "epoch": 8.434959349593496, "grad_norm": 4.598880646165203, "learning_rate": 1.2575646586191126e-06, "loss": 0.4154, "step": 4150 }, { "epoch": 8.4369918699187, "grad_norm": 4.885383026832489, "learning_rate": 1.2543704000153645e-06, "loss": 0.4857, "step": 4151 }, { "epoch": 8.439024390243903, "grad_norm": 4.190661689804591, "learning_rate": 1.2511799318543493e-06, "loss": 0.3467, "step": 4152 }, { "epoch": 8.441056910569106, "grad_norm": 4.146464924102766, "learning_rate": 1.247993255518849e-06, "loss": 0.4831, "step": 4153 }, { "epoch": 8.44308943089431, "grad_norm": 5.923913415723275, "learning_rate": 1.2448103723899984e-06, "loss": 0.5352, "step": 4154 }, { "epoch": 8.445121951219512, "grad_norm": 4.631497385513207, "learning_rate": 1.24163128384729e-06, "loss": 0.3668, "step": 4155 }, { "epoch": 8.447154471544716, "grad_norm": 5.220415775123457, "learning_rate": 1.2384559912685768e-06, "loss": 0.5681, "step": 4156 }, { "epoch": 8.449186991869919, "grad_norm": 4.621315223684983, "learning_rate": 1.2352844960300592e-06, "loss": 0.4043, "step": 4157 }, { "epoch": 8.451219512195122, "grad_norm": 4.144552125304515, "learning_rate": 1.2321167995062954e-06, "loss": 0.2971, "step": 4158 }, { "epoch": 8.453252032520325, "grad_norm": 5.147319460215155, "learning_rate": 1.2289529030701963e-06, "loss": 0.4882, "step": 4159 }, { "epoch": 8.455284552845528, "grad_norm": 4.760554686631416, "learning_rate": 1.2257928080930236e-06, "loss": 0.3927, "step": 4160 }, { "epoch": 8.457317073170731, "grad_norm": 4.867317636789221, "learning_rate": 1.2226365159443976e-06, "loss": 0.3485, "step": 4161 }, { "epoch": 8.459349593495935, "grad_norm": 4.331542266166103, "learning_rate": 1.2194840279922871e-06, "loss": 0.3592, "step": 4162 }, { "epoch": 8.461382113821138, "grad_norm": 4.625857808754792, "learning_rate": 1.216335345603008e-06, "loss": 0.4341, "step": 4163 }, { "epoch": 8.463414634146341, "grad_norm": 4.796490077186588, "learning_rate": 1.2131904701412345e-06, "loss": 0.5389, "step": 4164 }, { "epoch": 8.465447154471544, "grad_norm": 5.191683455597217, "learning_rate": 1.2100494029699805e-06, "loss": 0.5015, "step": 4165 }, { "epoch": 8.467479674796747, "grad_norm": 4.581740105177654, "learning_rate": 1.2069121454506238e-06, "loss": 0.4692, "step": 4166 }, { "epoch": 8.46951219512195, "grad_norm": 4.25453659208822, "learning_rate": 1.203778698942879e-06, "loss": 0.4318, "step": 4167 }, { "epoch": 8.471544715447154, "grad_norm": 4.872952608959665, "learning_rate": 1.2006490648048118e-06, "loss": 0.4377, "step": 4168 }, { "epoch": 8.473577235772357, "grad_norm": 4.737146006102281, "learning_rate": 1.1975232443928397e-06, "loss": 0.427, "step": 4169 }, { "epoch": 8.475609756097562, "grad_norm": 4.370697931777999, "learning_rate": 1.1944012390617198e-06, "loss": 0.4452, "step": 4170 }, { "epoch": 8.477642276422765, "grad_norm": 4.907395597713298, "learning_rate": 1.1912830501645667e-06, "loss": 0.5002, "step": 4171 }, { "epoch": 8.479674796747968, "grad_norm": 4.3941160839046, "learning_rate": 1.1881686790528279e-06, "loss": 0.4238, "step": 4172 }, { "epoch": 8.481707317073171, "grad_norm": 5.218786364964129, "learning_rate": 1.1850581270763062e-06, "loss": 0.4591, "step": 4173 }, { "epoch": 8.483739837398375, "grad_norm": 4.537600296518425, "learning_rate": 1.1819513955831462e-06, "loss": 0.3903, "step": 4174 }, { "epoch": 8.485772357723578, "grad_norm": 5.178290318252191, "learning_rate": 1.1788484859198334e-06, "loss": 0.3438, "step": 4175 }, { "epoch": 8.487804878048781, "grad_norm": 4.896626736840315, "learning_rate": 1.1757493994312052e-06, "loss": 0.4009, "step": 4176 }, { "epoch": 8.489837398373984, "grad_norm": 6.384333075196905, "learning_rate": 1.1726541374604282e-06, "loss": 0.5023, "step": 4177 }, { "epoch": 8.491869918699187, "grad_norm": 5.257455107521733, "learning_rate": 1.1695627013490262e-06, "loss": 0.4102, "step": 4178 }, { "epoch": 8.49390243902439, "grad_norm": 5.871794411719829, "learning_rate": 1.1664750924368573e-06, "loss": 0.4236, "step": 4179 }, { "epoch": 8.495934959349594, "grad_norm": 5.792528577673828, "learning_rate": 1.1633913120621188e-06, "loss": 0.4302, "step": 4180 }, { "epoch": 8.497967479674797, "grad_norm": 5.211964441787625, "learning_rate": 1.160311361561357e-06, "loss": 0.4712, "step": 4181 }, { "epoch": 8.5, "grad_norm": 4.485243017721971, "learning_rate": 1.1572352422694455e-06, "loss": 0.4075, "step": 4182 }, { "epoch": 8.502032520325203, "grad_norm": 5.277583065157922, "learning_rate": 1.154162955519612e-06, "loss": 0.5236, "step": 4183 }, { "epoch": 8.504065040650406, "grad_norm": 5.226338903864239, "learning_rate": 1.151094502643414e-06, "loss": 0.414, "step": 4184 }, { "epoch": 8.50609756097561, "grad_norm": 4.745212934433794, "learning_rate": 1.1480298849707483e-06, "loss": 0.4029, "step": 4185 }, { "epoch": 8.508130081300813, "grad_norm": 4.649416952446947, "learning_rate": 1.1449691038298516e-06, "loss": 0.3783, "step": 4186 }, { "epoch": 8.510162601626016, "grad_norm": 4.888781288898997, "learning_rate": 1.141912160547294e-06, "loss": 0.5058, "step": 4187 }, { "epoch": 8.512195121951219, "grad_norm": 4.763963142378667, "learning_rate": 1.1388590564479895e-06, "loss": 0.4692, "step": 4188 }, { "epoch": 8.514227642276422, "grad_norm": 4.145146541387193, "learning_rate": 1.1358097928551792e-06, "loss": 0.405, "step": 4189 }, { "epoch": 8.516260162601625, "grad_norm": 4.108154630492136, "learning_rate": 1.1327643710904523e-06, "loss": 0.2479, "step": 4190 }, { "epoch": 8.518292682926829, "grad_norm": 4.442055692709415, "learning_rate": 1.1297227924737164e-06, "loss": 0.3678, "step": 4191 }, { "epoch": 8.520325203252032, "grad_norm": 4.946738407675823, "learning_rate": 1.1266850583232224e-06, "loss": 0.4448, "step": 4192 }, { "epoch": 8.522357723577235, "grad_norm": 4.758608656347139, "learning_rate": 1.1236511699555596e-06, "loss": 0.4171, "step": 4193 }, { "epoch": 8.524390243902438, "grad_norm": 5.578754588913169, "learning_rate": 1.1206211286856405e-06, "loss": 0.4273, "step": 4194 }, { "epoch": 8.526422764227643, "grad_norm": 5.243003846321519, "learning_rate": 1.1175949358267214e-06, "loss": 0.419, "step": 4195 }, { "epoch": 8.528455284552846, "grad_norm": 6.13830035547345, "learning_rate": 1.1145725926903772e-06, "loss": 0.3718, "step": 4196 }, { "epoch": 8.53048780487805, "grad_norm": 4.35804627670907, "learning_rate": 1.1115541005865237e-06, "loss": 0.4272, "step": 4197 }, { "epoch": 8.532520325203253, "grad_norm": 4.8147059803286005, "learning_rate": 1.1085394608234067e-06, "loss": 0.4236, "step": 4198 }, { "epoch": 8.534552845528456, "grad_norm": 4.861726175699086, "learning_rate": 1.1055286747076e-06, "loss": 0.5347, "step": 4199 }, { "epoch": 8.536585365853659, "grad_norm": 4.728253344813897, "learning_rate": 1.1025217435440116e-06, "loss": 0.4071, "step": 4200 }, { "epoch": 8.538617886178862, "grad_norm": 5.695858336638396, "learning_rate": 1.0995186686358682e-06, "loss": 0.4361, "step": 4201 }, { "epoch": 8.540650406504065, "grad_norm": 5.248562705364429, "learning_rate": 1.0965194512847387e-06, "loss": 0.4226, "step": 4202 }, { "epoch": 8.542682926829269, "grad_norm": 5.265100144722205, "learning_rate": 1.093524092790511e-06, "loss": 0.4522, "step": 4203 }, { "epoch": 8.544715447154472, "grad_norm": 4.7597254896953896, "learning_rate": 1.0905325944514034e-06, "loss": 0.3293, "step": 4204 }, { "epoch": 8.546747967479675, "grad_norm": 5.894825628149396, "learning_rate": 1.0875449575639618e-06, "loss": 0.5369, "step": 4205 }, { "epoch": 8.548780487804878, "grad_norm": 5.143184550889033, "learning_rate": 1.0845611834230551e-06, "loss": 0.397, "step": 4206 }, { "epoch": 8.550813008130081, "grad_norm": 4.5582941525880685, "learning_rate": 1.0815812733218845e-06, "loss": 0.4708, "step": 4207 }, { "epoch": 8.552845528455284, "grad_norm": 5.736423147540173, "learning_rate": 1.078605228551971e-06, "loss": 0.4867, "step": 4208 }, { "epoch": 8.554878048780488, "grad_norm": 5.8781399842616455, "learning_rate": 1.075633050403162e-06, "loss": 0.5193, "step": 4209 }, { "epoch": 8.55691056910569, "grad_norm": 6.623473410212815, "learning_rate": 1.072664740163628e-06, "loss": 0.3789, "step": 4210 }, { "epoch": 8.558943089430894, "grad_norm": 5.586988659839521, "learning_rate": 1.0697002991198634e-06, "loss": 0.4669, "step": 4211 }, { "epoch": 8.560975609756097, "grad_norm": 4.1214222396341444, "learning_rate": 1.0667397285566893e-06, "loss": 0.4305, "step": 4212 }, { "epoch": 8.5630081300813, "grad_norm": 4.95908974062345, "learning_rate": 1.063783029757246e-06, "loss": 0.546, "step": 4213 }, { "epoch": 8.565040650406504, "grad_norm": 4.664484361493165, "learning_rate": 1.0608302040029938e-06, "loss": 0.4273, "step": 4214 }, { "epoch": 8.567073170731707, "grad_norm": 5.6953020632914555, "learning_rate": 1.0578812525737192e-06, "loss": 0.5234, "step": 4215 }, { "epoch": 8.56910569105691, "grad_norm": 5.098903557848146, "learning_rate": 1.0549361767475241e-06, "loss": 0.3534, "step": 4216 }, { "epoch": 8.571138211382113, "grad_norm": 4.623421101487863, "learning_rate": 1.0519949778008366e-06, "loss": 0.4247, "step": 4217 }, { "epoch": 8.573170731707316, "grad_norm": 4.45028159449285, "learning_rate": 1.0490576570083999e-06, "loss": 0.565, "step": 4218 }, { "epoch": 8.57520325203252, "grad_norm": 5.016047830008406, "learning_rate": 1.0461242156432793e-06, "loss": 0.3978, "step": 4219 }, { "epoch": 8.577235772357724, "grad_norm": 4.9422752120080675, "learning_rate": 1.0431946549768567e-06, "loss": 0.4334, "step": 4220 }, { "epoch": 8.579268292682928, "grad_norm": 5.521770104795977, "learning_rate": 1.0402689762788298e-06, "loss": 0.378, "step": 4221 }, { "epoch": 8.58130081300813, "grad_norm": 4.894544957636206, "learning_rate": 1.0373471808172219e-06, "loss": 0.4997, "step": 4222 }, { "epoch": 8.583333333333334, "grad_norm": 4.967205169686898, "learning_rate": 1.0344292698583658e-06, "loss": 0.4039, "step": 4223 }, { "epoch": 8.585365853658537, "grad_norm": 5.532376412404028, "learning_rate": 1.0315152446669142e-06, "loss": 0.4865, "step": 4224 }, { "epoch": 8.58739837398374, "grad_norm": 4.572570528035204, "learning_rate": 1.0286051065058323e-06, "loss": 0.4998, "step": 4225 }, { "epoch": 8.589430894308943, "grad_norm": 4.548865612588322, "learning_rate": 1.0256988566364034e-06, "loss": 0.4366, "step": 4226 }, { "epoch": 8.591463414634147, "grad_norm": 4.448457610401731, "learning_rate": 1.0227964963182269e-06, "loss": 0.3734, "step": 4227 }, { "epoch": 8.59349593495935, "grad_norm": 4.280585012595993, "learning_rate": 1.019898026809214e-06, "loss": 0.4233, "step": 4228 }, { "epoch": 8.595528455284553, "grad_norm": 4.897693882141388, "learning_rate": 1.0170034493655888e-06, "loss": 0.4596, "step": 4229 }, { "epoch": 8.597560975609756, "grad_norm": 5.611715745853799, "learning_rate": 1.0141127652418914e-06, "loss": 0.6108, "step": 4230 }, { "epoch": 8.59959349593496, "grad_norm": 4.819351643884217, "learning_rate": 1.0112259756909714e-06, "loss": 0.4078, "step": 4231 }, { "epoch": 8.601626016260163, "grad_norm": 5.0967249824700325, "learning_rate": 1.0083430819639962e-06, "loss": 0.4566, "step": 4232 }, { "epoch": 8.603658536585366, "grad_norm": 5.91737709703728, "learning_rate": 1.0054640853104325e-06, "loss": 0.4136, "step": 4233 }, { "epoch": 8.605691056910569, "grad_norm": 5.381442975352092, "learning_rate": 1.0025889869780747e-06, "loss": 0.3787, "step": 4234 }, { "epoch": 8.607723577235772, "grad_norm": 5.373677755550849, "learning_rate": 9.997177882130138e-07, "loss": 0.4073, "step": 4235 }, { "epoch": 8.609756097560975, "grad_norm": 4.7221661510409865, "learning_rate": 9.968504902596566e-07, "loss": 0.4349, "step": 4236 }, { "epoch": 8.611788617886178, "grad_norm": 4.018885280245114, "learning_rate": 9.939870943607222e-07, "loss": 0.2825, "step": 4237 }, { "epoch": 8.613821138211382, "grad_norm": 5.07936022025504, "learning_rate": 9.91127601757228e-07, "loss": 0.3773, "step": 4238 }, { "epoch": 8.615853658536585, "grad_norm": 4.860106064523858, "learning_rate": 9.882720136885126e-07, "loss": 0.4875, "step": 4239 }, { "epoch": 8.617886178861788, "grad_norm": 6.148303653660836, "learning_rate": 9.85420331392214e-07, "loss": 0.5003, "step": 4240 }, { "epoch": 8.619918699186991, "grad_norm": 5.0032126601762945, "learning_rate": 9.825725561042777e-07, "loss": 0.4949, "step": 4241 }, { "epoch": 8.621951219512194, "grad_norm": 5.166879235094712, "learning_rate": 9.797286890589642e-07, "loss": 0.4482, "step": 4242 }, { "epoch": 8.6239837398374, "grad_norm": 4.295590298885517, "learning_rate": 9.768887314888242e-07, "loss": 0.4775, "step": 4243 }, { "epoch": 8.6260162601626, "grad_norm": 4.972732349157325, "learning_rate": 9.74052684624731e-07, "loss": 0.3735, "step": 4244 }, { "epoch": 8.628048780487806, "grad_norm": 4.257050841041059, "learning_rate": 9.712205496958526e-07, "loss": 0.365, "step": 4245 }, { "epoch": 8.630081300813009, "grad_norm": 4.62800269709774, "learning_rate": 9.683923279296637e-07, "loss": 0.5543, "step": 4246 }, { "epoch": 8.632113821138212, "grad_norm": 4.9491499693191585, "learning_rate": 9.65568020551948e-07, "loss": 0.4939, "step": 4247 }, { "epoch": 8.634146341463415, "grad_norm": 4.29922201770183, "learning_rate": 9.62747628786782e-07, "loss": 0.6856, "step": 4248 }, { "epoch": 8.636178861788618, "grad_norm": 4.901090127982818, "learning_rate": 9.59931153856557e-07, "loss": 0.4592, "step": 4249 }, { "epoch": 8.638211382113822, "grad_norm": 4.358763738380932, "learning_rate": 9.571185969819575e-07, "loss": 0.4383, "step": 4250 }, { "epoch": 8.640243902439025, "grad_norm": 5.362558887496368, "learning_rate": 9.54309959381976e-07, "loss": 0.4269, "step": 4251 }, { "epoch": 8.642276422764228, "grad_norm": 4.254253356833036, "learning_rate": 9.515052422739035e-07, "loss": 0.5271, "step": 4252 }, { "epoch": 8.644308943089431, "grad_norm": 4.8819985642449515, "learning_rate": 9.487044468733286e-07, "loss": 0.5289, "step": 4253 }, { "epoch": 8.646341463414634, "grad_norm": 5.011622990002717, "learning_rate": 9.459075743941493e-07, "loss": 0.5485, "step": 4254 }, { "epoch": 8.648373983739837, "grad_norm": 4.933638894410415, "learning_rate": 9.431146260485558e-07, "loss": 0.4196, "step": 4255 }, { "epoch": 8.65040650406504, "grad_norm": 4.546863118777677, "learning_rate": 9.403256030470386e-07, "loss": 0.4161, "step": 4256 }, { "epoch": 8.652439024390244, "grad_norm": 4.3240319799437925, "learning_rate": 9.3754050659839e-07, "loss": 0.4641, "step": 4257 }, { "epoch": 8.654471544715447, "grad_norm": 5.396843019840459, "learning_rate": 9.347593379096942e-07, "loss": 0.4035, "step": 4258 }, { "epoch": 8.65650406504065, "grad_norm": 4.206959329639145, "learning_rate": 9.319820981863426e-07, "loss": 0.367, "step": 4259 }, { "epoch": 8.658536585365853, "grad_norm": 4.821017659165803, "learning_rate": 9.292087886320166e-07, "loss": 0.4901, "step": 4260 }, { "epoch": 8.660569105691057, "grad_norm": 5.105469827398396, "learning_rate": 9.264394104486963e-07, "loss": 0.4395, "step": 4261 }, { "epoch": 8.66260162601626, "grad_norm": 5.5940408654374645, "learning_rate": 9.236739648366566e-07, "loss": 0.5741, "step": 4262 }, { "epoch": 8.664634146341463, "grad_norm": 4.744024895956741, "learning_rate": 9.209124529944669e-07, "loss": 0.465, "step": 4263 }, { "epoch": 8.666666666666666, "grad_norm": 4.971473652285971, "learning_rate": 9.181548761189996e-07, "loss": 0.4608, "step": 4264 }, { "epoch": 8.66869918699187, "grad_norm": 5.229232282058707, "learning_rate": 9.154012354054121e-07, "loss": 0.4125, "step": 4265 }, { "epoch": 8.670731707317072, "grad_norm": 5.012989794509547, "learning_rate": 9.126515320471618e-07, "loss": 0.3844, "step": 4266 }, { "epoch": 8.672764227642276, "grad_norm": 5.375801927683618, "learning_rate": 9.099057672359945e-07, "loss": 0.3925, "step": 4267 }, { "epoch": 8.67479674796748, "grad_norm": 4.5449858182898115, "learning_rate": 9.071639421619527e-07, "loss": 0.416, "step": 4268 }, { "epoch": 8.676829268292684, "grad_norm": 4.705418596748461, "learning_rate": 9.044260580133734e-07, "loss": 0.5198, "step": 4269 }, { "epoch": 8.678861788617887, "grad_norm": 5.060435543355853, "learning_rate": 9.01692115976881e-07, "loss": 0.419, "step": 4270 }, { "epoch": 8.68089430894309, "grad_norm": 5.273763654373436, "learning_rate": 8.989621172373919e-07, "loss": 0.5152, "step": 4271 }, { "epoch": 8.682926829268293, "grad_norm": 5.080208384459305, "learning_rate": 8.962360629781164e-07, "loss": 0.3214, "step": 4272 }, { "epoch": 8.684959349593496, "grad_norm": 4.482035484214764, "learning_rate": 8.935139543805516e-07, "loss": 0.3973, "step": 4273 }, { "epoch": 8.6869918699187, "grad_norm": 4.757506563870643, "learning_rate": 8.907957926244892e-07, "loss": 0.4675, "step": 4274 }, { "epoch": 8.689024390243903, "grad_norm": 5.004706322779929, "learning_rate": 8.880815788880059e-07, "loss": 0.3775, "step": 4275 }, { "epoch": 8.691056910569106, "grad_norm": 5.0505939410114715, "learning_rate": 8.853713143474685e-07, "loss": 0.4962, "step": 4276 }, { "epoch": 8.69308943089431, "grad_norm": 4.687657778747023, "learning_rate": 8.826650001775327e-07, "loss": 0.3931, "step": 4277 }, { "epoch": 8.695121951219512, "grad_norm": 5.572381044385339, "learning_rate": 8.799626375511416e-07, "loss": 0.401, "step": 4278 }, { "epoch": 8.697154471544716, "grad_norm": 5.045599791920973, "learning_rate": 8.772642276395293e-07, "loss": 0.4411, "step": 4279 }, { "epoch": 8.699186991869919, "grad_norm": 5.462991739097739, "learning_rate": 8.745697716122081e-07, "loss": 0.4695, "step": 4280 }, { "epoch": 8.701219512195122, "grad_norm": 4.514460652816207, "learning_rate": 8.718792706369861e-07, "loss": 0.5669, "step": 4281 }, { "epoch": 8.703252032520325, "grad_norm": 4.174043804741593, "learning_rate": 8.691927258799504e-07, "loss": 0.3872, "step": 4282 }, { "epoch": 8.705284552845528, "grad_norm": 4.988722497428056, "learning_rate": 8.665101385054786e-07, "loss": 0.5217, "step": 4283 }, { "epoch": 8.707317073170731, "grad_norm": 5.444829868859583, "learning_rate": 8.638315096762318e-07, "loss": 0.5433, "step": 4284 }, { "epoch": 8.709349593495935, "grad_norm": 4.789732917745901, "learning_rate": 8.61156840553149e-07, "loss": 0.4578, "step": 4285 }, { "epoch": 8.711382113821138, "grad_norm": 5.067522969248511, "learning_rate": 8.58486132295463e-07, "loss": 0.5075, "step": 4286 }, { "epoch": 8.713414634146341, "grad_norm": 5.579265350120685, "learning_rate": 8.558193860606822e-07, "loss": 0.4555, "step": 4287 }, { "epoch": 8.715447154471544, "grad_norm": 5.3830900152749255, "learning_rate": 8.531566030046035e-07, "loss": 0.2872, "step": 4288 }, { "epoch": 8.717479674796747, "grad_norm": 5.04552960706346, "learning_rate": 8.504977842813056e-07, "loss": 0.4043, "step": 4289 }, { "epoch": 8.71951219512195, "grad_norm": 5.526569229726972, "learning_rate": 8.478429310431402e-07, "loss": 0.4431, "step": 4290 }, { "epoch": 8.721544715447154, "grad_norm": 4.542318485923492, "learning_rate": 8.451920444407524e-07, "loss": 0.5491, "step": 4291 }, { "epoch": 8.723577235772357, "grad_norm": 5.213827160458608, "learning_rate": 8.425451256230588e-07, "loss": 0.4485, "step": 4292 }, { "epoch": 8.725609756097562, "grad_norm": 5.303112493383718, "learning_rate": 8.399021757372672e-07, "loss": 0.3811, "step": 4293 }, { "epoch": 8.727642276422765, "grad_norm": 4.9615662552159385, "learning_rate": 8.37263195928849e-07, "loss": 0.4245, "step": 4294 }, { "epoch": 8.729674796747968, "grad_norm": 4.2838063836453735, "learning_rate": 8.346281873415719e-07, "loss": 0.4897, "step": 4295 }, { "epoch": 8.731707317073171, "grad_norm": 5.891947206574083, "learning_rate": 8.319971511174718e-07, "loss": 0.4829, "step": 4296 }, { "epoch": 8.733739837398375, "grad_norm": 4.808945836389324, "learning_rate": 8.293700883968648e-07, "loss": 0.4129, "step": 4297 }, { "epoch": 8.735772357723578, "grad_norm": 4.4045240748344865, "learning_rate": 8.267470003183498e-07, "loss": 0.4739, "step": 4298 }, { "epoch": 8.737804878048781, "grad_norm": 5.243411496686164, "learning_rate": 8.24127888018793e-07, "loss": 0.5006, "step": 4299 }, { "epoch": 8.739837398373984, "grad_norm": 4.846964226574533, "learning_rate": 8.215127526333499e-07, "loss": 0.4253, "step": 4300 }, { "epoch": 8.741869918699187, "grad_norm": 4.399571313857928, "learning_rate": 8.189015952954438e-07, "loss": 0.3959, "step": 4301 }, { "epoch": 8.74390243902439, "grad_norm": 4.522815867813755, "learning_rate": 8.162944171367737e-07, "loss": 0.4582, "step": 4302 }, { "epoch": 8.745934959349594, "grad_norm": 4.684011566960084, "learning_rate": 8.136912192873237e-07, "loss": 0.5258, "step": 4303 }, { "epoch": 8.747967479674797, "grad_norm": 5.8716308119660185, "learning_rate": 8.110920028753355e-07, "loss": 0.5991, "step": 4304 }, { "epoch": 8.75, "grad_norm": 4.995492711445619, "learning_rate": 8.084967690273437e-07, "loss": 0.3855, "step": 4305 }, { "epoch": 8.752032520325203, "grad_norm": 4.260214492917698, "learning_rate": 8.059055188681442e-07, "loss": 0.4712, "step": 4306 }, { "epoch": 8.754065040650406, "grad_norm": 5.3990549572688, "learning_rate": 8.033182535208106e-07, "loss": 0.4046, "step": 4307 }, { "epoch": 8.75609756097561, "grad_norm": 4.941563804430459, "learning_rate": 8.007349741066939e-07, "loss": 0.5024, "step": 4308 }, { "epoch": 8.758130081300813, "grad_norm": 4.6861587825814635, "learning_rate": 7.981556817454061e-07, "loss": 0.4993, "step": 4309 }, { "epoch": 8.760162601626016, "grad_norm": 5.4582496205664315, "learning_rate": 7.955803775548431e-07, "loss": 0.5242, "step": 4310 }, { "epoch": 8.762195121951219, "grad_norm": 5.053129297673929, "learning_rate": 7.930090626511667e-07, "loss": 0.4078, "step": 4311 }, { "epoch": 8.764227642276422, "grad_norm": 4.560859570143764, "learning_rate": 7.904417381488083e-07, "loss": 0.339, "step": 4312 }, { "epoch": 8.766260162601625, "grad_norm": 5.755118888791832, "learning_rate": 7.878784051604738e-07, "loss": 0.4801, "step": 4313 }, { "epoch": 8.768292682926829, "grad_norm": 5.02484152977885, "learning_rate": 7.853190647971343e-07, "loss": 0.4317, "step": 4314 }, { "epoch": 8.770325203252032, "grad_norm": 5.387453938009264, "learning_rate": 7.82763718168038e-07, "loss": 0.3722, "step": 4315 }, { "epoch": 8.772357723577235, "grad_norm": 5.751444274731012, "learning_rate": 7.802123663806938e-07, "loss": 0.5378, "step": 4316 }, { "epoch": 8.774390243902438, "grad_norm": 5.326214815919761, "learning_rate": 7.776650105408845e-07, "loss": 0.4938, "step": 4317 }, { "epoch": 8.776422764227643, "grad_norm": 4.45074442553728, "learning_rate": 7.751216517526594e-07, "loss": 0.4186, "step": 4318 }, { "epoch": 8.778455284552846, "grad_norm": 4.672141519026592, "learning_rate": 7.725822911183322e-07, "loss": 0.4598, "step": 4319 }, { "epoch": 8.78048780487805, "grad_norm": 5.074816333952639, "learning_rate": 7.700469297384927e-07, "loss": 0.5062, "step": 4320 }, { "epoch": 8.782520325203253, "grad_norm": 6.2323891281702215, "learning_rate": 7.675155687119884e-07, "loss": 0.4401, "step": 4321 }, { "epoch": 8.784552845528456, "grad_norm": 5.66424705198246, "learning_rate": 7.649882091359373e-07, "loss": 0.4055, "step": 4322 }, { "epoch": 8.786585365853659, "grad_norm": 5.364476470273918, "learning_rate": 7.624648521057209e-07, "loss": 0.4897, "step": 4323 }, { "epoch": 8.788617886178862, "grad_norm": 5.516697133845202, "learning_rate": 7.599454987149868e-07, "loss": 0.4618, "step": 4324 }, { "epoch": 8.790650406504065, "grad_norm": 5.743563496340588, "learning_rate": 7.574301500556502e-07, "loss": 0.4584, "step": 4325 }, { "epoch": 8.792682926829269, "grad_norm": 4.84625655715568, "learning_rate": 7.54918807217887e-07, "loss": 0.4224, "step": 4326 }, { "epoch": 8.794715447154472, "grad_norm": 4.473619300668532, "learning_rate": 7.524114712901376e-07, "loss": 0.4851, "step": 4327 }, { "epoch": 8.796747967479675, "grad_norm": 4.2028681149271465, "learning_rate": 7.499081433591071e-07, "loss": 0.4053, "step": 4328 }, { "epoch": 8.798780487804878, "grad_norm": 4.809439465050151, "learning_rate": 7.4740882450976e-07, "loss": 0.4757, "step": 4329 }, { "epoch": 8.800813008130081, "grad_norm": 5.424715255932885, "learning_rate": 7.449135158253306e-07, "loss": 0.4788, "step": 4330 }, { "epoch": 8.802845528455284, "grad_norm": 5.892418423148387, "learning_rate": 7.424222183873087e-07, "loss": 0.381, "step": 4331 }, { "epoch": 8.804878048780488, "grad_norm": 5.14996943469567, "learning_rate": 7.399349332754458e-07, "loss": 0.4926, "step": 4332 }, { "epoch": 8.80691056910569, "grad_norm": 4.872763516314516, "learning_rate": 7.374516615677563e-07, "loss": 0.5369, "step": 4333 }, { "epoch": 8.808943089430894, "grad_norm": 4.553003863506499, "learning_rate": 7.349724043405149e-07, "loss": 0.3483, "step": 4334 }, { "epoch": 8.810975609756097, "grad_norm": 6.437052992464861, "learning_rate": 7.324971626682587e-07, "loss": 0.4791, "step": 4335 }, { "epoch": 8.8130081300813, "grad_norm": 4.891049394502255, "learning_rate": 7.300259376237795e-07, "loss": 0.4724, "step": 4336 }, { "epoch": 8.815040650406504, "grad_norm": 5.427148340440266, "learning_rate": 7.275587302781318e-07, "loss": 0.3931, "step": 4337 }, { "epoch": 8.817073170731707, "grad_norm": 4.782152411299281, "learning_rate": 7.250955417006267e-07, "loss": 0.4189, "step": 4338 }, { "epoch": 8.81910569105691, "grad_norm": 4.5055862997985505, "learning_rate": 7.226363729588337e-07, "loss": 0.5028, "step": 4339 }, { "epoch": 8.821138211382113, "grad_norm": 4.752702384627825, "learning_rate": 7.201812251185869e-07, "loss": 0.3819, "step": 4340 }, { "epoch": 8.823170731707316, "grad_norm": 5.824144118046563, "learning_rate": 7.177300992439629e-07, "loss": 0.367, "step": 4341 }, { "epoch": 8.82520325203252, "grad_norm": 5.362573455913562, "learning_rate": 7.152829963973106e-07, "loss": 0.3626, "step": 4342 }, { "epoch": 8.827235772357724, "grad_norm": 5.50046476909845, "learning_rate": 7.128399176392275e-07, "loss": 0.4585, "step": 4343 }, { "epoch": 8.829268292682928, "grad_norm": 5.35685941982312, "learning_rate": 7.104008640285642e-07, "loss": 0.4256, "step": 4344 }, { "epoch": 8.83130081300813, "grad_norm": 5.113689921962224, "learning_rate": 7.079658366224384e-07, "loss": 0.4868, "step": 4345 }, { "epoch": 8.833333333333334, "grad_norm": 5.063050201663218, "learning_rate": 7.055348364762071e-07, "loss": 0.4063, "step": 4346 }, { "epoch": 8.835365853658537, "grad_norm": 5.7525527223941255, "learning_rate": 7.031078646434941e-07, "loss": 0.5714, "step": 4347 }, { "epoch": 8.83739837398374, "grad_norm": 5.288213449431181, "learning_rate": 7.006849221761736e-07, "loss": 0.455, "step": 4348 }, { "epoch": 8.839430894308943, "grad_norm": 6.4553691913101305, "learning_rate": 6.982660101243698e-07, "loss": 0.5562, "step": 4349 }, { "epoch": 8.841463414634147, "grad_norm": 5.513501170647107, "learning_rate": 6.958511295364678e-07, "loss": 0.4455, "step": 4350 }, { "epoch": 8.84349593495935, "grad_norm": 6.0566282456892475, "learning_rate": 6.934402814590957e-07, "loss": 0.5284, "step": 4351 }, { "epoch": 8.845528455284553, "grad_norm": 5.800582999832849, "learning_rate": 6.910334669371433e-07, "loss": 0.4757, "step": 4352 }, { "epoch": 8.847560975609756, "grad_norm": 5.082596234242842, "learning_rate": 6.886306870137449e-07, "loss": 0.4648, "step": 4353 }, { "epoch": 8.84959349593496, "grad_norm": 6.319843780070029, "learning_rate": 6.862319427302899e-07, "loss": 0.5197, "step": 4354 }, { "epoch": 8.851626016260163, "grad_norm": 4.739516804744794, "learning_rate": 6.83837235126421e-07, "loss": 0.4667, "step": 4355 }, { "epoch": 8.853658536585366, "grad_norm": 5.929080935709462, "learning_rate": 6.814465652400237e-07, "loss": 0.2821, "step": 4356 }, { "epoch": 8.855691056910569, "grad_norm": 4.25102947367709, "learning_rate": 6.790599341072401e-07, "loss": 0.4291, "step": 4357 }, { "epoch": 8.857723577235772, "grad_norm": 4.456992719533615, "learning_rate": 6.766773427624585e-07, "loss": 0.4297, "step": 4358 }, { "epoch": 8.859756097560975, "grad_norm": 5.138940156184707, "learning_rate": 6.742987922383226e-07, "loss": 0.4121, "step": 4359 }, { "epoch": 8.861788617886178, "grad_norm": 4.612554410005175, "learning_rate": 6.719242835657147e-07, "loss": 0.5649, "step": 4360 }, { "epoch": 8.863821138211382, "grad_norm": 4.870797846480153, "learning_rate": 6.695538177737704e-07, "loss": 0.5788, "step": 4361 }, { "epoch": 8.865853658536585, "grad_norm": 5.207855424309213, "learning_rate": 6.671873958898756e-07, "loss": 0.4724, "step": 4362 }, { "epoch": 8.867886178861788, "grad_norm": 5.955739780543205, "learning_rate": 6.648250189396587e-07, "loss": 0.4072, "step": 4363 }, { "epoch": 8.869918699186991, "grad_norm": 5.825583477627128, "learning_rate": 6.62466687947001e-07, "loss": 0.4471, "step": 4364 }, { "epoch": 8.871951219512194, "grad_norm": 5.533709634640845, "learning_rate": 6.601124039340235e-07, "loss": 0.361, "step": 4365 }, { "epoch": 8.8739837398374, "grad_norm": 5.054749710800501, "learning_rate": 6.577621679210955e-07, "loss": 0.4067, "step": 4366 }, { "epoch": 8.8760162601626, "grad_norm": 4.569075257302053, "learning_rate": 6.554159809268357e-07, "loss": 0.3971, "step": 4367 }, { "epoch": 8.878048780487806, "grad_norm": 5.173725876684425, "learning_rate": 6.530738439681017e-07, "loss": 0.5345, "step": 4368 }, { "epoch": 8.880081300813009, "grad_norm": 5.316439198064189, "learning_rate": 6.507357580600049e-07, "loss": 0.5043, "step": 4369 }, { "epoch": 8.882113821138212, "grad_norm": 5.279527066002336, "learning_rate": 6.484017242158857e-07, "loss": 0.4194, "step": 4370 }, { "epoch": 8.884146341463415, "grad_norm": 4.725924357015053, "learning_rate": 6.460717434473451e-07, "loss": 0.4885, "step": 4371 }, { "epoch": 8.886178861788618, "grad_norm": 4.9741956944740995, "learning_rate": 6.437458167642164e-07, "loss": 0.4061, "step": 4372 }, { "epoch": 8.888211382113822, "grad_norm": 4.947334582218455, "learning_rate": 6.414239451745807e-07, "loss": 0.47, "step": 4373 }, { "epoch": 8.890243902439025, "grad_norm": 5.394353808943312, "learning_rate": 6.3910612968476e-07, "loss": 0.3788, "step": 4374 }, { "epoch": 8.892276422764228, "grad_norm": 5.078545087717936, "learning_rate": 6.367923712993163e-07, "loss": 0.5009, "step": 4375 }, { "epoch": 8.894308943089431, "grad_norm": 4.3538382925528145, "learning_rate": 6.344826710210584e-07, "loss": 0.436, "step": 4376 }, { "epoch": 8.896341463414634, "grad_norm": 6.101453395350926, "learning_rate": 6.321770298510332e-07, "loss": 0.4396, "step": 4377 }, { "epoch": 8.898373983739837, "grad_norm": 5.125523659027087, "learning_rate": 6.298754487885272e-07, "loss": 0.4398, "step": 4378 }, { "epoch": 8.90040650406504, "grad_norm": 5.281633506800064, "learning_rate": 6.27577928831069e-07, "loss": 0.3972, "step": 4379 }, { "epoch": 8.902439024390244, "grad_norm": 5.60637287242486, "learning_rate": 6.252844709744255e-07, "loss": 0.5272, "step": 4380 }, { "epoch": 8.904471544715447, "grad_norm": 5.810229449985162, "learning_rate": 6.229950762126058e-07, "loss": 0.431, "step": 4381 }, { "epoch": 8.90650406504065, "grad_norm": 4.330924353403116, "learning_rate": 6.207097455378563e-07, "loss": 0.4689, "step": 4382 }, { "epoch": 8.908536585365853, "grad_norm": 5.42343737971838, "learning_rate": 6.184284799406615e-07, "loss": 0.48, "step": 4383 }, { "epoch": 8.910569105691057, "grad_norm": 4.603522958251836, "learning_rate": 6.161512804097436e-07, "loss": 0.3326, "step": 4384 }, { "epoch": 8.91260162601626, "grad_norm": 5.0554771203241815, "learning_rate": 6.138781479320632e-07, "loss": 0.3474, "step": 4385 }, { "epoch": 8.914634146341463, "grad_norm": 5.599858922678689, "learning_rate": 6.116090834928213e-07, "loss": 0.4523, "step": 4386 }, { "epoch": 8.916666666666666, "grad_norm": 4.30695733951631, "learning_rate": 6.093440880754508e-07, "loss": 0.4447, "step": 4387 }, { "epoch": 8.91869918699187, "grad_norm": 5.80018752260982, "learning_rate": 6.070831626616236e-07, "loss": 0.446, "step": 4388 }, { "epoch": 8.920731707317072, "grad_norm": 5.538079741816438, "learning_rate": 6.048263082312478e-07, "loss": 0.4882, "step": 4389 }, { "epoch": 8.922764227642276, "grad_norm": 4.606880746337986, "learning_rate": 6.025735257624654e-07, "loss": 0.5007, "step": 4390 }, { "epoch": 8.92479674796748, "grad_norm": 4.908952576407026, "learning_rate": 6.003248162316555e-07, "loss": 0.4071, "step": 4391 }, { "epoch": 8.926829268292684, "grad_norm": 4.627905192595351, "learning_rate": 5.980801806134318e-07, "loss": 0.4947, "step": 4392 }, { "epoch": 8.928861788617887, "grad_norm": 4.983180828033927, "learning_rate": 5.958396198806405e-07, "loss": 0.4525, "step": 4393 }, { "epoch": 8.93089430894309, "grad_norm": 4.516935843139752, "learning_rate": 5.936031350043625e-07, "loss": 0.3835, "step": 4394 }, { "epoch": 8.932926829268293, "grad_norm": 5.488692116223383, "learning_rate": 5.913707269539126e-07, "loss": 0.4304, "step": 4395 }, { "epoch": 8.934959349593496, "grad_norm": 4.785684735956892, "learning_rate": 5.891423966968413e-07, "loss": 0.4248, "step": 4396 }, { "epoch": 8.9369918699187, "grad_norm": 5.127313223244118, "learning_rate": 5.869181451989259e-07, "loss": 0.4436, "step": 4397 }, { "epoch": 8.939024390243903, "grad_norm": 5.29428781816751, "learning_rate": 5.846979734241809e-07, "loss": 0.5161, "step": 4398 }, { "epoch": 8.941056910569106, "grad_norm": 5.241561527851893, "learning_rate": 5.8248188233485e-07, "loss": 0.4273, "step": 4399 }, { "epoch": 8.94308943089431, "grad_norm": 4.792578283572765, "learning_rate": 5.80269872891408e-07, "loss": 0.4946, "step": 4400 }, { "epoch": 8.945121951219512, "grad_norm": 4.486934152593047, "learning_rate": 5.780619460525661e-07, "loss": 0.3819, "step": 4401 }, { "epoch": 8.947154471544716, "grad_norm": 5.188700958002736, "learning_rate": 5.758581027752552e-07, "loss": 0.4377, "step": 4402 }, { "epoch": 8.949186991869919, "grad_norm": 5.196018421589996, "learning_rate": 5.736583440146482e-07, "loss": 0.4636, "step": 4403 }, { "epoch": 8.951219512195122, "grad_norm": 5.397343851590606, "learning_rate": 5.714626707241411e-07, "loss": 0.4036, "step": 4404 }, { "epoch": 8.953252032520325, "grad_norm": 5.040187755854167, "learning_rate": 5.692710838553572e-07, "loss": 0.4736, "step": 4405 }, { "epoch": 8.955284552845528, "grad_norm": 4.345093344969024, "learning_rate": 5.67083584358159e-07, "loss": 0.4475, "step": 4406 }, { "epoch": 8.957317073170731, "grad_norm": 5.226774103255528, "learning_rate": 5.649001731806226e-07, "loss": 0.5535, "step": 4407 }, { "epoch": 8.959349593495935, "grad_norm": 5.572737447196633, "learning_rate": 5.627208512690641e-07, "loss": 0.5172, "step": 4408 }, { "epoch": 8.961382113821138, "grad_norm": 5.0815458513978475, "learning_rate": 5.605456195680248e-07, "loss": 0.4004, "step": 4409 }, { "epoch": 8.963414634146341, "grad_norm": 4.696172596648737, "learning_rate": 5.583744790202672e-07, "loss": 0.4386, "step": 4410 }, { "epoch": 8.965447154471544, "grad_norm": 5.4955060498071715, "learning_rate": 5.562074305667908e-07, "loss": 0.42, "step": 4411 }, { "epoch": 8.967479674796747, "grad_norm": 4.781897261674849, "learning_rate": 5.5404447514681e-07, "loss": 0.4244, "step": 4412 }, { "epoch": 8.96951219512195, "grad_norm": 5.448393182396848, "learning_rate": 5.518856136977758e-07, "loss": 0.4363, "step": 4413 }, { "epoch": 8.971544715447154, "grad_norm": 5.784878021478179, "learning_rate": 5.497308471553586e-07, "loss": 0.5315, "step": 4414 }, { "epoch": 8.973577235772357, "grad_norm": 4.884931049956898, "learning_rate": 5.475801764534538e-07, "loss": 0.4487, "step": 4415 }, { "epoch": 8.975609756097562, "grad_norm": 5.375388600487724, "learning_rate": 5.45433602524188e-07, "loss": 0.4807, "step": 4416 }, { "epoch": 8.977642276422765, "grad_norm": 5.886292838089242, "learning_rate": 5.432911262979024e-07, "loss": 0.4378, "step": 4417 }, { "epoch": 8.979674796747968, "grad_norm": 5.261132672452963, "learning_rate": 5.411527487031709e-07, "loss": 0.3806, "step": 4418 }, { "epoch": 8.981707317073171, "grad_norm": 5.240744206569963, "learning_rate": 5.390184706667867e-07, "loss": 0.3733, "step": 4419 }, { "epoch": 8.983739837398375, "grad_norm": 4.164394963912711, "learning_rate": 5.368882931137675e-07, "loss": 0.4541, "step": 4420 }, { "epoch": 8.985772357723578, "grad_norm": 5.58088009389505, "learning_rate": 5.347622169673516e-07, "loss": 0.5105, "step": 4421 }, { "epoch": 8.987804878048781, "grad_norm": 4.540040333285712, "learning_rate": 5.32640243149003e-07, "loss": 0.4419, "step": 4422 }, { "epoch": 8.989837398373984, "grad_norm": 4.857473803886191, "learning_rate": 5.305223725784059e-07, "loss": 0.3958, "step": 4423 }, { "epoch": 8.991869918699187, "grad_norm": 4.700046946030152, "learning_rate": 5.284086061734672e-07, "loss": 0.432, "step": 4424 }, { "epoch": 8.99390243902439, "grad_norm": 5.2801315719397675, "learning_rate": 5.262989448503142e-07, "loss": 0.3978, "step": 4425 }, { "epoch": 8.995934959349594, "grad_norm": 4.863535192900564, "learning_rate": 5.241933895232931e-07, "loss": 0.4627, "step": 4426 }, { "epoch": 8.997967479674797, "grad_norm": 4.7696766785215825, "learning_rate": 5.220919411049741e-07, "loss": 0.4771, "step": 4427 }, { "epoch": 9.0, "grad_norm": 6.346615344317735, "learning_rate": 5.199946005061462e-07, "loss": 0.5389, "step": 4428 }, { "epoch": 9.0, "eval_loss": 1.053712248802185, "eval_runtime": 132.122, "eval_samples_per_second": 7.448, "eval_steps_per_second": 0.931, "step": 4428 }, { "epoch": 9.002032520325203, "grad_norm": 4.9739355056094, "learning_rate": 5.17901368635817e-07, "loss": 0.5309, "step": 4429 }, { "epoch": 9.004065040650406, "grad_norm": 4.8167306005144646, "learning_rate": 5.158122464012138e-07, "loss": 0.5466, "step": 4430 }, { "epoch": 9.00609756097561, "grad_norm": 4.587792022800118, "learning_rate": 5.137272347077838e-07, "loss": 0.4702, "step": 4431 }, { "epoch": 9.008130081300813, "grad_norm": 4.198963280857847, "learning_rate": 5.116463344591893e-07, "loss": 0.4196, "step": 4432 }, { "epoch": 9.010162601626016, "grad_norm": 4.923867111225552, "learning_rate": 5.095695465573158e-07, "loss": 0.4404, "step": 4433 }, { "epoch": 9.012195121951219, "grad_norm": 4.765354993637875, "learning_rate": 5.074968719022644e-07, "loss": 0.4637, "step": 4434 }, { "epoch": 9.014227642276422, "grad_norm": 4.728779322896657, "learning_rate": 5.054283113923508e-07, "loss": 0.3795, "step": 4435 }, { "epoch": 9.016260162601625, "grad_norm": 5.678579951931546, "learning_rate": 5.033638659241102e-07, "loss": 0.4142, "step": 4436 }, { "epoch": 9.018292682926829, "grad_norm": 4.588773968544693, "learning_rate": 5.013035363922925e-07, "loss": 0.3464, "step": 4437 }, { "epoch": 9.020325203252032, "grad_norm": 4.788468654905465, "learning_rate": 4.992473236898676e-07, "loss": 0.3649, "step": 4438 }, { "epoch": 9.022357723577235, "grad_norm": 4.287673114420762, "learning_rate": 4.971952287080173e-07, "loss": 0.5012, "step": 4439 }, { "epoch": 9.024390243902438, "grad_norm": 5.418775022411461, "learning_rate": 4.951472523361401e-07, "loss": 0.4093, "step": 4440 }, { "epoch": 9.026422764227643, "grad_norm": 4.592229085293213, "learning_rate": 4.931033954618492e-07, "loss": 0.4835, "step": 4441 }, { "epoch": 9.028455284552846, "grad_norm": 5.2970179767102366, "learning_rate": 4.910636589709716e-07, "loss": 0.4574, "step": 4442 }, { "epoch": 9.03048780487805, "grad_norm": 5.164531612712752, "learning_rate": 4.890280437475503e-07, "loss": 0.473, "step": 4443 }, { "epoch": 9.032520325203253, "grad_norm": 4.28507916280863, "learning_rate": 4.869965506738416e-07, "loss": 0.4189, "step": 4444 }, { "epoch": 9.034552845528456, "grad_norm": 4.992798263975871, "learning_rate": 4.849691806303147e-07, "loss": 0.4718, "step": 4445 }, { "epoch": 9.036585365853659, "grad_norm": 5.536311563828069, "learning_rate": 4.829459344956511e-07, "loss": 0.4053, "step": 4446 }, { "epoch": 9.038617886178862, "grad_norm": 4.36924146342536, "learning_rate": 4.809268131467448e-07, "loss": 0.3377, "step": 4447 }, { "epoch": 9.040650406504065, "grad_norm": 5.508799256514077, "learning_rate": 4.789118174587071e-07, "loss": 0.4484, "step": 4448 }, { "epoch": 9.042682926829269, "grad_norm": 4.633493378971268, "learning_rate": 4.769009483048515e-07, "loss": 0.5297, "step": 4449 }, { "epoch": 9.044715447154472, "grad_norm": 4.761786420757763, "learning_rate": 4.748942065567119e-07, "loss": 0.4803, "step": 4450 }, { "epoch": 9.046747967479675, "grad_norm": 5.452689087014923, "learning_rate": 4.7289159308402943e-07, "loss": 0.5047, "step": 4451 }, { "epoch": 9.048780487804878, "grad_norm": 4.669574611450418, "learning_rate": 4.7089310875475856e-07, "loss": 0.3543, "step": 4452 }, { "epoch": 9.050813008130081, "grad_norm": 5.551974611975841, "learning_rate": 4.6889875443506207e-07, "loss": 0.3749, "step": 4453 }, { "epoch": 9.052845528455284, "grad_norm": 5.056471078778258, "learning_rate": 4.6690853098930844e-07, "loss": 0.5253, "step": 4454 }, { "epoch": 9.054878048780488, "grad_norm": 4.050508472276154, "learning_rate": 4.649224392800844e-07, "loss": 0.3617, "step": 4455 }, { "epoch": 9.05691056910569, "grad_norm": 4.522340675643052, "learning_rate": 4.6294048016817917e-07, "loss": 0.3858, "step": 4456 }, { "epoch": 9.058943089430894, "grad_norm": 5.407643277916597, "learning_rate": 4.6096265451259783e-07, "loss": 0.3172, "step": 4457 }, { "epoch": 9.060975609756097, "grad_norm": 4.612735225378065, "learning_rate": 4.5898896317054686e-07, "loss": 0.4968, "step": 4458 }, { "epoch": 9.0630081300813, "grad_norm": 6.004703357457538, "learning_rate": 4.570194069974421e-07, "loss": 0.4607, "step": 4459 }, { "epoch": 9.065040650406504, "grad_norm": 5.082747992868454, "learning_rate": 4.550539868469106e-07, "loss": 0.509, "step": 4460 }, { "epoch": 9.067073170731707, "grad_norm": 5.142589617833051, "learning_rate": 4.530927035707844e-07, "loss": 0.5075, "step": 4461 }, { "epoch": 9.06910569105691, "grad_norm": 5.318391709225177, "learning_rate": 4.511355580191057e-07, "loss": 0.4931, "step": 4462 }, { "epoch": 9.071138211382113, "grad_norm": 5.303798651436456, "learning_rate": 4.491825510401171e-07, "loss": 0.4357, "step": 4463 }, { "epoch": 9.073170731707316, "grad_norm": 4.765884479637659, "learning_rate": 4.4723368348027375e-07, "loss": 0.4636, "step": 4464 }, { "epoch": 9.075203252032521, "grad_norm": 4.349774261823624, "learning_rate": 4.452889561842333e-07, "loss": 0.3978, "step": 4465 }, { "epoch": 9.077235772357724, "grad_norm": 4.85070539193404, "learning_rate": 4.433483699948582e-07, "loss": 0.3842, "step": 4466 }, { "epoch": 9.079268292682928, "grad_norm": 4.509229583034046, "learning_rate": 4.4141192575322233e-07, "loss": 0.4564, "step": 4467 }, { "epoch": 9.08130081300813, "grad_norm": 5.455826424748683, "learning_rate": 4.394796242985933e-07, "loss": 0.4575, "step": 4468 }, { "epoch": 9.083333333333334, "grad_norm": 5.898574192400908, "learning_rate": 4.3755146646845546e-07, "loss": 0.3811, "step": 4469 }, { "epoch": 9.085365853658537, "grad_norm": 4.543732134620541, "learning_rate": 4.356274530984905e-07, "loss": 0.3862, "step": 4470 }, { "epoch": 9.08739837398374, "grad_norm": 5.244479354080338, "learning_rate": 4.337075850225814e-07, "loss": 0.4484, "step": 4471 }, { "epoch": 9.089430894308943, "grad_norm": 4.285386818582205, "learning_rate": 4.317918630728235e-07, "loss": 0.4262, "step": 4472 }, { "epoch": 9.091463414634147, "grad_norm": 5.0334738899850135, "learning_rate": 4.2988028807950386e-07, "loss": 0.4123, "step": 4473 }, { "epoch": 9.09349593495935, "grad_norm": 4.74525016408013, "learning_rate": 4.2797286087112313e-07, "loss": 0.3843, "step": 4474 }, { "epoch": 9.095528455284553, "grad_norm": 5.395753998228139, "learning_rate": 4.2606958227437655e-07, "loss": 0.386, "step": 4475 }, { "epoch": 9.097560975609756, "grad_norm": 5.395395079180936, "learning_rate": 4.241704531141633e-07, "loss": 0.443, "step": 4476 }, { "epoch": 9.09959349593496, "grad_norm": 4.581758012339319, "learning_rate": 4.222754742135893e-07, "loss": 0.4158, "step": 4477 }, { "epoch": 9.101626016260163, "grad_norm": 5.543299145501715, "learning_rate": 4.203846463939498e-07, "loss": 0.4884, "step": 4478 }, { "epoch": 9.103658536585366, "grad_norm": 4.52904423659004, "learning_rate": 4.184979704747549e-07, "loss": 0.3578, "step": 4479 }, { "epoch": 9.105691056910569, "grad_norm": 4.232977268740833, "learning_rate": 4.166154472737061e-07, "loss": 0.3976, "step": 4480 }, { "epoch": 9.107723577235772, "grad_norm": 5.227096216144695, "learning_rate": 4.147370776067083e-07, "loss": 0.4119, "step": 4481 }, { "epoch": 9.109756097560975, "grad_norm": 4.498015103939749, "learning_rate": 4.12862862287865e-07, "loss": 0.3784, "step": 4482 }, { "epoch": 9.111788617886178, "grad_norm": 4.401029951151349, "learning_rate": 4.1099280212947846e-07, "loss": 0.4354, "step": 4483 }, { "epoch": 9.113821138211382, "grad_norm": 4.356192399239374, "learning_rate": 4.091268979420537e-07, "loss": 0.3669, "step": 4484 }, { "epoch": 9.115853658536585, "grad_norm": 5.320108513673478, "learning_rate": 4.072651505342917e-07, "loss": 0.4546, "step": 4485 }, { "epoch": 9.117886178861788, "grad_norm": 4.95416184999179, "learning_rate": 4.054075607130914e-07, "loss": 0.4476, "step": 4486 }, { "epoch": 9.119918699186991, "grad_norm": 5.169717752762151, "learning_rate": 4.03554129283551e-07, "loss": 0.5741, "step": 4487 }, { "epoch": 9.121951219512194, "grad_norm": 4.968436798652821, "learning_rate": 4.0170485704896453e-07, "loss": 0.4238, "step": 4488 }, { "epoch": 9.123983739837398, "grad_norm": 4.66757296626012, "learning_rate": 3.998597448108277e-07, "loss": 0.5159, "step": 4489 }, { "epoch": 9.126016260162602, "grad_norm": 5.775659294197042, "learning_rate": 3.980187933688306e-07, "loss": 0.3083, "step": 4490 }, { "epoch": 9.128048780487806, "grad_norm": 5.0309672017372415, "learning_rate": 3.9618200352085745e-07, "loss": 0.3802, "step": 4491 }, { "epoch": 9.130081300813009, "grad_norm": 5.0176395362125366, "learning_rate": 3.943493760629924e-07, "loss": 0.3894, "step": 4492 }, { "epoch": 9.132113821138212, "grad_norm": 5.224518557754616, "learning_rate": 3.925209117895135e-07, "loss": 0.3973, "step": 4493 }, { "epoch": 9.134146341463415, "grad_norm": 4.54833930033871, "learning_rate": 3.9069661149289674e-07, "loss": 0.413, "step": 4494 }, { "epoch": 9.136178861788618, "grad_norm": 5.324194141595845, "learning_rate": 3.8887647596381195e-07, "loss": 0.3963, "step": 4495 }, { "epoch": 9.138211382113822, "grad_norm": 5.4246434566124995, "learning_rate": 3.8706050599112363e-07, "loss": 0.4139, "step": 4496 }, { "epoch": 9.140243902439025, "grad_norm": 4.74109910612423, "learning_rate": 3.8524870236189005e-07, "loss": 0.3827, "step": 4497 }, { "epoch": 9.142276422764228, "grad_norm": 4.789463064058026, "learning_rate": 3.834410658613652e-07, "loss": 0.389, "step": 4498 }, { "epoch": 9.144308943089431, "grad_norm": 4.875665735496934, "learning_rate": 3.8163759727299796e-07, "loss": 0.3689, "step": 4499 }, { "epoch": 9.146341463414634, "grad_norm": 4.766931071690428, "learning_rate": 3.798382973784298e-07, "loss": 0.339, "step": 4500 }, { "epoch": 9.148373983739837, "grad_norm": 5.343474772140771, "learning_rate": 3.780431669574935e-07, "loss": 0.3548, "step": 4501 }, { "epoch": 9.15040650406504, "grad_norm": 4.7651701567616636, "learning_rate": 3.7625220678821795e-07, "loss": 0.3195, "step": 4502 }, { "epoch": 9.152439024390244, "grad_norm": 5.4372591042896445, "learning_rate": 3.74465417646821e-07, "loss": 0.3818, "step": 4503 }, { "epoch": 9.154471544715447, "grad_norm": 5.288530997420765, "learning_rate": 3.7268280030771655e-07, "loss": 0.3952, "step": 4504 }, { "epoch": 9.15650406504065, "grad_norm": 5.259729685550877, "learning_rate": 3.709043555435088e-07, "loss": 0.413, "step": 4505 }, { "epoch": 9.158536585365853, "grad_norm": 5.30909554262666, "learning_rate": 3.6913008412499227e-07, "loss": 0.3398, "step": 4506 }, { "epoch": 9.160569105691057, "grad_norm": 4.548190856711865, "learning_rate": 3.6735998682115416e-07, "loss": 0.3397, "step": 4507 }, { "epoch": 9.16260162601626, "grad_norm": 4.338874921401737, "learning_rate": 3.655940643991718e-07, "loss": 0.4062, "step": 4508 }, { "epoch": 9.164634146341463, "grad_norm": 5.830553117179361, "learning_rate": 3.638323176244152e-07, "loss": 0.4495, "step": 4509 }, { "epoch": 9.166666666666666, "grad_norm": 4.744653213814627, "learning_rate": 3.620747472604413e-07, "loss": 0.3356, "step": 4510 }, { "epoch": 9.16869918699187, "grad_norm": 4.960472908103585, "learning_rate": 3.603213540689998e-07, "loss": 0.3612, "step": 4511 }, { "epoch": 9.170731707317072, "grad_norm": 5.45593980720699, "learning_rate": 3.585721388100283e-07, "loss": 0.4288, "step": 4512 }, { "epoch": 9.172764227642276, "grad_norm": 5.84952813844074, "learning_rate": 3.5682710224165143e-07, "loss": 0.4593, "step": 4513 }, { "epoch": 9.17479674796748, "grad_norm": 5.211499005598525, "learning_rate": 3.550862451201909e-07, "loss": 0.3827, "step": 4514 }, { "epoch": 9.176829268292684, "grad_norm": 5.866083123419872, "learning_rate": 3.533495682001453e-07, "loss": 0.4348, "step": 4515 }, { "epoch": 9.178861788617887, "grad_norm": 5.337438043193671, "learning_rate": 3.516170722342127e-07, "loss": 0.4431, "step": 4516 }, { "epoch": 9.18089430894309, "grad_norm": 5.41709430896333, "learning_rate": 3.4988875797327125e-07, "loss": 0.4183, "step": 4517 }, { "epoch": 9.182926829268293, "grad_norm": 5.151713066206893, "learning_rate": 3.4816462616638847e-07, "loss": 0.5494, "step": 4518 }, { "epoch": 9.184959349593496, "grad_norm": 5.174763762215458, "learning_rate": 3.4644467756082543e-07, "loss": 0.4385, "step": 4519 }, { "epoch": 9.1869918699187, "grad_norm": 5.262723745987425, "learning_rate": 3.4472891290201927e-07, "loss": 0.433, "step": 4520 }, { "epoch": 9.189024390243903, "grad_norm": 5.178116790951913, "learning_rate": 3.4301733293360175e-07, "loss": 0.4969, "step": 4521 }, { "epoch": 9.191056910569106, "grad_norm": 5.5920793687815245, "learning_rate": 3.4130993839738944e-07, "loss": 0.3751, "step": 4522 }, { "epoch": 9.19308943089431, "grad_norm": 4.691094536612961, "learning_rate": 3.396067300333816e-07, "loss": 0.3828, "step": 4523 }, { "epoch": 9.195121951219512, "grad_norm": 4.466942522696814, "learning_rate": 3.3790770857976995e-07, "loss": 0.4728, "step": 4524 }, { "epoch": 9.197154471544716, "grad_norm": 5.170396850238235, "learning_rate": 3.362128747729221e-07, "loss": 0.3321, "step": 4525 }, { "epoch": 9.199186991869919, "grad_norm": 5.742781117579047, "learning_rate": 3.3452222934739933e-07, "loss": 0.3748, "step": 4526 }, { "epoch": 9.201219512195122, "grad_norm": 4.788712625587141, "learning_rate": 3.328357730359422e-07, "loss": 0.4976, "step": 4527 }, { "epoch": 9.203252032520325, "grad_norm": 5.678708638076202, "learning_rate": 3.3115350656948043e-07, "loss": 0.4878, "step": 4528 }, { "epoch": 9.205284552845528, "grad_norm": 5.535324694315797, "learning_rate": 3.2947543067712195e-07, "loss": 0.4087, "step": 4529 }, { "epoch": 9.207317073170731, "grad_norm": 4.588163537670723, "learning_rate": 3.278015460861628e-07, "loss": 0.5674, "step": 4530 }, { "epoch": 9.209349593495935, "grad_norm": 5.028188743861141, "learning_rate": 3.2613185352208143e-07, "loss": 0.5677, "step": 4531 }, { "epoch": 9.211382113821138, "grad_norm": 5.828781031277548, "learning_rate": 3.2446635370853686e-07, "loss": 0.4815, "step": 4532 }, { "epoch": 9.213414634146341, "grad_norm": 4.897057575497604, "learning_rate": 3.2280504736737936e-07, "loss": 0.4136, "step": 4533 }, { "epoch": 9.215447154471544, "grad_norm": 5.834167303503153, "learning_rate": 3.2114793521862864e-07, "loss": 0.3507, "step": 4534 }, { "epoch": 9.217479674796747, "grad_norm": 4.981959682864659, "learning_rate": 3.1949501798049675e-07, "loss": 0.356, "step": 4535 }, { "epoch": 9.21951219512195, "grad_norm": 4.963467121346392, "learning_rate": 3.1784629636937404e-07, "loss": 0.3451, "step": 4536 }, { "epoch": 9.221544715447154, "grad_norm": 4.846517948474432, "learning_rate": 3.162017710998322e-07, "loss": 0.7385, "step": 4537 }, { "epoch": 9.223577235772357, "grad_norm": 5.303573805401992, "learning_rate": 3.1456144288462773e-07, "loss": 0.4533, "step": 4538 }, { "epoch": 9.225609756097562, "grad_norm": 5.230442952623761, "learning_rate": 3.1292531243469204e-07, "loss": 0.3954, "step": 4539 }, { "epoch": 9.227642276422765, "grad_norm": 5.095393892090179, "learning_rate": 3.1129338045914004e-07, "loss": 0.467, "step": 4540 }, { "epoch": 9.229674796747968, "grad_norm": 5.076572673156048, "learning_rate": 3.096656476652693e-07, "loss": 0.4518, "step": 4541 }, { "epoch": 9.231707317073171, "grad_norm": 4.558537387706208, "learning_rate": 3.0804211475855436e-07, "loss": 0.483, "step": 4542 }, { "epoch": 9.233739837398375, "grad_norm": 6.001935084716186, "learning_rate": 3.064227824426513e-07, "loss": 0.4142, "step": 4543 }, { "epoch": 9.235772357723578, "grad_norm": 4.341942916138309, "learning_rate": 3.0480765141939316e-07, "loss": 0.4356, "step": 4544 }, { "epoch": 9.237804878048781, "grad_norm": 4.771368752038076, "learning_rate": 3.031967223887955e-07, "loss": 0.4081, "step": 4545 }, { "epoch": 9.239837398373984, "grad_norm": 4.589090021995307, "learning_rate": 3.015899960490498e-07, "loss": 0.4277, "step": 4546 }, { "epoch": 9.241869918699187, "grad_norm": 5.417193126647664, "learning_rate": 2.999874730965291e-07, "loss": 0.4788, "step": 4547 }, { "epoch": 9.24390243902439, "grad_norm": 4.756047841944045, "learning_rate": 2.9838915422578e-07, "loss": 0.495, "step": 4548 }, { "epoch": 9.245934959349594, "grad_norm": 4.721926665903183, "learning_rate": 2.967950401295305e-07, "loss": 0.4922, "step": 4549 }, { "epoch": 9.247967479674797, "grad_norm": 4.650961107642811, "learning_rate": 2.95205131498687e-07, "loss": 0.3485, "step": 4550 }, { "epoch": 9.25, "grad_norm": 4.6864549982937955, "learning_rate": 2.9361942902233155e-07, "loss": 0.5005, "step": 4551 }, { "epoch": 9.252032520325203, "grad_norm": 5.275340363241378, "learning_rate": 2.920379333877221e-07, "loss": 0.4513, "step": 4552 }, { "epoch": 9.254065040650406, "grad_norm": 5.182804055396432, "learning_rate": 2.9046064528029493e-07, "loss": 0.3585, "step": 4553 }, { "epoch": 9.25609756097561, "grad_norm": 5.215771437540787, "learning_rate": 2.8888756538366203e-07, "loss": 0.4447, "step": 4554 }, { "epoch": 9.258130081300813, "grad_norm": 5.4169331480112834, "learning_rate": 2.8731869437961246e-07, "loss": 0.4168, "step": 4555 }, { "epoch": 9.260162601626016, "grad_norm": 5.435935517246761, "learning_rate": 2.8575403294811123e-07, "loss": 0.4149, "step": 4556 }, { "epoch": 9.262195121951219, "grad_norm": 5.269575935791066, "learning_rate": 2.8419358176729694e-07, "loss": 0.4881, "step": 4557 }, { "epoch": 9.264227642276422, "grad_norm": 5.344712967549321, "learning_rate": 2.8263734151348533e-07, "loss": 0.4657, "step": 4558 }, { "epoch": 9.266260162601625, "grad_norm": 4.8459882072561395, "learning_rate": 2.810853128611646e-07, "loss": 0.482, "step": 4559 }, { "epoch": 9.268292682926829, "grad_norm": 4.37543836111265, "learning_rate": 2.795374964830022e-07, "loss": 0.4429, "step": 4560 }, { "epoch": 9.270325203252032, "grad_norm": 4.989512182164597, "learning_rate": 2.779938930498349e-07, "loss": 0.4077, "step": 4561 }, { "epoch": 9.272357723577235, "grad_norm": 5.464577557066134, "learning_rate": 2.7645450323067756e-07, "loss": 0.4122, "step": 4562 }, { "epoch": 9.274390243902438, "grad_norm": 4.977321255156109, "learning_rate": 2.749193276927153e-07, "loss": 0.4313, "step": 4563 }, { "epoch": 9.276422764227643, "grad_norm": 4.446948624198841, "learning_rate": 2.733883671013082e-07, "loss": 0.4183, "step": 4564 }, { "epoch": 9.278455284552846, "grad_norm": 5.712701956718451, "learning_rate": 2.718616221199921e-07, "loss": 0.5348, "step": 4565 }, { "epoch": 9.28048780487805, "grad_norm": 4.985747807913763, "learning_rate": 2.7033909341047213e-07, "loss": 0.4259, "step": 4566 }, { "epoch": 9.282520325203253, "grad_norm": 4.311896632493715, "learning_rate": 2.6882078163262717e-07, "loss": 0.4901, "step": 4567 }, { "epoch": 9.284552845528456, "grad_norm": 5.568037749981777, "learning_rate": 2.673066874445096e-07, "loss": 0.5156, "step": 4568 }, { "epoch": 9.286585365853659, "grad_norm": 5.251174224719762, "learning_rate": 2.657968115023402e-07, "loss": 0.3911, "step": 4569 }, { "epoch": 9.288617886178862, "grad_norm": 4.652450243509443, "learning_rate": 2.6429115446051757e-07, "loss": 0.4573, "step": 4570 }, { "epoch": 9.290650406504065, "grad_norm": 5.722801931270064, "learning_rate": 2.6278971697160647e-07, "loss": 0.4026, "step": 4571 }, { "epoch": 9.292682926829269, "grad_norm": 5.890696912125531, "learning_rate": 2.612924996863453e-07, "loss": 0.3686, "step": 4572 }, { "epoch": 9.294715447154472, "grad_norm": 5.660495136494396, "learning_rate": 2.597995032536438e-07, "loss": 0.3918, "step": 4573 }, { "epoch": 9.296747967479675, "grad_norm": 4.703178992942444, "learning_rate": 2.583107283205788e-07, "loss": 0.5142, "step": 4574 }, { "epoch": 9.298780487804878, "grad_norm": 4.933513012233035, "learning_rate": 2.568261755324042e-07, "loss": 0.3082, "step": 4575 }, { "epoch": 9.300813008130081, "grad_norm": 4.772258352487176, "learning_rate": 2.5534584553253526e-07, "loss": 0.5157, "step": 4576 }, { "epoch": 9.302845528455284, "grad_norm": 4.753357409059537, "learning_rate": 2.538697389625644e-07, "loss": 0.4, "step": 4577 }, { "epoch": 9.304878048780488, "grad_norm": 5.593884371276785, "learning_rate": 2.52397856462252e-07, "loss": 0.4859, "step": 4578 }, { "epoch": 9.30691056910569, "grad_norm": 4.2343266002957485, "learning_rate": 2.509301986695234e-07, "loss": 0.4362, "step": 4579 }, { "epoch": 9.308943089430894, "grad_norm": 5.002783974924321, "learning_rate": 2.494667662204797e-07, "loss": 0.5055, "step": 4580 }, { "epoch": 9.310975609756097, "grad_norm": 5.284561337895392, "learning_rate": 2.480075597493825e-07, "loss": 0.3876, "step": 4581 }, { "epoch": 9.3130081300813, "grad_norm": 5.326290658937444, "learning_rate": 2.465525798886703e-07, "loss": 0.5011, "step": 4582 }, { "epoch": 9.315040650406504, "grad_norm": 4.76696996548753, "learning_rate": 2.451018272689454e-07, "loss": 0.3818, "step": 4583 }, { "epoch": 9.317073170731707, "grad_norm": 4.485293478111565, "learning_rate": 2.436553025189758e-07, "loss": 0.4514, "step": 4584 }, { "epoch": 9.31910569105691, "grad_norm": 5.335070649662152, "learning_rate": 2.4221300626570354e-07, "loss": 0.5542, "step": 4585 }, { "epoch": 9.321138211382113, "grad_norm": 4.581087263673888, "learning_rate": 2.4077493913422955e-07, "loss": 0.3821, "step": 4586 }, { "epoch": 9.323170731707316, "grad_norm": 4.693479675216827, "learning_rate": 2.393411017478309e-07, "loss": 0.4044, "step": 4587 }, { "epoch": 9.32520325203252, "grad_norm": 4.083479891509775, "learning_rate": 2.3791149472794373e-07, "loss": 0.4723, "step": 4588 }, { "epoch": 9.327235772357724, "grad_norm": 5.757030490310403, "learning_rate": 2.3648611869417582e-07, "loss": 0.3529, "step": 4589 }, { "epoch": 9.329268292682928, "grad_norm": 5.528939114031861, "learning_rate": 2.3506497426429853e-07, "loss": 0.4972, "step": 4590 }, { "epoch": 9.33130081300813, "grad_norm": 5.217345272394032, "learning_rate": 2.3364806205424917e-07, "loss": 0.3884, "step": 4591 }, { "epoch": 9.333333333333334, "grad_norm": 4.681047037124985, "learning_rate": 2.3223538267813317e-07, "loss": 0.4301, "step": 4592 }, { "epoch": 9.335365853658537, "grad_norm": 4.650884037111433, "learning_rate": 2.3082693674821855e-07, "loss": 0.432, "step": 4593 }, { "epoch": 9.33739837398374, "grad_norm": 5.725739501572547, "learning_rate": 2.2942272487494145e-07, "loss": 0.4785, "step": 4594 }, { "epoch": 9.339430894308943, "grad_norm": 4.574143909555959, "learning_rate": 2.2802274766689835e-07, "loss": 0.2712, "step": 4595 }, { "epoch": 9.341463414634147, "grad_norm": 5.080099230212669, "learning_rate": 2.2662700573085505e-07, "loss": 0.437, "step": 4596 }, { "epoch": 9.34349593495935, "grad_norm": 5.125086347489345, "learning_rate": 2.2523549967174096e-07, "loss": 0.3249, "step": 4597 }, { "epoch": 9.345528455284553, "grad_norm": 5.580331289177464, "learning_rate": 2.2384823009264811e-07, "loss": 0.4573, "step": 4598 }, { "epoch": 9.347560975609756, "grad_norm": 6.125450918577479, "learning_rate": 2.2246519759483108e-07, "loss": 0.5114, "step": 4599 }, { "epoch": 9.34959349593496, "grad_norm": 4.496563901983162, "learning_rate": 2.2108640277771153e-07, "loss": 0.5194, "step": 4600 }, { "epoch": 9.351626016260163, "grad_norm": 5.55875414847196, "learning_rate": 2.1971184623887253e-07, "loss": 0.4812, "step": 4601 }, { "epoch": 9.353658536585366, "grad_norm": 4.593129293767076, "learning_rate": 2.1834152857406088e-07, "loss": 0.4356, "step": 4602 }, { "epoch": 9.355691056910569, "grad_norm": 4.5098063217068445, "learning_rate": 2.1697545037718482e-07, "loss": 0.3447, "step": 4603 }, { "epoch": 9.357723577235772, "grad_norm": 5.136298061666212, "learning_rate": 2.156136122403174e-07, "loss": 0.4935, "step": 4604 }, { "epoch": 9.359756097560975, "grad_norm": 4.497130113925968, "learning_rate": 2.1425601475369207e-07, "loss": 0.4231, "step": 4605 }, { "epoch": 9.361788617886178, "grad_norm": 4.366528350225903, "learning_rate": 2.1290265850570478e-07, "loss": 0.4907, "step": 4606 }, { "epoch": 9.363821138211382, "grad_norm": 4.127463533030646, "learning_rate": 2.1155354408291418e-07, "loss": 0.358, "step": 4607 }, { "epoch": 9.365853658536585, "grad_norm": 5.539150587427233, "learning_rate": 2.1020867207004026e-07, "loss": 0.3881, "step": 4608 }, { "epoch": 9.367886178861788, "grad_norm": 5.378565768593519, "learning_rate": 2.0886804304996344e-07, "loss": 0.3469, "step": 4609 }, { "epoch": 9.369918699186991, "grad_norm": 4.831107627109776, "learning_rate": 2.0753165760372563e-07, "loss": 0.4992, "step": 4610 }, { "epoch": 9.371951219512194, "grad_norm": 4.881685576038564, "learning_rate": 2.0619951631052791e-07, "loss": 0.3998, "step": 4611 }, { "epoch": 9.373983739837398, "grad_norm": 3.9344038220134694, "learning_rate": 2.048716197477374e-07, "loss": 0.418, "step": 4612 }, { "epoch": 9.376016260162602, "grad_norm": 4.611097351483098, "learning_rate": 2.0354796849087476e-07, "loss": 0.4173, "step": 4613 }, { "epoch": 9.378048780487806, "grad_norm": 4.7968529405024505, "learning_rate": 2.022285631136234e-07, "loss": 0.4808, "step": 4614 }, { "epoch": 9.380081300813009, "grad_norm": 5.200831854332043, "learning_rate": 2.00913404187828e-07, "loss": 0.3877, "step": 4615 }, { "epoch": 9.382113821138212, "grad_norm": 5.432237509076892, "learning_rate": 1.996024922834905e-07, "loss": 0.3508, "step": 4616 }, { "epoch": 9.384146341463415, "grad_norm": 4.922234744339083, "learning_rate": 1.9829582796877522e-07, "loss": 0.567, "step": 4617 }, { "epoch": 9.386178861788618, "grad_norm": 4.979606019257048, "learning_rate": 1.969934118100003e-07, "loss": 0.3671, "step": 4618 }, { "epoch": 9.388211382113822, "grad_norm": 5.284283118901699, "learning_rate": 1.9569524437164867e-07, "loss": 0.3802, "step": 4619 }, { "epoch": 9.390243902439025, "grad_norm": 4.477235010652847, "learning_rate": 1.9440132621635687e-07, "loss": 0.3893, "step": 4620 }, { "epoch": 9.392276422764228, "grad_norm": 5.416807881149513, "learning_rate": 1.9311165790492303e-07, "loss": 0.4891, "step": 4621 }, { "epoch": 9.394308943089431, "grad_norm": 5.225478918110791, "learning_rate": 1.9182623999630335e-07, "loss": 0.4608, "step": 4622 }, { "epoch": 9.396341463414634, "grad_norm": 4.829334210237229, "learning_rate": 1.9054507304760772e-07, "loss": 0.4521, "step": 4623 }, { "epoch": 9.398373983739837, "grad_norm": 4.812056729879311, "learning_rate": 1.8926815761410867e-07, "loss": 0.399, "step": 4624 }, { "epoch": 9.40040650406504, "grad_norm": 6.179164876268586, "learning_rate": 1.8799549424923346e-07, "loss": 0.4396, "step": 4625 }, { "epoch": 9.402439024390244, "grad_norm": 4.982544160683931, "learning_rate": 1.8672708350456759e-07, "loss": 0.3738, "step": 4626 }, { "epoch": 9.404471544715447, "grad_norm": 5.8245573214691655, "learning_rate": 1.8546292592985238e-07, "loss": 0.5678, "step": 4627 }, { "epoch": 9.40650406504065, "grad_norm": 4.926753590348459, "learning_rate": 1.8420302207298623e-07, "loss": 0.4201, "step": 4628 }, { "epoch": 9.408536585365853, "grad_norm": 5.734210626899772, "learning_rate": 1.8294737248002347e-07, "loss": 0.44, "step": 4629 }, { "epoch": 9.410569105691057, "grad_norm": 5.125159832845257, "learning_rate": 1.8169597769517655e-07, "loss": 0.6232, "step": 4630 }, { "epoch": 9.41260162601626, "grad_norm": 5.178947768928123, "learning_rate": 1.804488382608116e-07, "loss": 0.4755, "step": 4631 }, { "epoch": 9.414634146341463, "grad_norm": 4.615780136318358, "learning_rate": 1.792059547174507e-07, "loss": 0.507, "step": 4632 }, { "epoch": 9.416666666666666, "grad_norm": 5.820804995344709, "learning_rate": 1.77967327603773e-07, "loss": 0.4082, "step": 4633 }, { "epoch": 9.41869918699187, "grad_norm": 5.6751934337508905, "learning_rate": 1.7673295745661124e-07, "loss": 0.4589, "step": 4634 }, { "epoch": 9.420731707317072, "grad_norm": 5.312880838586742, "learning_rate": 1.7550284481095302e-07, "loss": 0.3402, "step": 4635 }, { "epoch": 9.422764227642276, "grad_norm": 4.494043595953911, "learning_rate": 1.7427699019994415e-07, "loss": 0.2417, "step": 4636 }, { "epoch": 9.42479674796748, "grad_norm": 5.182490036329675, "learning_rate": 1.7305539415487736e-07, "loss": 0.5114, "step": 4637 }, { "epoch": 9.426829268292684, "grad_norm": 4.980021877655785, "learning_rate": 1.718380572052092e-07, "loss": 0.5939, "step": 4638 }, { "epoch": 9.428861788617887, "grad_norm": 4.975943025701276, "learning_rate": 1.7062497987854309e-07, "loss": 0.5406, "step": 4639 }, { "epoch": 9.43089430894309, "grad_norm": 5.217950194823423, "learning_rate": 1.6941616270063854e-07, "loss": 0.3493, "step": 4640 }, { "epoch": 9.432926829268293, "grad_norm": 5.183089036093734, "learning_rate": 1.6821160619541087e-07, "loss": 0.5278, "step": 4641 }, { "epoch": 9.434959349593496, "grad_norm": 5.02683432376853, "learning_rate": 1.6701131088492473e-07, "loss": 0.331, "step": 4642 }, { "epoch": 9.4369918699187, "grad_norm": 4.691851300526822, "learning_rate": 1.6581527728940173e-07, "loss": 0.3744, "step": 4643 }, { "epoch": 9.439024390243903, "grad_norm": 5.920126150392016, "learning_rate": 1.6462350592721498e-07, "loss": 0.3989, "step": 4644 }, { "epoch": 9.441056910569106, "grad_norm": 5.606632624481097, "learning_rate": 1.634359973148869e-07, "loss": 0.5642, "step": 4645 }, { "epoch": 9.44308943089431, "grad_norm": 5.7173887306233695, "learning_rate": 1.6225275196709911e-07, "loss": 0.3749, "step": 4646 }, { "epoch": 9.445121951219512, "grad_norm": 5.7255672013317005, "learning_rate": 1.6107377039668026e-07, "loss": 0.3491, "step": 4647 }, { "epoch": 9.447154471544716, "grad_norm": 4.967520037593613, "learning_rate": 1.5989905311461274e-07, "loss": 0.4753, "step": 4648 }, { "epoch": 9.449186991869919, "grad_norm": 4.460507072940082, "learning_rate": 1.5872860063003038e-07, "loss": 0.4647, "step": 4649 }, { "epoch": 9.451219512195122, "grad_norm": 5.074534823855154, "learning_rate": 1.5756241345021962e-07, "loss": 0.4694, "step": 4650 }, { "epoch": 9.453252032520325, "grad_norm": 5.3750710304409335, "learning_rate": 1.5640049208061724e-07, "loss": 0.3526, "step": 4651 }, { "epoch": 9.455284552845528, "grad_norm": 5.379799637061595, "learning_rate": 1.5524283702481158e-07, "loss": 0.4879, "step": 4652 }, { "epoch": 9.457317073170731, "grad_norm": 5.647451621753497, "learning_rate": 1.5408944878454125e-07, "loss": 0.4464, "step": 4653 }, { "epoch": 9.459349593495935, "grad_norm": 5.126010612520733, "learning_rate": 1.5294032785969748e-07, "loss": 0.4511, "step": 4654 }, { "epoch": 9.461382113821138, "grad_norm": 5.534937165026975, "learning_rate": 1.517954747483197e-07, "loss": 0.4375, "step": 4655 }, { "epoch": 9.463414634146341, "grad_norm": 5.144149901031977, "learning_rate": 1.5065488994659983e-07, "loss": 0.407, "step": 4656 }, { "epoch": 9.465447154471544, "grad_norm": 4.771948861685037, "learning_rate": 1.495185739488758e-07, "loss": 0.4, "step": 4657 }, { "epoch": 9.467479674796747, "grad_norm": 5.381564251398954, "learning_rate": 1.4838652724764146e-07, "loss": 0.3799, "step": 4658 }, { "epoch": 9.46951219512195, "grad_norm": 5.358029508382461, "learning_rate": 1.4725875033353542e-07, "loss": 0.5345, "step": 4659 }, { "epoch": 9.471544715447154, "grad_norm": 4.879353774718753, "learning_rate": 1.461352436953478e-07, "loss": 0.4066, "step": 4660 }, { "epoch": 9.473577235772357, "grad_norm": 5.285151209237403, "learning_rate": 1.450160078200169e-07, "loss": 0.4798, "step": 4661 }, { "epoch": 9.475609756097562, "grad_norm": 4.756031781039572, "learning_rate": 1.4390104319263132e-07, "loss": 0.654, "step": 4662 }, { "epoch": 9.477642276422765, "grad_norm": 4.951935614853728, "learning_rate": 1.4279035029642786e-07, "loss": 0.4479, "step": 4663 }, { "epoch": 9.479674796747968, "grad_norm": 4.878904709694279, "learning_rate": 1.4168392961279254e-07, "loss": 0.4044, "step": 4664 }, { "epoch": 9.481707317073171, "grad_norm": 5.316118598145147, "learning_rate": 1.4058178162125735e-07, "loss": 0.4003, "step": 4665 }, { "epoch": 9.483739837398375, "grad_norm": 4.6138721623090015, "learning_rate": 1.3948390679950575e-07, "loss": 0.4503, "step": 4666 }, { "epoch": 9.485772357723578, "grad_norm": 3.87040864811262, "learning_rate": 1.3839030562336596e-07, "loss": 0.2726, "step": 4667 }, { "epoch": 9.487804878048781, "grad_norm": 4.707629290149031, "learning_rate": 1.3730097856681668e-07, "loss": 0.4548, "step": 4668 }, { "epoch": 9.489837398373984, "grad_norm": 6.571442148890783, "learning_rate": 1.3621592610198354e-07, "loss": 0.4935, "step": 4669 }, { "epoch": 9.491869918699187, "grad_norm": 5.809538833048218, "learning_rate": 1.3513514869913923e-07, "loss": 0.5426, "step": 4670 }, { "epoch": 9.49390243902439, "grad_norm": 5.122408946515273, "learning_rate": 1.3405864682670133e-07, "loss": 0.4373, "step": 4671 }, { "epoch": 9.495934959349594, "grad_norm": 5.630737505329509, "learning_rate": 1.329864209512377e-07, "loss": 0.3601, "step": 4672 }, { "epoch": 9.497967479674797, "grad_norm": 4.88929568098237, "learning_rate": 1.3191847153746217e-07, "loss": 0.4295, "step": 4673 }, { "epoch": 9.5, "grad_norm": 5.0999789239682585, "learning_rate": 1.308547990482345e-07, "loss": 0.3885, "step": 4674 }, { "epoch": 9.502032520325203, "grad_norm": 5.720075323986883, "learning_rate": 1.2979540394456036e-07, "loss": 0.4337, "step": 4675 }, { "epoch": 9.504065040650406, "grad_norm": 4.895036883104883, "learning_rate": 1.2874028668559247e-07, "loss": 0.5568, "step": 4676 }, { "epoch": 9.50609756097561, "grad_norm": 5.498145860147273, "learning_rate": 1.2768944772862834e-07, "loss": 0.3164, "step": 4677 }, { "epoch": 9.508130081300813, "grad_norm": 4.645708206634186, "learning_rate": 1.2664288752911257e-07, "loss": 0.426, "step": 4678 }, { "epoch": 9.510162601626016, "grad_norm": 5.489617065636494, "learning_rate": 1.2560060654063565e-07, "loss": 0.4826, "step": 4679 }, { "epoch": 9.512195121951219, "grad_norm": 5.811221811152633, "learning_rate": 1.245626052149318e-07, "loss": 0.3642, "step": 4680 }, { "epoch": 9.514227642276422, "grad_norm": 4.963331060699771, "learning_rate": 1.2352888400188112e-07, "loss": 0.5352, "step": 4681 }, { "epoch": 9.516260162601625, "grad_norm": 5.8999645418140165, "learning_rate": 1.2249944334950748e-07, "loss": 0.3525, "step": 4682 }, { "epoch": 9.518292682926829, "grad_norm": 4.9637058798990354, "learning_rate": 1.2147428370398283e-07, "loss": 0.4066, "step": 4683 }, { "epoch": 9.520325203252032, "grad_norm": 4.835041947754773, "learning_rate": 1.2045340550961958e-07, "loss": 0.3776, "step": 4684 }, { "epoch": 9.522357723577235, "grad_norm": 5.172608741073681, "learning_rate": 1.1943680920887823e-07, "loss": 0.4571, "step": 4685 }, { "epoch": 9.524390243902438, "grad_norm": 4.463896058298046, "learning_rate": 1.1842449524236077e-07, "loss": 0.4329, "step": 4686 }, { "epoch": 9.526422764227643, "grad_norm": 5.481639419136768, "learning_rate": 1.1741646404881291e-07, "loss": 0.5548, "step": 4687 }, { "epoch": 9.528455284552846, "grad_norm": 5.960740706214499, "learning_rate": 1.164127160651285e-07, "loss": 0.4308, "step": 4688 }, { "epoch": 9.53048780487805, "grad_norm": 5.411009853852519, "learning_rate": 1.1541325172633844e-07, "loss": 0.5137, "step": 4689 }, { "epoch": 9.532520325203253, "grad_norm": 5.334446048511247, "learning_rate": 1.1441807146562401e-07, "loss": 0.3534, "step": 4690 }, { "epoch": 9.534552845528456, "grad_norm": 5.794413413071296, "learning_rate": 1.134271757143024e-07, "loss": 0.4448, "step": 4691 }, { "epoch": 9.536585365853659, "grad_norm": 5.201322451696726, "learning_rate": 1.1244056490184008e-07, "loss": 0.474, "step": 4692 }, { "epoch": 9.538617886178862, "grad_norm": 5.054531338474024, "learning_rate": 1.1145823945584499e-07, "loss": 0.5268, "step": 4693 }, { "epoch": 9.540650406504065, "grad_norm": 5.174423708914472, "learning_rate": 1.1048019980206325e-07, "loss": 0.4945, "step": 4694 }, { "epoch": 9.542682926829269, "grad_norm": 5.2900085589509, "learning_rate": 1.0950644636439023e-07, "loss": 0.4192, "step": 4695 }, { "epoch": 9.544715447154472, "grad_norm": 5.145227435722561, "learning_rate": 1.0853697956485942e-07, "loss": 0.5527, "step": 4696 }, { "epoch": 9.546747967479675, "grad_norm": 5.51432534463755, "learning_rate": 1.0757179982364586e-07, "loss": 0.373, "step": 4697 }, { "epoch": 9.548780487804878, "grad_norm": 5.142584087043826, "learning_rate": 1.0661090755907045e-07, "loss": 0.4117, "step": 4698 }, { "epoch": 9.550813008130081, "grad_norm": 5.236992453555665, "learning_rate": 1.0565430318759007e-07, "loss": 0.4382, "step": 4699 }, { "epoch": 9.552845528455284, "grad_norm": 4.847022665550203, "learning_rate": 1.0470198712381086e-07, "loss": 0.4442, "step": 4700 }, { "epoch": 9.554878048780488, "grad_norm": 5.026579775720752, "learning_rate": 1.0375395978047154e-07, "loss": 0.5106, "step": 4701 }, { "epoch": 9.55691056910569, "grad_norm": 4.7649028427292, "learning_rate": 1.0281022156846121e-07, "loss": 0.3602, "step": 4702 }, { "epoch": 9.558943089430894, "grad_norm": 4.576371891720818, "learning_rate": 1.0187077289680269e-07, "loss": 0.4492, "step": 4703 }, { "epoch": 9.560975609756097, "grad_norm": 5.097926475511236, "learning_rate": 1.009356141726614e-07, "loss": 0.571, "step": 4704 }, { "epoch": 9.5630081300813, "grad_norm": 5.281014135606403, "learning_rate": 1.0000474580134755e-07, "loss": 0.4504, "step": 4705 }, { "epoch": 9.565040650406504, "grad_norm": 5.201710001522933, "learning_rate": 9.907816818630512e-08, "loss": 0.4045, "step": 4706 }, { "epoch": 9.567073170731707, "grad_norm": 5.782963224154593, "learning_rate": 9.815588172912616e-08, "loss": 0.4709, "step": 4707 }, { "epoch": 9.56910569105691, "grad_norm": 4.587600397719616, "learning_rate": 9.723788682953539e-08, "loss": 0.4128, "step": 4708 }, { "epoch": 9.571138211382113, "grad_norm": 4.529433678122993, "learning_rate": 9.632418388540121e-08, "loss": 0.3603, "step": 4709 }, { "epoch": 9.573170731707316, "grad_norm": 4.5373734928731695, "learning_rate": 9.541477329273241e-08, "loss": 0.3487, "step": 4710 }, { "epoch": 9.57520325203252, "grad_norm": 5.200435200688203, "learning_rate": 9.450965544567702e-08, "loss": 0.4813, "step": 4711 }, { "epoch": 9.577235772357724, "grad_norm": 4.605763255386185, "learning_rate": 9.360883073652238e-08, "loss": 0.3995, "step": 4712 }, { "epoch": 9.579268292682928, "grad_norm": 4.619884754665459, "learning_rate": 9.271229955569283e-08, "loss": 0.309, "step": 4713 }, { "epoch": 9.58130081300813, "grad_norm": 4.925503489985135, "learning_rate": 9.182006229175645e-08, "loss": 0.5244, "step": 4714 }, { "epoch": 9.583333333333334, "grad_norm": 5.4401595587559575, "learning_rate": 9.09321193314161e-08, "loss": 0.5263, "step": 4715 }, { "epoch": 9.585365853658537, "grad_norm": 4.7551633863554414, "learning_rate": 9.004847105951509e-08, "loss": 0.3547, "step": 4716 }, { "epoch": 9.58739837398374, "grad_norm": 5.2985142633648445, "learning_rate": 8.916911785903593e-08, "loss": 0.4178, "step": 4717 }, { "epoch": 9.589430894308943, "grad_norm": 4.882625621739354, "learning_rate": 8.829406011109821e-08, "loss": 0.3522, "step": 4718 }, { "epoch": 9.591463414634147, "grad_norm": 5.3684851406331555, "learning_rate": 8.742329819496187e-08, "loss": 0.4264, "step": 4719 }, { "epoch": 9.59349593495935, "grad_norm": 5.114162234189964, "learning_rate": 8.655683248802282e-08, "loss": 0.4843, "step": 4720 }, { "epoch": 9.595528455284553, "grad_norm": 4.767145988960451, "learning_rate": 8.569466336581623e-08, "loss": 0.5158, "step": 4721 }, { "epoch": 9.597560975609756, "grad_norm": 4.639205595681253, "learning_rate": 8.483679120201538e-08, "loss": 0.4187, "step": 4722 }, { "epoch": 9.59959349593496, "grad_norm": 5.092150410517615, "learning_rate": 8.398321636842844e-08, "loss": 0.3549, "step": 4723 }, { "epoch": 9.601626016260163, "grad_norm": 5.307218630673954, "learning_rate": 8.313393923500613e-08, "loss": 0.4601, "step": 4724 }, { "epoch": 9.603658536585366, "grad_norm": 5.152906791543132, "learning_rate": 8.228896016983179e-08, "loss": 0.4036, "step": 4725 }, { "epoch": 9.605691056910569, "grad_norm": 5.488164479459363, "learning_rate": 8.144827953912915e-08, "loss": 0.3387, "step": 4726 }, { "epoch": 9.607723577235772, "grad_norm": 5.169091394503133, "learning_rate": 8.061189770725564e-08, "loss": 0.4709, "step": 4727 }, { "epoch": 9.609756097560975, "grad_norm": 4.760491475366661, "learning_rate": 7.977981503670795e-08, "loss": 0.5355, "step": 4728 }, { "epoch": 9.611788617886178, "grad_norm": 5.602423223987862, "learning_rate": 7.895203188811984e-08, "loss": 0.4755, "step": 4729 }, { "epoch": 9.613821138211382, "grad_norm": 5.605050484810725, "learning_rate": 7.812854862026098e-08, "loss": 0.464, "step": 4730 }, { "epoch": 9.615853658536585, "grad_norm": 6.099929229126327, "learning_rate": 7.730936559003588e-08, "loss": 0.4882, "step": 4731 }, { "epoch": 9.617886178861788, "grad_norm": 5.10098973348246, "learning_rate": 7.64944831524872e-08, "loss": 0.3889, "step": 4732 }, { "epoch": 9.619918699186991, "grad_norm": 4.003787806646006, "learning_rate": 7.568390166079243e-08, "loss": 0.4355, "step": 4733 }, { "epoch": 9.621951219512194, "grad_norm": 4.414330724241839, "learning_rate": 7.48776214662672e-08, "loss": 0.3106, "step": 4734 }, { "epoch": 9.6239837398374, "grad_norm": 4.681459600538443, "learning_rate": 7.40756429183609e-08, "loss": 0.5317, "step": 4735 }, { "epoch": 9.6260162601626, "grad_norm": 4.909133257257988, "learning_rate": 7.327796636465767e-08, "loss": 0.3338, "step": 4736 }, { "epoch": 9.628048780487806, "grad_norm": 6.054052741586877, "learning_rate": 7.248459215087988e-08, "loss": 0.5305, "step": 4737 }, { "epoch": 9.630081300813009, "grad_norm": 4.440865544654074, "learning_rate": 7.169552062088247e-08, "loss": 0.3967, "step": 4738 }, { "epoch": 9.632113821138212, "grad_norm": 5.492028246509806, "learning_rate": 7.091075211665854e-08, "loss": 0.4428, "step": 4739 }, { "epoch": 9.634146341463415, "grad_norm": 5.066783759511349, "learning_rate": 7.01302869783338e-08, "loss": 0.4513, "step": 4740 }, { "epoch": 9.636178861788618, "grad_norm": 5.050086980610208, "learning_rate": 6.935412554416876e-08, "loss": 0.4095, "step": 4741 }, { "epoch": 9.638211382113822, "grad_norm": 4.255943889889312, "learning_rate": 6.858226815056101e-08, "loss": 0.4524, "step": 4742 }, { "epoch": 9.640243902439025, "grad_norm": 4.549821676164648, "learning_rate": 6.781471513203964e-08, "loss": 0.5027, "step": 4743 }, { "epoch": 9.642276422764228, "grad_norm": 5.318707209045543, "learning_rate": 6.705146682127184e-08, "loss": 0.4723, "step": 4744 }, { "epoch": 9.644308943089431, "grad_norm": 4.524330290627297, "learning_rate": 6.629252354905413e-08, "loss": 0.3444, "step": 4745 }, { "epoch": 9.646341463414634, "grad_norm": 5.622164666593851, "learning_rate": 6.55378856443234e-08, "loss": 0.4561, "step": 4746 }, { "epoch": 9.648373983739837, "grad_norm": 4.678574614966206, "learning_rate": 6.478755343414355e-08, "loss": 0.4569, "step": 4747 }, { "epoch": 9.65040650406504, "grad_norm": 5.362928874049421, "learning_rate": 6.404152724371892e-08, "loss": 0.3759, "step": 4748 }, { "epoch": 9.652439024390244, "grad_norm": 5.498762199146369, "learning_rate": 6.329980739638419e-08, "loss": 0.4449, "step": 4749 }, { "epoch": 9.654471544715447, "grad_norm": 4.580334951639198, "learning_rate": 6.256239421360666e-08, "loss": 0.5456, "step": 4750 }, { "epoch": 9.65650406504065, "grad_norm": 4.921666128315923, "learning_rate": 6.182928801498955e-08, "loss": 0.3658, "step": 4751 }, { "epoch": 9.658536585365853, "grad_norm": 4.835364193897308, "learning_rate": 6.110048911826871e-08, "loss": 0.4109, "step": 4752 }, { "epoch": 9.660569105691057, "grad_norm": 5.565362486460845, "learning_rate": 6.037599783931147e-08, "loss": 0.3597, "step": 4753 }, { "epoch": 9.66260162601626, "grad_norm": 4.812445307724306, "learning_rate": 5.965581449212221e-08, "loss": 0.3657, "step": 4754 }, { "epoch": 9.664634146341463, "grad_norm": 5.444989942906382, "learning_rate": 5.8939939388832355e-08, "loss": 0.4995, "step": 4755 }, { "epoch": 9.666666666666666, "grad_norm": 4.835716645400037, "learning_rate": 5.82283728397115e-08, "loss": 0.4048, "step": 4756 }, { "epoch": 9.66869918699187, "grad_norm": 5.505368909409704, "learning_rate": 5.75211151531585e-08, "loss": 0.366, "step": 4757 }, { "epoch": 9.670731707317072, "grad_norm": 5.249495119284508, "learning_rate": 5.681816663570594e-08, "loss": 0.3739, "step": 4758 }, { "epoch": 9.672764227642276, "grad_norm": 4.785984038972887, "learning_rate": 5.6119527592020106e-08, "loss": 0.3321, "step": 4759 }, { "epoch": 9.67479674796748, "grad_norm": 5.036814047081234, "learning_rate": 5.542519832489546e-08, "loss": 0.3282, "step": 4760 }, { "epoch": 9.676829268292684, "grad_norm": 5.058117826446424, "learning_rate": 5.473517913526349e-08, "loss": 0.399, "step": 4761 }, { "epoch": 9.678861788617887, "grad_norm": 5.650120625378251, "learning_rate": 5.4049470322184996e-08, "loss": 0.4802, "step": 4762 }, { "epoch": 9.68089430894309, "grad_norm": 5.266967333130306, "learning_rate": 5.3368072182852227e-08, "loss": 0.4205, "step": 4763 }, { "epoch": 9.682926829268293, "grad_norm": 5.808046281903578, "learning_rate": 5.269098501259007e-08, "loss": 0.4545, "step": 4764 }, { "epoch": 9.684959349593496, "grad_norm": 5.329473673782954, "learning_rate": 5.201820910485489e-08, "loss": 0.4572, "step": 4765 }, { "epoch": 9.6869918699187, "grad_norm": 5.532488731971633, "learning_rate": 5.134974475123566e-08, "loss": 0.4086, "step": 4766 }, { "epoch": 9.689024390243903, "grad_norm": 5.205992449510503, "learning_rate": 5.068559224144953e-08, "loss": 0.3531, "step": 4767 }, { "epoch": 9.691056910569106, "grad_norm": 5.604862067139619, "learning_rate": 5.002575186334735e-08, "loss": 0.5037, "step": 4768 }, { "epoch": 9.69308943089431, "grad_norm": 4.412928835148684, "learning_rate": 4.9370223902911466e-08, "loss": 0.3531, "step": 4769 }, { "epoch": 9.695121951219512, "grad_norm": 5.437188327271127, "learning_rate": 4.8719008644253495e-08, "loss": 0.4146, "step": 4770 }, { "epoch": 9.697154471544716, "grad_norm": 4.666325516444621, "learning_rate": 4.807210636961657e-08, "loss": 0.4661, "step": 4771 }, { "epoch": 9.699186991869919, "grad_norm": 5.4656771944930265, "learning_rate": 4.742951735937418e-08, "loss": 0.5827, "step": 4772 }, { "epoch": 9.701219512195122, "grad_norm": 4.971011405452489, "learning_rate": 4.679124189203243e-08, "loss": 0.3963, "step": 4773 }, { "epoch": 9.703252032520325, "grad_norm": 5.098683222036838, "learning_rate": 4.61572802442245e-08, "loss": 0.3251, "step": 4774 }, { "epoch": 9.705284552845528, "grad_norm": 6.204502349877874, "learning_rate": 4.5527632690715026e-08, "loss": 0.4998, "step": 4775 }, { "epoch": 9.707317073170731, "grad_norm": 4.85275327642955, "learning_rate": 4.490229950440239e-08, "loss": 0.3887, "step": 4776 }, { "epoch": 9.709349593495935, "grad_norm": 4.471483283922551, "learning_rate": 4.4281280956309794e-08, "loss": 0.4818, "step": 4777 }, { "epoch": 9.711382113821138, "grad_norm": 4.993555675104413, "learning_rate": 4.3664577315593036e-08, "loss": 0.4226, "step": 4778 }, { "epoch": 9.713414634146341, "grad_norm": 5.488634685130595, "learning_rate": 4.305218884953943e-08, "loss": 0.4509, "step": 4779 }, { "epoch": 9.715447154471544, "grad_norm": 5.170507424543804, "learning_rate": 4.2444115823562226e-08, "loss": 0.4601, "step": 4780 }, { "epoch": 9.717479674796747, "grad_norm": 4.986713111217568, "learning_rate": 4.1840358501207266e-08, "loss": 0.4815, "step": 4781 }, { "epoch": 9.71951219512195, "grad_norm": 5.027203561128033, "learning_rate": 4.124091714414857e-08, "loss": 0.4462, "step": 4782 }, { "epoch": 9.721544715447154, "grad_norm": 5.083930065271835, "learning_rate": 4.064579201218943e-08, "loss": 0.3875, "step": 4783 }, { "epoch": 9.723577235772357, "grad_norm": 5.660017717186132, "learning_rate": 4.005498336326463e-08, "loss": 0.5381, "step": 4784 }, { "epoch": 9.725609756097562, "grad_norm": 4.949614271797442, "learning_rate": 3.9468491453434896e-08, "loss": 0.4985, "step": 4785 }, { "epoch": 9.727642276422765, "grad_norm": 5.02380429712434, "learning_rate": 3.8886316536893564e-08, "loss": 0.4743, "step": 4786 }, { "epoch": 9.729674796747968, "grad_norm": 4.436889145798545, "learning_rate": 3.8308458865958795e-08, "loss": 0.5201, "step": 4787 }, { "epoch": 9.731707317073171, "grad_norm": 5.485509590651052, "learning_rate": 3.773491869108137e-08, "loss": 0.4677, "step": 4788 }, { "epoch": 9.733739837398375, "grad_norm": 5.4820216825731904, "learning_rate": 3.716569626083799e-08, "loss": 0.3796, "step": 4789 }, { "epoch": 9.735772357723578, "grad_norm": 4.625192006282044, "learning_rate": 3.660079182193688e-08, "loss": 0.4328, "step": 4790 }, { "epoch": 9.737804878048781, "grad_norm": 5.218883220331888, "learning_rate": 3.604020561921218e-08, "loss": 0.5127, "step": 4791 }, { "epoch": 9.739837398373984, "grad_norm": 5.107658851470709, "learning_rate": 3.548393789562732e-08, "loss": 0.329, "step": 4792 }, { "epoch": 9.741869918699187, "grad_norm": 4.849970198275561, "learning_rate": 3.4931988892275e-08, "loss": 0.4498, "step": 4793 }, { "epoch": 9.74390243902439, "grad_norm": 4.899577970164322, "learning_rate": 3.4384358848374985e-08, "loss": 0.3512, "step": 4794 }, { "epoch": 9.745934959349594, "grad_norm": 5.264247264540461, "learning_rate": 3.38410480012763e-08, "loss": 0.4501, "step": 4795 }, { "epoch": 9.747967479674797, "grad_norm": 4.882486927661266, "learning_rate": 3.3302056586453916e-08, "loss": 0.4666, "step": 4796 }, { "epoch": 9.75, "grad_norm": 4.085702902011094, "learning_rate": 3.2767384837513185e-08, "loss": 0.4313, "step": 4797 }, { "epoch": 9.752032520325203, "grad_norm": 5.168880454675395, "learning_rate": 3.2237032986185415e-08, "loss": 0.5485, "step": 4798 }, { "epoch": 9.754065040650406, "grad_norm": 4.302732474200469, "learning_rate": 3.1711001262331175e-08, "loss": 0.4221, "step": 4799 }, { "epoch": 9.75609756097561, "grad_norm": 4.691621676750135, "learning_rate": 3.118928989393699e-08, "loss": 0.4278, "step": 4800 }, { "epoch": 9.758130081300813, "grad_norm": 5.366248541168582, "learning_rate": 3.0671899107118654e-08, "loss": 0.4068, "step": 4801 }, { "epoch": 9.760162601626016, "grad_norm": 4.867061552702161, "learning_rate": 3.015882912611789e-08, "loss": 0.3893, "step": 4802 }, { "epoch": 9.762195121951219, "grad_norm": 5.398997409190775, "learning_rate": 2.965008017330462e-08, "loss": 0.4538, "step": 4803 }, { "epoch": 9.764227642276422, "grad_norm": 4.634219062943048, "learning_rate": 2.9145652469174666e-08, "loss": 0.5126, "step": 4804 }, { "epoch": 9.766260162601625, "grad_norm": 5.220049030884407, "learning_rate": 2.864554623235316e-08, "loss": 0.513, "step": 4805 }, { "epoch": 9.768292682926829, "grad_norm": 4.777742722961318, "learning_rate": 2.814976167959116e-08, "loss": 0.4103, "step": 4806 }, { "epoch": 9.770325203252032, "grad_norm": 5.623525983238057, "learning_rate": 2.7658299025765668e-08, "loss": 0.4041, "step": 4807 }, { "epoch": 9.772357723577235, "grad_norm": 4.112275309669542, "learning_rate": 2.7171158483882963e-08, "loss": 0.4641, "step": 4808 }, { "epoch": 9.774390243902438, "grad_norm": 5.224103946153961, "learning_rate": 2.6688340265071943e-08, "loss": 0.5289, "step": 4809 }, { "epoch": 9.776422764227643, "grad_norm": 4.366344089900083, "learning_rate": 2.6209844578594102e-08, "loss": 0.3739, "step": 4810 }, { "epoch": 9.778455284552846, "grad_norm": 4.205315737796296, "learning_rate": 2.573567163183244e-08, "loss": 0.4697, "step": 4811 }, { "epoch": 9.78048780487805, "grad_norm": 4.8564192035318055, "learning_rate": 2.5265821630298116e-08, "loss": 0.4746, "step": 4812 }, { "epoch": 9.782520325203253, "grad_norm": 4.5086493574429385, "learning_rate": 2.4800294777629352e-08, "loss": 0.4602, "step": 4813 }, { "epoch": 9.784552845528456, "grad_norm": 4.569148154325319, "learning_rate": 2.43390912755892e-08, "loss": 0.4712, "step": 4814 }, { "epoch": 9.786585365853659, "grad_norm": 4.7506999169587765, "learning_rate": 2.388221132406998e-08, "loss": 0.3639, "step": 4815 }, { "epoch": 9.788617886178862, "grad_norm": 5.2131643912978465, "learning_rate": 2.3429655121085525e-08, "loss": 0.541, "step": 4816 }, { "epoch": 9.790650406504065, "grad_norm": 5.724044136932834, "learning_rate": 2.2981422862778934e-08, "loss": 0.524, "step": 4817 }, { "epoch": 9.792682926829269, "grad_norm": 5.13479582167489, "learning_rate": 2.2537514743419252e-08, "loss": 0.4141, "step": 4818 }, { "epoch": 9.794715447154472, "grad_norm": 5.259235251014618, "learning_rate": 2.209793095540036e-08, "loss": 0.4322, "step": 4819 }, { "epoch": 9.796747967479675, "grad_norm": 4.350432974351853, "learning_rate": 2.1662671689242076e-08, "loss": 0.4949, "step": 4820 }, { "epoch": 9.798780487804878, "grad_norm": 5.430779778928208, "learning_rate": 2.123173713358906e-08, "loss": 0.5308, "step": 4821 }, { "epoch": 9.800813008130081, "grad_norm": 5.476694465080225, "learning_rate": 2.0805127475214128e-08, "loss": 0.3929, "step": 4822 }, { "epoch": 9.802845528455284, "grad_norm": 5.012205877985282, "learning_rate": 2.0382842899014933e-08, "loss": 0.3551, "step": 4823 }, { "epoch": 9.804878048780488, "grad_norm": 5.491293210317716, "learning_rate": 1.996488358801174e-08, "loss": 0.3782, "step": 4824 }, { "epoch": 9.80691056910569, "grad_norm": 5.125772818420642, "learning_rate": 1.9551249723352982e-08, "loss": 0.4386, "step": 4825 }, { "epoch": 9.808943089430894, "grad_norm": 4.724904609069452, "learning_rate": 1.9141941484310812e-08, "loss": 0.4278, "step": 4826 }, { "epoch": 9.810975609756097, "grad_norm": 5.331252102963821, "learning_rate": 1.873695904828554e-08, "loss": 0.436, "step": 4827 }, { "epoch": 9.8130081300813, "grad_norm": 4.424161303561169, "learning_rate": 1.8336302590798992e-08, "loss": 0.3993, "step": 4828 }, { "epoch": 9.815040650406504, "grad_norm": 4.871588554982781, "learning_rate": 1.7939972285501152e-08, "loss": 0.4699, "step": 4829 }, { "epoch": 9.817073170731707, "grad_norm": 5.79658388425697, "learning_rate": 1.75479683041635e-08, "loss": 0.4605, "step": 4830 }, { "epoch": 9.81910569105691, "grad_norm": 4.8029073950299646, "learning_rate": 1.7160290816685688e-08, "loss": 0.5202, "step": 4831 }, { "epoch": 9.821138211382113, "grad_norm": 4.770275819341984, "learning_rate": 1.677693999109109e-08, "loss": 0.4372, "step": 4832 }, { "epoch": 9.823170731707316, "grad_norm": 5.385251687251179, "learning_rate": 1.6397915993526804e-08, "loss": 0.292, "step": 4833 }, { "epoch": 9.82520325203252, "grad_norm": 5.270950161777868, "learning_rate": 1.6023218988266977e-08, "loss": 0.4909, "step": 4834 }, { "epoch": 9.827235772357724, "grad_norm": 5.328197489629693, "learning_rate": 1.565284913770837e-08, "loss": 0.4835, "step": 4835 }, { "epoch": 9.829268292682928, "grad_norm": 5.099898437875519, "learning_rate": 1.5286806602372583e-08, "loss": 0.4625, "step": 4836 }, { "epoch": 9.83130081300813, "grad_norm": 5.748268465739985, "learning_rate": 1.4925091540906044e-08, "loss": 0.552, "step": 4837 }, { "epoch": 9.833333333333334, "grad_norm": 5.355532781915136, "learning_rate": 1.4567704110080016e-08, "loss": 0.4618, "step": 4838 }, { "epoch": 9.835365853658537, "grad_norm": 4.783622943368642, "learning_rate": 1.4214644464789485e-08, "loss": 0.3653, "step": 4839 }, { "epoch": 9.83739837398374, "grad_norm": 5.507382241807975, "learning_rate": 1.3865912758054267e-08, "loss": 0.3573, "step": 4840 }, { "epoch": 9.839430894308943, "grad_norm": 5.197414325377537, "learning_rate": 1.3521509141019017e-08, "loss": 0.4382, "step": 4841 }, { "epoch": 9.841463414634147, "grad_norm": 5.394147227018569, "learning_rate": 1.318143376294989e-08, "loss": 0.4326, "step": 4842 }, { "epoch": 9.84349593495935, "grad_norm": 5.034238484769971, "learning_rate": 1.2845686771240096e-08, "loss": 0.484, "step": 4843 }, { "epoch": 9.845528455284553, "grad_norm": 4.4956715963711655, "learning_rate": 1.2514268311405452e-08, "loss": 0.3881, "step": 4844 }, { "epoch": 9.847560975609756, "grad_norm": 4.705446243704091, "learning_rate": 1.2187178527085509e-08, "loss": 0.4935, "step": 4845 }, { "epoch": 9.84959349593496, "grad_norm": 5.6470178325725895, "learning_rate": 1.1864417560045749e-08, "loss": 0.4094, "step": 4846 }, { "epoch": 9.851626016260163, "grad_norm": 4.616066862356433, "learning_rate": 1.1545985550172056e-08, "loss": 0.3844, "step": 4847 }, { "epoch": 9.853658536585366, "grad_norm": 4.961942981220772, "learning_rate": 1.1231882635477364e-08, "loss": 0.4571, "step": 4848 }, { "epoch": 9.855691056910569, "grad_norm": 4.761875819060113, "learning_rate": 1.0922108952096111e-08, "loss": 0.3708, "step": 4849 }, { "epoch": 9.857723577235772, "grad_norm": 5.004818795101893, "learning_rate": 1.0616664634288675e-08, "loss": 0.4789, "step": 4850 }, { "epoch": 9.859756097560975, "grad_norm": 5.006966871826638, "learning_rate": 1.0315549814435832e-08, "loss": 0.6304, "step": 4851 }, { "epoch": 9.861788617886178, "grad_norm": 5.539079538432507, "learning_rate": 1.0018764623045407e-08, "loss": 0.5607, "step": 4852 }, { "epoch": 9.863821138211382, "grad_norm": 5.2977376579616, "learning_rate": 9.72630918874562e-09, "loss": 0.4468, "step": 4853 }, { "epoch": 9.865853658536585, "grad_norm": 4.921526991969381, "learning_rate": 9.438183638289523e-09, "loss": 0.4275, "step": 4854 }, { "epoch": 9.867886178861788, "grad_norm": 5.139419595067469, "learning_rate": 9.154388096555e-09, "loss": 0.4461, "step": 4855 }, { "epoch": 9.869918699186991, "grad_norm": 5.250103244003162, "learning_rate": 8.874922686541442e-09, "loss": 0.4445, "step": 4856 }, { "epoch": 9.871951219512194, "grad_norm": 4.942794536523991, "learning_rate": 8.599787529371962e-09, "loss": 0.4198, "step": 4857 }, { "epoch": 9.8739837398374, "grad_norm": 4.646253038616247, "learning_rate": 8.32898274429117e-09, "loss": 0.3567, "step": 4858 }, { "epoch": 9.8760162601626, "grad_norm": 4.896855737079579, "learning_rate": 8.062508448670736e-09, "loss": 0.446, "step": 4859 }, { "epoch": 9.878048780487806, "grad_norm": 5.957825556174205, "learning_rate": 7.800364758002721e-09, "loss": 0.3788, "step": 4860 }, { "epoch": 9.880081300813009, "grad_norm": 4.8868235662574815, "learning_rate": 7.542551785901798e-09, "loss": 0.4973, "step": 4861 }, { "epoch": 9.882113821138212, "grad_norm": 5.957657102956786, "learning_rate": 7.289069644107471e-09, "loss": 0.4513, "step": 4862 }, { "epoch": 9.884146341463415, "grad_norm": 5.337886443435937, "learning_rate": 7.039918442481863e-09, "loss": 0.4009, "step": 4863 }, { "epoch": 9.886178861788618, "grad_norm": 4.258118876233858, "learning_rate": 6.795098289008595e-09, "loss": 0.5445, "step": 4864 }, { "epoch": 9.888211382113822, "grad_norm": 5.2843199343095835, "learning_rate": 6.554609289796121e-09, "loss": 0.48, "step": 4865 }, { "epoch": 9.890243902439025, "grad_norm": 5.429638951887208, "learning_rate": 6.3184515490732904e-09, "loss": 0.3718, "step": 4866 }, { "epoch": 9.892276422764228, "grad_norm": 4.743864154244451, "learning_rate": 6.086625169194893e-09, "loss": 0.444, "step": 4867 }, { "epoch": 9.894308943089431, "grad_norm": 5.507170434416024, "learning_rate": 5.859130250636113e-09, "loss": 0.403, "step": 4868 }, { "epoch": 9.896341463414634, "grad_norm": 5.256774497154064, "learning_rate": 5.635966891994749e-09, "loss": 0.5713, "step": 4869 }, { "epoch": 9.898373983739837, "grad_norm": 4.860365501841098, "learning_rate": 5.417135189993428e-09, "loss": 0.4536, "step": 4870 }, { "epoch": 9.90040650406504, "grad_norm": 5.239308171869635, "learning_rate": 5.202635239475173e-09, "loss": 0.4075, "step": 4871 }, { "epoch": 9.902439024390244, "grad_norm": 4.414289316933377, "learning_rate": 4.992467133406731e-09, "loss": 0.3735, "step": 4872 }, { "epoch": 9.904471544715447, "grad_norm": 5.056935441948214, "learning_rate": 4.786630962876348e-09, "loss": 0.4332, "step": 4873 }, { "epoch": 9.90650406504065, "grad_norm": 5.177224730993976, "learning_rate": 4.585126817095997e-09, "loss": 0.3848, "step": 4874 }, { "epoch": 9.908536585365853, "grad_norm": 5.246471147736484, "learning_rate": 4.387954783400261e-09, "loss": 0.3149, "step": 4875 }, { "epoch": 9.910569105691057, "grad_norm": 5.01125199147235, "learning_rate": 4.195114947244117e-09, "loss": 0.4317, "step": 4876 }, { "epoch": 9.91260162601626, "grad_norm": 4.7376264257620075, "learning_rate": 4.006607392206263e-09, "loss": 0.4338, "step": 4877 }, { "epoch": 9.914634146341463, "grad_norm": 4.590881959430625, "learning_rate": 3.822432199989123e-09, "loss": 0.4507, "step": 4878 }, { "epoch": 9.916666666666666, "grad_norm": 6.375730556951399, "learning_rate": 3.6425894504155124e-09, "loss": 0.4414, "step": 4879 }, { "epoch": 9.91869918699187, "grad_norm": 4.708330467101292, "learning_rate": 3.4670792214297476e-09, "loss": 0.431, "step": 4880 }, { "epoch": 9.920731707317072, "grad_norm": 4.805475906541682, "learning_rate": 3.2959015891020905e-09, "loss": 0.3591, "step": 4881 }, { "epoch": 9.922764227642276, "grad_norm": 4.678644104753161, "learning_rate": 3.1290566276198643e-09, "loss": 0.4186, "step": 4882 }, { "epoch": 9.92479674796748, "grad_norm": 4.465422943993956, "learning_rate": 2.966544409298555e-09, "loss": 0.328, "step": 4883 }, { "epoch": 9.926829268292684, "grad_norm": 5.088865889400908, "learning_rate": 2.808365004569602e-09, "loss": 0.3883, "step": 4884 }, { "epoch": 9.928861788617887, "grad_norm": 6.022062265968364, "learning_rate": 2.6545184819903867e-09, "loss": 0.4548, "step": 4885 }, { "epoch": 9.93089430894309, "grad_norm": 5.314353151731265, "learning_rate": 2.5050049082409046e-09, "loss": 0.4413, "step": 4886 }, { "epoch": 9.932926829268293, "grad_norm": 5.676845565763262, "learning_rate": 2.3598243481204318e-09, "loss": 0.5172, "step": 4887 }, { "epoch": 9.934959349593496, "grad_norm": 5.444557872511462, "learning_rate": 2.2189768645519693e-09, "loss": 0.6576, "step": 4888 }, { "epoch": 9.9369918699187, "grad_norm": 5.440894939874042, "learning_rate": 2.0824625185800195e-09, "loss": 0.4924, "step": 4889 }, { "epoch": 9.939024390243903, "grad_norm": 4.7944702282597405, "learning_rate": 1.950281369371698e-09, "loss": 0.4628, "step": 4890 }, { "epoch": 9.941056910569106, "grad_norm": 5.4272712824181575, "learning_rate": 1.8224334742156235e-09, "loss": 0.4602, "step": 4891 }, { "epoch": 9.94308943089431, "grad_norm": 5.030380146603478, "learning_rate": 1.6989188885219165e-09, "loss": 0.3953, "step": 4892 }, { "epoch": 9.945121951219512, "grad_norm": 3.884605044565729, "learning_rate": 1.579737665823311e-09, "loss": 0.4291, "step": 4893 }, { "epoch": 9.947154471544716, "grad_norm": 4.685137821997196, "learning_rate": 1.4648898577740434e-09, "loss": 0.4881, "step": 4894 }, { "epoch": 9.949186991869919, "grad_norm": 5.643622541153047, "learning_rate": 1.3543755141498527e-09, "loss": 0.5411, "step": 4895 }, { "epoch": 9.951219512195122, "grad_norm": 5.421858984162554, "learning_rate": 1.2481946828502011e-09, "loss": 0.5696, "step": 4896 }, { "epoch": 9.953252032520325, "grad_norm": 4.744561760082286, "learning_rate": 1.146347409892723e-09, "loss": 0.447, "step": 4897 }, { "epoch": 9.955284552845528, "grad_norm": 5.124432493939918, "learning_rate": 1.0488337394221059e-09, "loss": 0.4571, "step": 4898 }, { "epoch": 9.957317073170731, "grad_norm": 4.9996995635485115, "learning_rate": 9.556537136978794e-10, "loss": 0.4326, "step": 4899 }, { "epoch": 9.959349593495935, "grad_norm": 5.518044866200494, "learning_rate": 8.668073731088467e-10, "loss": 0.3461, "step": 4900 }, { "epoch": 9.961382113821138, "grad_norm": 4.506193896885158, "learning_rate": 7.822947561597627e-10, "loss": 0.4443, "step": 4901 }, { "epoch": 9.963414634146341, "grad_norm": 4.99898050666623, "learning_rate": 7.021158994791055e-10, "loss": 0.4708, "step": 4902 }, { "epoch": 9.965447154471544, "grad_norm": 6.157697037698256, "learning_rate": 6.262708378179661e-10, "loss": 0.3786, "step": 4903 }, { "epoch": 9.967479674796747, "grad_norm": 4.217304316508705, "learning_rate": 5.547596040489378e-10, "loss": 0.4655, "step": 4904 }, { "epoch": 9.96951219512195, "grad_norm": 5.169639068394747, "learning_rate": 4.875822291638965e-10, "loss": 0.5577, "step": 4905 }, { "epoch": 9.971544715447154, "grad_norm": 5.653333483692406, "learning_rate": 4.247387422806615e-10, "loss": 0.4647, "step": 4906 }, { "epoch": 9.973577235772357, "grad_norm": 5.158196858169236, "learning_rate": 3.6622917063411367e-10, "loss": 0.3475, "step": 4907 }, { "epoch": 9.975609756097562, "grad_norm": 5.3337701037775815, "learning_rate": 3.1205353958285724e-10, "loss": 0.3942, "step": 4908 }, { "epoch": 9.977642276422765, "grad_norm": 5.814710937887969, "learning_rate": 2.622118726092193e-10, "loss": 0.3076, "step": 4909 }, { "epoch": 9.979674796747968, "grad_norm": 5.535096059498302, "learning_rate": 2.167041913125889e-10, "loss": 0.4314, "step": 4910 }, { "epoch": 9.981707317073171, "grad_norm": 4.854098053561679, "learning_rate": 1.7553051541829847e-10, "loss": 0.4577, "step": 4911 }, { "epoch": 9.983739837398375, "grad_norm": 5.030223605673868, "learning_rate": 1.3869086276985243e-10, "loss": 0.3696, "step": 4912 }, { "epoch": 9.985772357723578, "grad_norm": 4.915940180186763, "learning_rate": 1.061852493355886e-10, "loss": 0.4411, "step": 4913 }, { "epoch": 9.987804878048781, "grad_norm": 5.045023933509812, "learning_rate": 7.801368920201669e-11, "loss": 0.4291, "step": 4914 }, { "epoch": 9.989837398373984, "grad_norm": 5.193205930637915, "learning_rate": 5.417619458158996e-11, "loss": 0.409, "step": 4915 }, { "epoch": 9.991869918699187, "grad_norm": 4.851940689278832, "learning_rate": 3.467277580271322e-11, "loss": 0.5348, "step": 4916 }, { "epoch": 9.99390243902439, "grad_norm": 4.376709838786734, "learning_rate": 1.9503441319734807e-11, "loss": 0.4216, "step": 4917 }, { "epoch": 9.995934959349594, "grad_norm": 4.1650726573814385, "learning_rate": 8.668197707395464e-12, "loss": 0.4022, "step": 4918 }, { "epoch": 9.997967479674797, "grad_norm": 5.770481678797226, "learning_rate": 2.1670496619385915e-12, "loss": 0.3748, "step": 4919 }, { "epoch": 10.0, "grad_norm": 5.390728193999499, "learning_rate": 0.0, "loss": 0.4344, "step": 4920 }, { "epoch": 10.0, "eval_loss": 1.0637468099594116, "eval_runtime": 131.9541, "eval_samples_per_second": 7.457, "eval_steps_per_second": 0.932, "step": 4920 }, { "epoch": 10.0, "step": 4920, "total_flos": 2934188109987840.0, "train_loss": 0.7051953146249298, "train_runtime": 20412.1785, "train_samples_per_second": 1.928, "train_steps_per_second": 0.241 } ], "logging_steps": 1.0, "max_steps": 4920, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2934188109987840.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }