|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 125.0, |
|
"eval_steps": 1000, |
|
"global_step": 250, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 3.334595203399658, |
|
"learning_rate": 0.0, |
|
"loss": 1.3402, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 3.309346914291382, |
|
"learning_rate": 1.5051499783199055e-07, |
|
"loss": 1.284, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 3.4695301055908203, |
|
"learning_rate": 2.385606273598312e-07, |
|
"loss": 1.3362, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 2.9988322257995605, |
|
"learning_rate": 3.010299956639811e-07, |
|
"loss": 1.2535, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 3.0436394214630127, |
|
"learning_rate": 3.494850021680093e-07, |
|
"loss": 1.2326, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 3.329270601272583, |
|
"learning_rate": 3.8907562519182173e-07, |
|
"loss": 1.3378, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"grad_norm": 3.0574333667755127, |
|
"learning_rate": 4.2254902000712834e-07, |
|
"loss": 1.4162, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 3.348349094390869, |
|
"learning_rate": 4.5154499349597166e-07, |
|
"loss": 1.3598, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"grad_norm": 2.9368350505828857, |
|
"learning_rate": 4.771212547196623e-07, |
|
"loss": 1.1551, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 3.3986520767211914, |
|
"learning_rate": 4.999999999999999e-07, |
|
"loss": 1.4285, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"grad_norm": 3.004727363586426, |
|
"learning_rate": 5.206963425791124e-07, |
|
"loss": 1.2629, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"grad_norm": 3.090939998626709, |
|
"learning_rate": 5.395906230238123e-07, |
|
"loss": 1.3378, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"grad_norm": 2.906198740005493, |
|
"learning_rate": 5.569716761534182e-07, |
|
"loss": 1.3384, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"grad_norm": 2.9036598205566406, |
|
"learning_rate": 5.730640178391189e-07, |
|
"loss": 1.2693, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"grad_norm": 2.924349069595337, |
|
"learning_rate": 5.880456295278405e-07, |
|
"loss": 1.2711, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 3.110586166381836, |
|
"learning_rate": 6.020599913279622e-07, |
|
"loss": 1.2708, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"grad_norm": 3.2055583000183105, |
|
"learning_rate": 6.15224460689137e-07, |
|
"loss": 1.4757, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"grad_norm": 2.972228527069092, |
|
"learning_rate": 6.276362525516529e-07, |
|
"loss": 1.3115, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"grad_norm": 2.447122573852539, |
|
"learning_rate": 6.393768004764143e-07, |
|
"loss": 1.241, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 2.3147361278533936, |
|
"learning_rate": 6.505149978319905e-07, |
|
"loss": 1.2135, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 10.5, |
|
"grad_norm": 2.9482200145721436, |
|
"learning_rate": 6.611096473669595e-07, |
|
"loss": 1.4155, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"grad_norm": 2.6261355876922607, |
|
"learning_rate": 6.712113404111031e-07, |
|
"loss": 1.2434, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 11.5, |
|
"grad_norm": 2.448061943054199, |
|
"learning_rate": 6.808639180087963e-07, |
|
"loss": 1.2424, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"grad_norm": 2.22385311126709, |
|
"learning_rate": 6.901056208558029e-07, |
|
"loss": 1.1388, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 12.5, |
|
"grad_norm": 2.7259082794189453, |
|
"learning_rate": 6.989700043360186e-07, |
|
"loss": 1.4238, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"grad_norm": 2.03949236869812, |
|
"learning_rate": 7.074866739854088e-07, |
|
"loss": 1.0655, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 13.5, |
|
"grad_norm": 2.341221809387207, |
|
"learning_rate": 7.156818820794935e-07, |
|
"loss": 1.2652, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"grad_norm": 2.2585113048553467, |
|
"learning_rate": 7.235790156711094e-07, |
|
"loss": 1.1248, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 14.5, |
|
"grad_norm": 2.2248682975769043, |
|
"learning_rate": 7.311989989494779e-07, |
|
"loss": 1.1755, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"grad_norm": 2.115250825881958, |
|
"learning_rate": 7.38560627359831e-07, |
|
"loss": 1.1932, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 15.5, |
|
"grad_norm": 2.030606269836426, |
|
"learning_rate": 7.456808469171361e-07, |
|
"loss": 1.2845, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"grad_norm": 2.0579230785369873, |
|
"learning_rate": 7.525749891599529e-07, |
|
"loss": 1.1964, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 16.5, |
|
"grad_norm": 2.074784517288208, |
|
"learning_rate": 7.592569699389436e-07, |
|
"loss": 1.1445, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"grad_norm": 2.064431667327881, |
|
"learning_rate": 7.657394585211274e-07, |
|
"loss": 1.2398, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 17.5, |
|
"grad_norm": 1.8380582332611084, |
|
"learning_rate": 7.720340221751376e-07, |
|
"loss": 1.1373, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"grad_norm": 2.3023219108581543, |
|
"learning_rate": 7.781512503836435e-07, |
|
"loss": 1.0978, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 18.5, |
|
"grad_norm": 2.0860466957092285, |
|
"learning_rate": 7.841008620334974e-07, |
|
"loss": 1.1017, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"grad_norm": 1.9423589706420898, |
|
"learning_rate": 7.89891798308405e-07, |
|
"loss": 1.1538, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 19.5, |
|
"grad_norm": 2.018376350402832, |
|
"learning_rate": 7.955323035132494e-07, |
|
"loss": 1.1345, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"grad_norm": 1.9834052324295044, |
|
"learning_rate": 8.01029995663981e-07, |
|
"loss": 1.1596, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 20.5, |
|
"grad_norm": 1.8433727025985718, |
|
"learning_rate": 8.063919283598676e-07, |
|
"loss": 0.9934, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"grad_norm": 2.0120983123779297, |
|
"learning_rate": 8.116246451989502e-07, |
|
"loss": 1.1849, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 21.5, |
|
"grad_norm": 1.7347400188446045, |
|
"learning_rate": 8.16734227789793e-07, |
|
"loss": 0.9969, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"grad_norm": 2.1413559913635254, |
|
"learning_rate": 8.217263382430935e-07, |
|
"loss": 1.2613, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 22.5, |
|
"grad_norm": 1.9589574337005615, |
|
"learning_rate": 8.266062568876716e-07, |
|
"loss": 1.134, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"grad_norm": 1.7819244861602783, |
|
"learning_rate": 8.313789158407869e-07, |
|
"loss": 1.0627, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 23.5, |
|
"grad_norm": 2.12137508392334, |
|
"learning_rate": 8.360489289678585e-07, |
|
"loss": 1.1799, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"grad_norm": 1.7399541139602661, |
|
"learning_rate": 8.406206186877934e-07, |
|
"loss": 0.9974, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 24.5, |
|
"grad_norm": 2.0634093284606934, |
|
"learning_rate": 8.450980400142567e-07, |
|
"loss": 0.9758, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"grad_norm": 1.8642668724060059, |
|
"learning_rate": 8.494850021680092e-07, |
|
"loss": 1.1103, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 25.5, |
|
"grad_norm": 1.8973793983459473, |
|
"learning_rate": 8.53785088048968e-07, |
|
"loss": 1.1013, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"grad_norm": 1.8340080976486206, |
|
"learning_rate": 8.580016718173995e-07, |
|
"loss": 0.9959, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 26.5, |
|
"grad_norm": 1.8867207765579224, |
|
"learning_rate": 8.621379348003944e-07, |
|
"loss": 1.1526, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"grad_norm": 1.7188408374786377, |
|
"learning_rate": 8.661968799114842e-07, |
|
"loss": 0.9184, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 27.5, |
|
"grad_norm": 1.8246105909347534, |
|
"learning_rate": 8.701813447471218e-07, |
|
"loss": 1.0568, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"grad_norm": 1.9179924726486206, |
|
"learning_rate": 8.740940135031001e-07, |
|
"loss": 1.1564, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 28.5, |
|
"grad_norm": 1.7202187776565552, |
|
"learning_rate": 8.779374278362456e-07, |
|
"loss": 0.9801, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"grad_norm": 1.891327977180481, |
|
"learning_rate": 8.817139967814684e-07, |
|
"loss": 1.1344, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 29.5, |
|
"grad_norm": 1.6394891738891602, |
|
"learning_rate": 8.854260058210719e-07, |
|
"loss": 0.9748, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"grad_norm": 1.8808232545852661, |
|
"learning_rate": 8.890756251918216e-07, |
|
"loss": 0.9423, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 30.5, |
|
"grad_norm": 1.8192319869995117, |
|
"learning_rate": 8.926649175053833e-07, |
|
"loss": 1.003, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"grad_norm": 1.7621928453445435, |
|
"learning_rate": 8.961958447491268e-07, |
|
"loss": 1.0498, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 31.5, |
|
"grad_norm": 1.8957831859588623, |
|
"learning_rate": 8.996702747267907e-07, |
|
"loss": 1.1259, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"grad_norm": 1.6946312189102173, |
|
"learning_rate": 9.030899869919433e-07, |
|
"loss": 0.9092, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 32.5, |
|
"grad_norm": 1.6719154119491577, |
|
"learning_rate": 9.064566783214276e-07, |
|
"loss": 0.9273, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"grad_norm": 1.6526826620101929, |
|
"learning_rate": 9.097719677709341e-07, |
|
"loss": 0.9469, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 33.5, |
|
"grad_norm": 1.7876310348510742, |
|
"learning_rate": 9.13037401350413e-07, |
|
"loss": 0.8734, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"grad_norm": 1.517971396446228, |
|
"learning_rate": 9.162544563531181e-07, |
|
"loss": 0.9527, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 34.5, |
|
"grad_norm": 1.7323296070098877, |
|
"learning_rate": 9.194245453686276e-07, |
|
"loss": 0.8279, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"grad_norm": 1.5928230285644531, |
|
"learning_rate": 9.225490200071283e-07, |
|
"loss": 0.9627, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 35.5, |
|
"grad_norm": 1.6561415195465088, |
|
"learning_rate": 9.256291743595375e-07, |
|
"loss": 0.9138, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"grad_norm": 1.628674030303955, |
|
"learning_rate": 9.28666248215634e-07, |
|
"loss": 0.8956, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 36.5, |
|
"grad_norm": 1.5087110996246338, |
|
"learning_rate": 9.316614300602277e-07, |
|
"loss": 0.8223, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"grad_norm": 1.6503626108169556, |
|
"learning_rate": 9.346158598654879e-07, |
|
"loss": 0.9776, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 37.5, |
|
"grad_norm": 1.6680738925933838, |
|
"learning_rate": 9.375306316958498e-07, |
|
"loss": 0.8301, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"grad_norm": 1.8239458799362183, |
|
"learning_rate": 9.404067961403955e-07, |
|
"loss": 0.9891, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 38.5, |
|
"grad_norm": 1.669043779373169, |
|
"learning_rate": 9.432453625862408e-07, |
|
"loss": 0.9085, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"grad_norm": 1.8124521970748901, |
|
"learning_rate": 9.4604730134524e-07, |
|
"loss": 0.87, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 39.5, |
|
"grad_norm": 1.6593950986862183, |
|
"learning_rate": 9.488135456452205e-07, |
|
"loss": 0.8142, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"grad_norm": 1.6837782859802246, |
|
"learning_rate": 9.515449934959715e-07, |
|
"loss": 0.8246, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 40.5, |
|
"grad_norm": 1.7322601079940796, |
|
"learning_rate": 9.542425094393247e-07, |
|
"loss": 0.8752, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"grad_norm": 1.5245649814605713, |
|
"learning_rate": 9.569069261918583e-07, |
|
"loss": 0.7944, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 41.5, |
|
"grad_norm": 1.713905692100525, |
|
"learning_rate": 9.59539046188037e-07, |
|
"loss": 0.8122, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"grad_norm": 1.7115576267242432, |
|
"learning_rate": 9.621396430309406e-07, |
|
"loss": 0.8612, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 42.5, |
|
"grad_norm": 1.770555019378662, |
|
"learning_rate": 9.647094628571462e-07, |
|
"loss": 0.9131, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"grad_norm": 1.8751977682113647, |
|
"learning_rate": 9.672492256217836e-07, |
|
"loss": 0.8756, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 43.5, |
|
"grad_norm": 1.8345999717712402, |
|
"learning_rate": 9.69759626309309e-07, |
|
"loss": 0.9098, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"grad_norm": 1.6411185264587402, |
|
"learning_rate": 9.722413360750842e-07, |
|
"loss": 0.7975, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 44.5, |
|
"grad_norm": 1.744828462600708, |
|
"learning_rate": 9.74695003322456e-07, |
|
"loss": 0.8444, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"grad_norm": 1.6588188409805298, |
|
"learning_rate": 9.771212547196622e-07, |
|
"loss": 0.8376, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 45.5, |
|
"grad_norm": 1.8046928644180298, |
|
"learning_rate": 9.795206961605466e-07, |
|
"loss": 0.7546, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 46.0, |
|
"grad_norm": 1.8828351497650146, |
|
"learning_rate": 9.818939136727774e-07, |
|
"loss": 0.8505, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 46.5, |
|
"grad_norm": 1.841956377029419, |
|
"learning_rate": 9.842414742769674e-07, |
|
"loss": 0.7847, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 47.0, |
|
"grad_norm": 1.682256817817688, |
|
"learning_rate": 9.865639267998492e-07, |
|
"loss": 0.7712, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 47.5, |
|
"grad_norm": 1.8375487327575684, |
|
"learning_rate": 9.888618026444236e-07, |
|
"loss": 0.7228, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"grad_norm": 1.7198117971420288, |
|
"learning_rate": 9.91135616519784e-07, |
|
"loss": 0.7755, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 48.5, |
|
"grad_norm": 1.760452389717102, |
|
"learning_rate": 9.933858671331222e-07, |
|
"loss": 0.7045, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 49.0, |
|
"grad_norm": 1.735704779624939, |
|
"learning_rate": 9.956130378462473e-07, |
|
"loss": 0.8024, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 49.5, |
|
"grad_norm": 1.6422948837280273, |
|
"learning_rate": 9.978175972987748e-07, |
|
"loss": 0.7368, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"grad_norm": 1.8960306644439697, |
|
"learning_rate": 9.999999999999997e-07, |
|
"loss": 0.7557, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 50.5, |
|
"grad_norm": 1.6727304458618164, |
|
"learning_rate": 1e-06, |
|
"loss": 0.7325, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 51.0, |
|
"grad_norm": 1.6515084505081177, |
|
"learning_rate": 1e-06, |
|
"loss": 0.7351, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 51.5, |
|
"grad_norm": 1.7705847024917603, |
|
"learning_rate": 1e-06, |
|
"loss": 0.7907, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 52.0, |
|
"grad_norm": 1.7057950496673584, |
|
"learning_rate": 1e-06, |
|
"loss": 0.6447, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 52.5, |
|
"grad_norm": 1.6130571365356445, |
|
"learning_rate": 1e-06, |
|
"loss": 0.7079, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 53.0, |
|
"grad_norm": 2.063298463821411, |
|
"learning_rate": 1e-06, |
|
"loss": 0.693, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 53.5, |
|
"grad_norm": 2.0730509757995605, |
|
"learning_rate": 1e-06, |
|
"loss": 0.8002, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 54.0, |
|
"grad_norm": 1.6381713151931763, |
|
"learning_rate": 1e-06, |
|
"loss": 0.657, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 54.5, |
|
"grad_norm": 1.5659828186035156, |
|
"learning_rate": 1e-06, |
|
"loss": 0.7202, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 55.0, |
|
"grad_norm": 1.594575047492981, |
|
"learning_rate": 1e-06, |
|
"loss": 0.6627, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 55.5, |
|
"grad_norm": 1.497917652130127, |
|
"learning_rate": 1e-06, |
|
"loss": 0.6693, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 56.0, |
|
"grad_norm": 2.5011086463928223, |
|
"learning_rate": 1e-06, |
|
"loss": 0.6946, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 56.5, |
|
"grad_norm": 1.9602758884429932, |
|
"learning_rate": 1e-06, |
|
"loss": 0.6926, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 57.0, |
|
"grad_norm": 1.4793510437011719, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5861, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 57.5, |
|
"grad_norm": 1.6028777360916138, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5788, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 58.0, |
|
"grad_norm": 1.6478813886642456, |
|
"learning_rate": 1e-06, |
|
"loss": 0.6395, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 58.5, |
|
"grad_norm": 1.5423738956451416, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5959, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 59.0, |
|
"grad_norm": 1.8497169017791748, |
|
"learning_rate": 1e-06, |
|
"loss": 0.584, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 59.5, |
|
"grad_norm": 1.6440547704696655, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5916, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"grad_norm": 1.765620231628418, |
|
"learning_rate": 1e-06, |
|
"loss": 0.6226, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 60.5, |
|
"grad_norm": 1.543800950050354, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5966, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 61.0, |
|
"grad_norm": 1.4944016933441162, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5305, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 61.5, |
|
"grad_norm": 1.968621850013733, |
|
"learning_rate": 1e-06, |
|
"loss": 0.6673, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 62.0, |
|
"grad_norm": 1.523604393005371, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5424, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 62.5, |
|
"grad_norm": 1.6466797590255737, |
|
"learning_rate": 1e-06, |
|
"loss": 0.6007, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 63.0, |
|
"grad_norm": 1.7836798429489136, |
|
"learning_rate": 1e-06, |
|
"loss": 0.6201, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 63.5, |
|
"grad_norm": 1.6673424243927002, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5786, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 64.0, |
|
"grad_norm": 1.6889145374298096, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5211, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 64.5, |
|
"grad_norm": 1.4834386110305786, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4521, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 65.0, |
|
"grad_norm": 1.743851661682129, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5363, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 65.5, |
|
"grad_norm": 1.8134723901748657, |
|
"learning_rate": 1e-06, |
|
"loss": 0.554, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 66.0, |
|
"grad_norm": 1.508358120918274, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5104, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 66.5, |
|
"grad_norm": 1.6829733848571777, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4658, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 67.0, |
|
"grad_norm": 1.526950716972351, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4892, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 67.5, |
|
"grad_norm": 1.8935024738311768, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4979, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 68.0, |
|
"grad_norm": 1.4638999700546265, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5144, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 68.5, |
|
"grad_norm": 1.9910645484924316, |
|
"learning_rate": 1e-06, |
|
"loss": 0.5521, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 69.0, |
|
"grad_norm": 1.6257317066192627, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4937, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 69.5, |
|
"grad_norm": 1.4498651027679443, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4749, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 70.0, |
|
"grad_norm": 1.8104501962661743, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4564, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 70.5, |
|
"grad_norm": 2.0244479179382324, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4228, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 71.0, |
|
"grad_norm": 1.5190598964691162, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4735, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 71.5, |
|
"grad_norm": 1.7180043458938599, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4776, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 72.0, |
|
"grad_norm": 1.5680577754974365, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4343, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 72.5, |
|
"grad_norm": 1.6798756122589111, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4074, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 73.0, |
|
"grad_norm": 1.4644179344177246, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4982, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 73.5, |
|
"grad_norm": 1.5461561679840088, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3539, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 74.0, |
|
"grad_norm": 1.7116854190826416, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4267, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 74.5, |
|
"grad_norm": 1.6357485055923462, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4563, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 75.0, |
|
"grad_norm": 1.3843780755996704, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4072, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 75.5, |
|
"grad_norm": 1.6510047912597656, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4619, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 76.0, |
|
"grad_norm": 1.5008376836776733, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3768, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 76.5, |
|
"grad_norm": 1.4433045387268066, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4502, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 77.0, |
|
"grad_norm": 1.4826611280441284, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3686, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 77.5, |
|
"grad_norm": 1.5890164375305176, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3642, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 78.0, |
|
"grad_norm": 1.5281238555908203, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4034, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 78.5, |
|
"grad_norm": 1.3185185194015503, |
|
"learning_rate": 1e-06, |
|
"loss": 0.336, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 79.0, |
|
"grad_norm": 1.6037932634353638, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4349, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 79.5, |
|
"grad_norm": 1.384059190750122, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3431, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"grad_norm": 1.339905858039856, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3745, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 80.5, |
|
"grad_norm": 1.2671548128128052, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3476, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 81.0, |
|
"grad_norm": 1.5032880306243896, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3297, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 81.5, |
|
"grad_norm": 1.432960033416748, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3829, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 82.0, |
|
"grad_norm": 1.785606026649475, |
|
"learning_rate": 1e-06, |
|
"loss": 0.4002, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 82.5, |
|
"grad_norm": 1.599700927734375, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3272, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 83.0, |
|
"grad_norm": 1.3606281280517578, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3462, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 83.5, |
|
"grad_norm": 1.311733603477478, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3459, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 84.0, |
|
"grad_norm": 1.577045202255249, |
|
"learning_rate": 1e-06, |
|
"loss": 0.278, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 84.5, |
|
"grad_norm": 1.5641367435455322, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3636, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 85.0, |
|
"grad_norm": 1.2674757242202759, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3074, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 85.5, |
|
"grad_norm": 1.423398494720459, |
|
"learning_rate": 1e-06, |
|
"loss": 0.32, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 86.0, |
|
"grad_norm": 1.149396538734436, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2577, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 86.5, |
|
"grad_norm": 1.687155842781067, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2998, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 87.0, |
|
"grad_norm": 1.0938485860824585, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2963, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 87.5, |
|
"grad_norm": 1.2464781999588013, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2691, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 88.0, |
|
"grad_norm": 1.259631633758545, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2815, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 88.5, |
|
"grad_norm": 1.2384026050567627, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1886, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 89.0, |
|
"grad_norm": 1.209479808807373, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3283, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 89.5, |
|
"grad_norm": 1.2056385278701782, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2795, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 90.0, |
|
"grad_norm": 1.256142258644104, |
|
"learning_rate": 1e-06, |
|
"loss": 0.3071, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 90.5, |
|
"grad_norm": 1.2020200490951538, |
|
"learning_rate": 1e-06, |
|
"loss": 0.242, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 91.0, |
|
"grad_norm": 1.275436520576477, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2505, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 91.5, |
|
"grad_norm": 1.1096285581588745, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2833, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 92.0, |
|
"grad_norm": 1.0823484659194946, |
|
"learning_rate": 1e-06, |
|
"loss": 0.216, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 92.5, |
|
"grad_norm": 1.112586498260498, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2292, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 93.0, |
|
"grad_norm": 1.004947543144226, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2399, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 93.5, |
|
"grad_norm": 1.10011887550354, |
|
"learning_rate": 1e-06, |
|
"loss": 0.263, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 94.0, |
|
"grad_norm": 0.9535015821456909, |
|
"learning_rate": 1e-06, |
|
"loss": 0.218, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 94.5, |
|
"grad_norm": 1.0121976137161255, |
|
"learning_rate": 1e-06, |
|
"loss": 0.191, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 95.0, |
|
"grad_norm": 0.9026556611061096, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2278, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 95.5, |
|
"grad_norm": 0.9730249643325806, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2422, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 96.0, |
|
"grad_norm": 0.9288642406463623, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2353, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 96.5, |
|
"grad_norm": 0.8509739637374878, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2292, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 97.0, |
|
"grad_norm": 0.9947998523712158, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2309, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 97.5, |
|
"grad_norm": 1.109282374382019, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2369, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 98.0, |
|
"grad_norm": 0.8555991053581238, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2011, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 98.5, |
|
"grad_norm": 0.9674638509750366, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2385, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 99.0, |
|
"grad_norm": 0.781050443649292, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1881, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 99.5, |
|
"grad_norm": 0.8599874377250671, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2031, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"grad_norm": 0.8711087703704834, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2214, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 100.5, |
|
"grad_norm": 0.9213354587554932, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2313, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 101.0, |
|
"grad_norm": 0.871462345123291, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1978, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 101.5, |
|
"grad_norm": 0.7935155630111694, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1873, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 102.0, |
|
"grad_norm": 0.9139618277549744, |
|
"learning_rate": 1e-06, |
|
"loss": 0.2283, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 102.5, |
|
"grad_norm": 0.8635255694389343, |
|
"learning_rate": 1e-06, |
|
"loss": 0.228, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 103.0, |
|
"grad_norm": 0.9213907122612, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1837, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 103.5, |
|
"grad_norm": 0.7787233591079712, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1652, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 104.0, |
|
"grad_norm": 0.8260976076126099, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1986, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 104.5, |
|
"grad_norm": 0.8949348330497742, |
|
"learning_rate": 1e-06, |
|
"loss": 0.172, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 105.0, |
|
"grad_norm": 0.8772971630096436, |
|
"learning_rate": 1e-06, |
|
"loss": 0.201, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 105.5, |
|
"grad_norm": 0.7942510843276978, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1754, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 106.0, |
|
"grad_norm": 0.8099932670593262, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1586, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 106.5, |
|
"grad_norm": 0.880547285079956, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1516, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 107.0, |
|
"grad_norm": 0.8132925033569336, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1657, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 107.5, |
|
"grad_norm": 0.8455451726913452, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1994, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 108.0, |
|
"grad_norm": 0.9202403426170349, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1486, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 108.5, |
|
"grad_norm": 0.8958231806755066, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1949, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 109.0, |
|
"grad_norm": 0.8252700567245483, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1645, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 109.5, |
|
"grad_norm": 0.796977698802948, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1297, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 110.0, |
|
"grad_norm": 0.8288230895996094, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1967, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 110.5, |
|
"grad_norm": 0.9239948987960815, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1546, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 111.0, |
|
"grad_norm": 0.8271680474281311, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1748, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 111.5, |
|
"grad_norm": 0.7675459980964661, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1262, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 112.0, |
|
"grad_norm": 0.7924964427947998, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1434, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 112.5, |
|
"grad_norm": 0.9103841185569763, |
|
"learning_rate": 1e-06, |
|
"loss": 0.186, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 113.0, |
|
"grad_norm": 0.8457869291305542, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1384, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 113.5, |
|
"grad_norm": 0.8478854298591614, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1743, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 114.0, |
|
"grad_norm": 0.8645926713943481, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1514, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 114.5, |
|
"grad_norm": 0.9108607769012451, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1759, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 115.0, |
|
"grad_norm": 0.8300096392631531, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1375, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 115.5, |
|
"grad_norm": 0.9206691384315491, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1369, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 116.0, |
|
"grad_norm": 0.7558128833770752, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1242, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 116.5, |
|
"grad_norm": 0.8597300052642822, |
|
"learning_rate": 1e-06, |
|
"loss": 0.128, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 117.0, |
|
"grad_norm": 0.8134746551513672, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1547, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 117.5, |
|
"grad_norm": 0.9657474160194397, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1534, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 118.0, |
|
"grad_norm": 0.7481112480163574, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1282, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 118.5, |
|
"grad_norm": 0.6953885555267334, |
|
"learning_rate": 1e-06, |
|
"loss": 0.0988, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 119.0, |
|
"grad_norm": 0.8225458860397339, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1135, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 119.5, |
|
"grad_norm": 0.7915026545524597, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1043, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 120.0, |
|
"grad_norm": 0.7963205575942993, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1474, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 120.5, |
|
"grad_norm": 0.796559751033783, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1295, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 121.0, |
|
"grad_norm": 0.8174726963043213, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1305, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 121.5, |
|
"grad_norm": 0.7992000579833984, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1078, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 122.0, |
|
"grad_norm": 0.7742059826850891, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1082, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 122.5, |
|
"grad_norm": 0.7738575339317322, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1075, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 123.0, |
|
"grad_norm": 0.7644574642181396, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1228, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 123.5, |
|
"grad_norm": 0.7060183882713318, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1115, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 124.0, |
|
"grad_norm": 0.7386205196380615, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1181, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 124.5, |
|
"grad_norm": 0.7068957090377808, |
|
"learning_rate": 1e-06, |
|
"loss": 0.1156, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 125.0, |
|
"grad_norm": 0.7993431687355042, |
|
"learning_rate": 1e-06, |
|
"loss": 0.088, |
|
"step": 250 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 10000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5000, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.317943074155397e+17, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|