{ "best_metric": null, "best_model_checkpoint": null, "epoch": 75.0, "eval_steps": 1000, "global_step": 150, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.5, "grad_norm": 0.0, "learning_rate": 0, "loss": 1.3402, "step": 1 }, { "epoch": 1.0, "grad_norm": 0.0, "learning_rate": 0, "loss": 1.286, "step": 2 }, { "epoch": 1.5, "grad_norm": 0.0, "learning_rate": 0, "loss": 1.3382, "step": 3 }, { "epoch": 2.0, "grad_norm": 6.078482627868652, "learning_rate": 0.0, "loss": 1.2555, "step": 4 }, { "epoch": 2.5, "grad_norm": 5.590641975402832, "learning_rate": 1.5051499783199055e-07, "loss": 1.2272, "step": 5 }, { "epoch": 3.0, "grad_norm": 5.590641975402832, "learning_rate": 1.5051499783199055e-07, "loss": 1.3339, "step": 6 }, { "epoch": 3.5, "grad_norm": 5.590641975402832, "learning_rate": 1.5051499783199055e-07, "loss": 1.4161, "step": 7 }, { "epoch": 4.0, "grad_norm": 11.268411636352539, "learning_rate": 2.385606273598312e-07, "loss": 1.362, "step": 8 }, { "epoch": 4.5, "grad_norm": 9.850939750671387, "learning_rate": 3.010299956639811e-07, "loss": 1.1579, "step": 9 }, { "epoch": 5.0, "grad_norm": 11.71530532836914, "learning_rate": 3.494850021680093e-07, "loss": 1.4335, "step": 10 }, { "epoch": 5.5, "grad_norm": 5.690868377685547, "learning_rate": 3.8907562519182173e-07, "loss": 1.2658, "step": 11 }, { "epoch": 6.0, "grad_norm": 10.700072288513184, "learning_rate": 4.2254902000712834e-07, "loss": 1.3482, "step": 12 }, { "epoch": 6.5, "grad_norm": 6.189509391784668, "learning_rate": 4.5154499349597166e-07, "loss": 1.3464, "step": 13 }, { "epoch": 7.0, "grad_norm": 6.247044086456299, "learning_rate": 4.771212547196623e-07, "loss": 1.2738, "step": 14 }, { "epoch": 7.5, "grad_norm": 6.324206829071045, "learning_rate": 4.999999999999999e-07, "loss": 1.2795, "step": 15 }, { "epoch": 8.0, "grad_norm": 5.119565486907959, "learning_rate": 5.206963425791124e-07, "loss": 1.281, "step": 16 }, { "epoch": 8.5, "grad_norm": 5.067975044250488, "learning_rate": 5.395906230238123e-07, "loss": 1.4589, "step": 17 }, { "epoch": 9.0, "grad_norm": 4.973406791687012, "learning_rate": 5.569716761534182e-07, "loss": 1.293, "step": 18 }, { "epoch": 9.5, "grad_norm": 4.727417945861816, "learning_rate": 5.730640178391189e-07, "loss": 1.2502, "step": 19 }, { "epoch": 10.0, "grad_norm": 4.404863357543945, "learning_rate": 5.880456295278405e-07, "loss": 1.2186, "step": 20 }, { "epoch": 10.5, "grad_norm": 6.084046840667725, "learning_rate": 6.020599913279622e-07, "loss": 1.4264, "step": 21 }, { "epoch": 11.0, "grad_norm": 4.772565841674805, "learning_rate": 6.15224460689137e-07, "loss": 1.2514, "step": 22 }, { "epoch": 11.5, "grad_norm": 4.340854167938232, "learning_rate": 6.276362525516529e-07, "loss": 1.2459, "step": 23 }, { "epoch": 12.0, "grad_norm": 3.630235433578491, "learning_rate": 6.393768004764143e-07, "loss": 1.0926, "step": 24 }, { "epoch": 12.5, "grad_norm": 7.564579963684082, "learning_rate": 6.505149978319905e-07, "loss": 1.3692, "step": 25 }, { "epoch": 13.0, "grad_norm": 4.304481029510498, "learning_rate": 6.611096473669595e-07, "loss": 1.0156, "step": 26 }, { "epoch": 13.5, "grad_norm": 5.493950366973877, "learning_rate": 6.712113404111031e-07, "loss": 1.2072, "step": 27 }, { "epoch": 14.0, "grad_norm": 3.82694411277771, "learning_rate": 6.808639180087963e-07, "loss": 1.0592, "step": 28 }, { "epoch": 14.5, "grad_norm": 3.7888131141662598, "learning_rate": 6.901056208558029e-07, "loss": 1.1043, "step": 29 }, { "epoch": 15.0, "grad_norm": 3.5746841430664062, "learning_rate": 6.989700043360186e-07, "loss": 1.1532, "step": 30 }, { "epoch": 15.5, "grad_norm": 4.591047286987305, "learning_rate": 7.074866739854088e-07, "loss": 1.2519, "step": 31 }, { "epoch": 16.0, "grad_norm": 10.526527404785156, "learning_rate": 7.156818820794935e-07, "loss": 1.1504, "step": 32 }, { "epoch": 16.5, "grad_norm": 4.354953289031982, "learning_rate": 7.235790156711094e-07, "loss": 1.098, "step": 33 }, { "epoch": 17.0, "grad_norm": 5.967930316925049, "learning_rate": 7.311989989494779e-07, "loss": 1.1957, "step": 34 }, { "epoch": 17.5, "grad_norm": 3.097174644470215, "learning_rate": 7.38560627359831e-07, "loss": 1.073, "step": 35 }, { "epoch": 18.0, "grad_norm": 3.9502739906311035, "learning_rate": 7.456808469171361e-07, "loss": 1.0042, "step": 36 }, { "epoch": 18.5, "grad_norm": 3.2742667198181152, "learning_rate": 7.525749891599529e-07, "loss": 1.004, "step": 37 }, { "epoch": 19.0, "grad_norm": 4.2321953773498535, "learning_rate": 7.592569699389436e-07, "loss": 1.0532, "step": 38 }, { "epoch": 19.5, "grad_norm": 4.322045803070068, "learning_rate": 7.657394585211274e-07, "loss": 1.0297, "step": 39 }, { "epoch": 20.0, "grad_norm": 3.5067355632781982, "learning_rate": 7.720340221751376e-07, "loss": 1.0541, "step": 40 }, { "epoch": 20.5, "grad_norm": 3.1070497035980225, "learning_rate": 7.781512503836435e-07, "loss": 0.8917, "step": 41 }, { "epoch": 21.0, "grad_norm": 3.413217067718506, "learning_rate": 7.841008620334974e-07, "loss": 1.0688, "step": 42 }, { "epoch": 21.5, "grad_norm": 3.413217067718506, "learning_rate": 7.841008620334974e-07, "loss": 0.8956, "step": 43 }, { "epoch": 22.0, "grad_norm": 31.582998275756836, "learning_rate": 7.89891798308405e-07, "loss": 1.1435, "step": 44 }, { "epoch": 22.5, "grad_norm": 3.2715117931365967, "learning_rate": 7.955323035132494e-07, "loss": 1.023, "step": 45 }, { "epoch": 23.0, "grad_norm": 3.081068754196167, "learning_rate": 8.01029995663981e-07, "loss": 0.9536, "step": 46 }, { "epoch": 23.5, "grad_norm": 5.033097267150879, "learning_rate": 8.063919283598676e-07, "loss": 1.0542, "step": 47 }, { "epoch": 24.0, "grad_norm": 3.2547528743743896, "learning_rate": 8.116246451989502e-07, "loss": 0.8736, "step": 48 }, { "epoch": 24.5, "grad_norm": 3.2148780822753906, "learning_rate": 8.16734227789793e-07, "loss": 0.8366, "step": 49 }, { "epoch": 25.0, "grad_norm": 3.271756887435913, "learning_rate": 8.217263382430935e-07, "loss": 0.9887, "step": 50 }, { "epoch": 25.5, "grad_norm": 3.1897552013397217, "learning_rate": 8.266062568876716e-07, "loss": 0.9634, "step": 51 }, { "epoch": 26.0, "grad_norm": 2.999152421951294, "learning_rate": 8.313789158407869e-07, "loss": 0.8479, "step": 52 }, { "epoch": 26.5, "grad_norm": 3.126117467880249, "learning_rate": 8.360489289678585e-07, "loss": 0.9963, "step": 53 }, { "epoch": 27.0, "grad_norm": 2.808807611465454, "learning_rate": 8.406206186877934e-07, "loss": 0.7693, "step": 54 }, { "epoch": 27.5, "grad_norm": 3.8470795154571533, "learning_rate": 8.450980400142567e-07, "loss": 0.8929, "step": 55 }, { "epoch": 28.0, "grad_norm": 4.059876918792725, "learning_rate": 8.494850021680092e-07, "loss": 0.9874, "step": 56 }, { "epoch": 28.5, "grad_norm": 3.1297614574432373, "learning_rate": 8.53785088048968e-07, "loss": 0.8148, "step": 57 }, { "epoch": 29.0, "grad_norm": 4.2221150398254395, "learning_rate": 8.580016718173995e-07, "loss": 0.9519, "step": 58 }, { "epoch": 29.5, "grad_norm": 3.0946404933929443, "learning_rate": 8.621379348003944e-07, "loss": 0.8067, "step": 59 }, { "epoch": 30.0, "grad_norm": 3.16498064994812, "learning_rate": 8.661968799114842e-07, "loss": 0.759, "step": 60 }, { "epoch": 30.5, "grad_norm": 3.2981464862823486, "learning_rate": 8.701813447471218e-07, "loss": 0.8196, "step": 61 }, { "epoch": 31.0, "grad_norm": 2.9492411613464355, "learning_rate": 8.740940135031001e-07, "loss": 0.8733, "step": 62 }, { "epoch": 31.5, "grad_norm": 4.684609413146973, "learning_rate": 8.779374278362456e-07, "loss": 0.9314, "step": 63 }, { "epoch": 32.0, "grad_norm": 2.6447761058807373, "learning_rate": 8.817139967814684e-07, "loss": 0.7206, "step": 64 }, { "epoch": 32.5, "grad_norm": 3.544161319732666, "learning_rate": 8.854260058210719e-07, "loss": 0.7301, "step": 65 }, { "epoch": 33.0, "grad_norm": 2.7136151790618896, "learning_rate": 8.890756251918216e-07, "loss": 0.7486, "step": 66 }, { "epoch": 33.5, "grad_norm": 3.6095616817474365, "learning_rate": 8.926649175053833e-07, "loss": 0.6688, "step": 67 }, { "epoch": 34.0, "grad_norm": 2.742135763168335, "learning_rate": 8.961958447491268e-07, "loss": 0.7577, "step": 68 }, { "epoch": 34.5, "grad_norm": 3.166233777999878, "learning_rate": 8.996702747267907e-07, "loss": 0.6201, "step": 69 }, { "epoch": 35.0, "grad_norm": 6.629209518432617, "learning_rate": 9.030899869919433e-07, "loss": 0.7546, "step": 70 }, { "epoch": 35.5, "grad_norm": 3.2534568309783936, "learning_rate": 9.064566783214276e-07, "loss": 0.6966, "step": 71 }, { "epoch": 36.0, "grad_norm": 2.9836554527282715, "learning_rate": 9.097719677709341e-07, "loss": 0.669, "step": 72 }, { "epoch": 36.5, "grad_norm": 2.8371522426605225, "learning_rate": 9.13037401350413e-07, "loss": 0.602, "step": 73 }, { "epoch": 37.0, "grad_norm": 3.1005606651306152, "learning_rate": 9.162544563531181e-07, "loss": 0.7351, "step": 74 }, { "epoch": 37.5, "grad_norm": 2.771270751953125, "learning_rate": 9.194245453686276e-07, "loss": 0.5892, "step": 75 }, { "epoch": 38.0, "grad_norm": 4.207065105438232, "learning_rate": 9.225490200071283e-07, "loss": 0.7159, "step": 76 }, { "epoch": 38.5, "grad_norm": 4.0643439292907715, "learning_rate": 9.256291743595375e-07, "loss": 0.6424, "step": 77 }, { "epoch": 39.0, "grad_norm": 3.463486433029175, "learning_rate": 9.28666248215634e-07, "loss": 0.6164, "step": 78 }, { "epoch": 39.5, "grad_norm": 3.14841890335083, "learning_rate": 9.316614300602277e-07, "loss": 0.5615, "step": 79 }, { "epoch": 40.0, "grad_norm": 2.958975076675415, "learning_rate": 9.346158598654879e-07, "loss": 0.5521, "step": 80 }, { "epoch": 40.5, "grad_norm": 2.7486228942871094, "learning_rate": 9.375306316958498e-07, "loss": 0.5937, "step": 81 }, { "epoch": 41.0, "grad_norm": 2.736597776412964, "learning_rate": 9.404067961403955e-07, "loss": 0.5238, "step": 82 }, { "epoch": 41.5, "grad_norm": 2.9119350910186768, "learning_rate": 9.432453625862408e-07, "loss": 0.5296, "step": 83 }, { "epoch": 42.0, "grad_norm": 3.870023727416992, "learning_rate": 9.4604730134524e-07, "loss": 0.5716, "step": 84 }, { "epoch": 42.5, "grad_norm": 3.2765142917633057, "learning_rate": 9.488135456452205e-07, "loss": 0.597, "step": 85 }, { "epoch": 43.0, "grad_norm": 3.575230598449707, "learning_rate": 9.515449934959715e-07, "loss": 0.5601, "step": 86 }, { "epoch": 43.5, "grad_norm": 3.2691493034362793, "learning_rate": 9.542425094393247e-07, "loss": 0.5662, "step": 87 }, { "epoch": 44.0, "grad_norm": 2.8934061527252197, "learning_rate": 9.569069261918583e-07, "loss": 0.5026, "step": 88 }, { "epoch": 44.5, "grad_norm": 3.517810821533203, "learning_rate": 9.59539046188037e-07, "loss": 0.5107, "step": 89 }, { "epoch": 45.0, "grad_norm": 3.1376760005950928, "learning_rate": 9.621396430309406e-07, "loss": 0.5166, "step": 90 }, { "epoch": 45.5, "grad_norm": 3.4940109252929688, "learning_rate": 9.647094628571462e-07, "loss": 0.4549, "step": 91 }, { "epoch": 46.0, "grad_norm": 6.771224498748779, "learning_rate": 9.672492256217836e-07, "loss": 0.4991, "step": 92 }, { "epoch": 46.5, "grad_norm": 4.691510200500488, "learning_rate": 9.69759626309309e-07, "loss": 0.4521, "step": 93 }, { "epoch": 47.0, "grad_norm": 3.1397156715393066, "learning_rate": 9.722413360750842e-07, "loss": 0.4287, "step": 94 }, { "epoch": 47.5, "grad_norm": 3.4504847526550293, "learning_rate": 9.74695003322456e-07, "loss": 0.3875, "step": 95 }, { "epoch": 48.0, "grad_norm": 2.7338361740112305, "learning_rate": 9.771212547196622e-07, "loss": 0.44, "step": 96 }, { "epoch": 48.5, "grad_norm": 4.007640838623047, "learning_rate": 9.795206961605466e-07, "loss": 0.3561, "step": 97 }, { "epoch": 49.0, "grad_norm": 3.637798309326172, "learning_rate": 9.818939136727774e-07, "loss": 0.4562, "step": 98 }, { "epoch": 49.5, "grad_norm": 4.988349914550781, "learning_rate": 9.842414742769674e-07, "loss": 0.4038, "step": 99 }, { "epoch": 50.0, "grad_norm": 3.512204885482788, "learning_rate": 9.865639267998492e-07, "loss": 0.3882, "step": 100 }, { "epoch": 50.5, "grad_norm": 2.9533920288085938, "learning_rate": 9.888618026444236e-07, "loss": 0.3843, "step": 101 }, { "epoch": 51.0, "grad_norm": 2.7523856163024902, "learning_rate": 9.91135616519784e-07, "loss": 0.3811, "step": 102 }, { "epoch": 51.5, "grad_norm": 6.668997764587402, "learning_rate": 9.933858671331222e-07, "loss": 0.4223, "step": 103 }, { "epoch": 52.0, "grad_norm": 3.053255319595337, "learning_rate": 9.956130378462473e-07, "loss": 0.3066, "step": 104 }, { "epoch": 52.5, "grad_norm": 3.2109737396240234, "learning_rate": 9.978175972987748e-07, "loss": 0.3446, "step": 105 }, { "epoch": 53.0, "grad_norm": 2.961881399154663, "learning_rate": 9.999999999999997e-07, "loss": 0.2935, "step": 106 }, { "epoch": 53.5, "grad_norm": 3.6308419704437256, "learning_rate": 1e-06, "loss": 0.3619, "step": 107 }, { "epoch": 54.0, "grad_norm": 3.809734582901001, "learning_rate": 1e-06, "loss": 0.3063, "step": 108 }, { "epoch": 54.5, "grad_norm": 2.6650562286376953, "learning_rate": 1e-06, "loss": 0.3504, "step": 109 }, { "epoch": 55.0, "grad_norm": 3.361117362976074, "learning_rate": 1e-06, "loss": 0.3084, "step": 110 }, { "epoch": 55.5, "grad_norm": 2.6281044483184814, "learning_rate": 1e-06, "loss": 0.3191, "step": 111 }, { "epoch": 56.0, "grad_norm": 3.0685670375823975, "learning_rate": 1e-06, "loss": 0.2472, "step": 112 }, { "epoch": 56.5, "grad_norm": 3.149151563644409, "learning_rate": 1e-06, "loss": 0.255, "step": 113 }, { "epoch": 57.0, "grad_norm": 3.0701425075531006, "learning_rate": 1e-06, "loss": 0.2392, "step": 114 }, { "epoch": 57.5, "grad_norm": 2.449301242828369, "learning_rate": 1e-06, "loss": 0.2239, "step": 115 }, { "epoch": 58.0, "grad_norm": 2.5579042434692383, "learning_rate": 1e-06, "loss": 0.265, "step": 116 }, { "epoch": 58.5, "grad_norm": 3.4325478076934814, "learning_rate": 1e-06, "loss": 0.2522, "step": 117 }, { "epoch": 59.0, "grad_norm": 3.4678075313568115, "learning_rate": 1e-06, "loss": 0.1957, "step": 118 }, { "epoch": 59.5, "grad_norm": 3.3002102375030518, "learning_rate": 1e-06, "loss": 0.2152, "step": 119 }, { "epoch": 60.0, "grad_norm": 2.939915180206299, "learning_rate": 1e-06, "loss": 0.228, "step": 120 }, { "epoch": 60.5, "grad_norm": 2.763664484024048, "learning_rate": 1e-06, "loss": 0.2261, "step": 121 }, { "epoch": 61.0, "grad_norm": 2.337585210800171, "learning_rate": 1e-06, "loss": 0.1851, "step": 122 }, { "epoch": 61.5, "grad_norm": 2.545644760131836, "learning_rate": 1e-06, "loss": 0.2182, "step": 123 }, { "epoch": 62.0, "grad_norm": 2.564753770828247, "learning_rate": 1e-06, "loss": 0.1962, "step": 124 }, { "epoch": 62.5, "grad_norm": 2.338416814804077, "learning_rate": 1e-06, "loss": 0.2051, "step": 125 }, { "epoch": 63.0, "grad_norm": 2.154878854751587, "learning_rate": 1e-06, "loss": 0.1879, "step": 126 }, { "epoch": 63.5, "grad_norm": 3.405646800994873, "learning_rate": 1e-06, "loss": 0.1655, "step": 127 }, { "epoch": 64.0, "grad_norm": 1.8845716714859009, "learning_rate": 1e-06, "loss": 0.161, "step": 128 }, { "epoch": 64.5, "grad_norm": 1.707482099533081, "learning_rate": 1e-06, "loss": 0.1342, "step": 129 }, { "epoch": 65.0, "grad_norm": 3.2093379497528076, "learning_rate": 1e-06, "loss": 0.1416, "step": 130 }, { "epoch": 65.5, "grad_norm": 3.438514232635498, "learning_rate": 1e-06, "loss": 0.1486, "step": 131 }, { "epoch": 66.0, "grad_norm": 1.9615154266357422, "learning_rate": 1e-06, "loss": 0.1461, "step": 132 }, { "epoch": 66.5, "grad_norm": 2.3800694942474365, "learning_rate": 1e-06, "loss": 0.1133, "step": 133 }, { "epoch": 67.0, "grad_norm": 2.241901159286499, "learning_rate": 1e-06, "loss": 0.1317, "step": 134 }, { "epoch": 67.5, "grad_norm": 2.024240255355835, "learning_rate": 1e-06, "loss": 0.1195, "step": 135 }, { "epoch": 68.0, "grad_norm": 2.2146880626678467, "learning_rate": 1e-06, "loss": 0.1413, "step": 136 }, { "epoch": 68.5, "grad_norm": 2.4345452785491943, "learning_rate": 1e-06, "loss": 0.1313, "step": 137 }, { "epoch": 69.0, "grad_norm": 3.0751936435699463, "learning_rate": 1e-06, "loss": 0.1231, "step": 138 }, { "epoch": 69.5, "grad_norm": 2.2487804889678955, "learning_rate": 1e-06, "loss": 0.1131, "step": 139 }, { "epoch": 70.0, "grad_norm": 3.5457677841186523, "learning_rate": 1e-06, "loss": 0.0812, "step": 140 }, { "epoch": 70.5, "grad_norm": 1.4189667701721191, "learning_rate": 1e-06, "loss": 0.0574, "step": 141 }, { "epoch": 71.0, "grad_norm": 2.4439244270324707, "learning_rate": 1e-06, "loss": 0.1146, "step": 142 }, { "epoch": 71.5, "grad_norm": 1.7094790935516357, "learning_rate": 1e-06, "loss": 0.0937, "step": 143 }, { "epoch": 72.0, "grad_norm": 1.4799623489379883, "learning_rate": 1e-06, "loss": 0.085, "step": 144 }, { "epoch": 72.5, "grad_norm": 1.4534505605697632, "learning_rate": 1e-06, "loss": 0.074, "step": 145 }, { "epoch": 73.0, "grad_norm": 1.7459020614624023, "learning_rate": 1e-06, "loss": 0.1062, "step": 146 }, { "epoch": 73.5, "grad_norm": 1.3827319145202637, "learning_rate": 1e-06, "loss": 0.0597, "step": 147 }, { "epoch": 74.0, "grad_norm": 1.4655884504318237, "learning_rate": 1e-06, "loss": 0.0749, "step": 148 }, { "epoch": 74.5, "grad_norm": 1.6126585006713867, "learning_rate": 1e-06, "loss": 0.0812, "step": 149 }, { "epoch": 75.0, "grad_norm": 1.4007526636123657, "learning_rate": 1e-06, "loss": 0.0724, "step": 150 } ], "logging_steps": 1.0, "max_steps": 10000, "num_input_tokens_seen": 0, "num_train_epochs": 5000, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.098224817063854e+17, "train_batch_size": 4, "trial_name": null, "trial_params": null }