{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9998107374929026, "eval_steps": 500, "global_step": 3962, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0025235000946312535, "grad_norm": 1.4534096717834473, "learning_rate": 1.2594458438287156e-06, "loss": 0.9056, "step": 10 }, { "epoch": 0.005047000189262507, "grad_norm": 0.9612669944763184, "learning_rate": 2.518891687657431e-06, "loss": 0.8732, "step": 20 }, { "epoch": 0.007570500283893761, "grad_norm": 0.7495620846748352, "learning_rate": 3.7783375314861467e-06, "loss": 0.8402, "step": 30 }, { "epoch": 0.010094000378525014, "grad_norm": 0.765338659286499, "learning_rate": 5.037783375314862e-06, "loss": 0.8493, "step": 40 }, { "epoch": 0.012617500473156268, "grad_norm": 0.5078374147415161, "learning_rate": 6.297229219143577e-06, "loss": 0.79, "step": 50 }, { "epoch": 0.015141000567787522, "grad_norm": 0.5078111290931702, "learning_rate": 7.556675062972293e-06, "loss": 0.8135, "step": 60 }, { "epoch": 0.017664500662418776, "grad_norm": 0.5493185520172119, "learning_rate": 8.816120906801008e-06, "loss": 0.8141, "step": 70 }, { "epoch": 0.020188000757050028, "grad_norm": 0.5199134945869446, "learning_rate": 1.0075566750629725e-05, "loss": 0.8042, "step": 80 }, { "epoch": 0.022711500851681284, "grad_norm": 0.5342094898223877, "learning_rate": 1.133501259445844e-05, "loss": 0.7793, "step": 90 }, { "epoch": 0.025235000946312536, "grad_norm": 0.4974937438964844, "learning_rate": 1.2594458438287154e-05, "loss": 0.7755, "step": 100 }, { "epoch": 0.027758501040943788, "grad_norm": 0.6357698440551758, "learning_rate": 1.385390428211587e-05, "loss": 0.8445, "step": 110 }, { "epoch": 0.030282001135575044, "grad_norm": 0.42305174469947815, "learning_rate": 1.5113350125944587e-05, "loss": 0.8141, "step": 120 }, { "epoch": 0.0328055012302063, "grad_norm": 0.5842427611351013, "learning_rate": 1.63727959697733e-05, "loss": 0.778, "step": 130 }, { "epoch": 0.03532900132483755, "grad_norm": 0.5549261569976807, "learning_rate": 1.7632241813602016e-05, "loss": 0.7565, "step": 140 }, { "epoch": 0.037852501419468804, "grad_norm": 0.4934644401073456, "learning_rate": 1.8891687657430733e-05, "loss": 0.8119, "step": 150 }, { "epoch": 0.040376001514100056, "grad_norm": 0.5441620945930481, "learning_rate": 2.015113350125945e-05, "loss": 0.799, "step": 160 }, { "epoch": 0.04289950160873131, "grad_norm": 0.4513319730758667, "learning_rate": 2.1410579345088162e-05, "loss": 0.7748, "step": 170 }, { "epoch": 0.04542300170336257, "grad_norm": 0.5183308124542236, "learning_rate": 2.267002518891688e-05, "loss": 0.8176, "step": 180 }, { "epoch": 0.04794650179799382, "grad_norm": 0.5419854521751404, "learning_rate": 2.392947103274559e-05, "loss": 0.8083, "step": 190 }, { "epoch": 0.05047000189262507, "grad_norm": 0.495588093996048, "learning_rate": 2.5188916876574308e-05, "loss": 0.8034, "step": 200 }, { "epoch": 0.052993501987256324, "grad_norm": 0.5842857956886292, "learning_rate": 2.6448362720403024e-05, "loss": 0.7939, "step": 210 }, { "epoch": 0.055517002081887576, "grad_norm": 0.47628337144851685, "learning_rate": 2.770780856423174e-05, "loss": 0.8022, "step": 220 }, { "epoch": 0.058040502176518835, "grad_norm": 0.513734757900238, "learning_rate": 2.8967254408060457e-05, "loss": 0.7521, "step": 230 }, { "epoch": 0.06056400227115009, "grad_norm": 0.4637106657028198, "learning_rate": 3.0226700251889174e-05, "loss": 0.7651, "step": 240 }, { "epoch": 0.06308750236578134, "grad_norm": 0.5194531083106995, "learning_rate": 3.148614609571788e-05, "loss": 0.8277, "step": 250 }, { "epoch": 0.0656110024604126, "grad_norm": 0.5461646318435669, "learning_rate": 3.27455919395466e-05, "loss": 0.8012, "step": 260 }, { "epoch": 0.06813450255504384, "grad_norm": 0.525729238986969, "learning_rate": 3.4005037783375316e-05, "loss": 0.7945, "step": 270 }, { "epoch": 0.0706580026496751, "grad_norm": 0.5770601034164429, "learning_rate": 3.526448362720403e-05, "loss": 0.7753, "step": 280 }, { "epoch": 0.07318150274430635, "grad_norm": 0.5252787470817566, "learning_rate": 3.652392947103275e-05, "loss": 0.7889, "step": 290 }, { "epoch": 0.07570500283893761, "grad_norm": 0.5346773266792297, "learning_rate": 3.7783375314861465e-05, "loss": 0.7592, "step": 300 }, { "epoch": 0.07822850293356887, "grad_norm": 0.5902045965194702, "learning_rate": 3.904282115869018e-05, "loss": 0.8066, "step": 310 }, { "epoch": 0.08075200302820011, "grad_norm": 0.5109593868255615, "learning_rate": 4.03022670025189e-05, "loss": 0.7715, "step": 320 }, { "epoch": 0.08327550312283137, "grad_norm": 0.4999026656150818, "learning_rate": 4.1561712846347615e-05, "loss": 0.795, "step": 330 }, { "epoch": 0.08579900321746262, "grad_norm": 0.4799485504627228, "learning_rate": 4.2821158690176324e-05, "loss": 0.7621, "step": 340 }, { "epoch": 0.08832250331209388, "grad_norm": 0.5258325338363647, "learning_rate": 4.408060453400504e-05, "loss": 0.7471, "step": 350 }, { "epoch": 0.09084600340672513, "grad_norm": 0.5144442319869995, "learning_rate": 4.534005037783376e-05, "loss": 0.7888, "step": 360 }, { "epoch": 0.09336950350135638, "grad_norm": 0.5580424666404724, "learning_rate": 4.659949622166247e-05, "loss": 0.7865, "step": 370 }, { "epoch": 0.09589300359598764, "grad_norm": 0.5704917907714844, "learning_rate": 4.785894206549118e-05, "loss": 0.8025, "step": 380 }, { "epoch": 0.09841650369061888, "grad_norm": 0.4671490490436554, "learning_rate": 4.91183879093199e-05, "loss": 0.8083, "step": 390 }, { "epoch": 0.10094000378525014, "grad_norm": 0.5287295579910278, "learning_rate": 4.999991263591223e-05, "loss": 0.7433, "step": 400 }, { "epoch": 0.1034635038798814, "grad_norm": 0.5582460165023804, "learning_rate": 4.9998359513560176e-05, "loss": 0.7645, "step": 410 }, { "epoch": 0.10598700397451265, "grad_norm": 0.48861467838287354, "learning_rate": 4.999486510586282e-05, "loss": 0.8056, "step": 420 }, { "epoch": 0.1085105040691439, "grad_norm": 0.565294086933136, "learning_rate": 4.9989429684183686e-05, "loss": 0.8072, "step": 430 }, { "epoch": 0.11103400416377515, "grad_norm": 0.5327048897743225, "learning_rate": 4.9982053670618626e-05, "loss": 0.7871, "step": 440 }, { "epoch": 0.11355750425840641, "grad_norm": 0.5085189342498779, "learning_rate": 4.997273763796312e-05, "loss": 0.7772, "step": 450 }, { "epoch": 0.11608100435303767, "grad_norm": 0.5682306885719299, "learning_rate": 4.996148230966775e-05, "loss": 0.7521, "step": 460 }, { "epoch": 0.11860450444766892, "grad_norm": 0.5310999751091003, "learning_rate": 4.994828855978202e-05, "loss": 0.8276, "step": 470 }, { "epoch": 0.12112800454230017, "grad_norm": 0.5132438540458679, "learning_rate": 4.99331574128865e-05, "loss": 0.7388, "step": 480 }, { "epoch": 0.12365150463693142, "grad_norm": 0.48004600405693054, "learning_rate": 4.991609004401324e-05, "loss": 0.7898, "step": 490 }, { "epoch": 0.12617500473156268, "grad_norm": 0.4942034184932709, "learning_rate": 4.989708777855453e-05, "loss": 0.7904, "step": 500 }, { "epoch": 0.12869850482619394, "grad_norm": 0.4329438805580139, "learning_rate": 4.9876152092159994e-05, "loss": 0.8144, "step": 510 }, { "epoch": 0.1312220049208252, "grad_norm": 0.4422023892402649, "learning_rate": 4.985328461062195e-05, "loss": 0.7894, "step": 520 }, { "epoch": 0.13374550501545643, "grad_norm": 0.4740963876247406, "learning_rate": 4.98284871097492e-05, "loss": 0.8185, "step": 530 }, { "epoch": 0.1362690051100877, "grad_norm": 0.5256137847900391, "learning_rate": 4.98017615152291e-05, "loss": 0.7917, "step": 540 }, { "epoch": 0.13879250520471895, "grad_norm": 0.48558464646339417, "learning_rate": 4.977310990247807e-05, "loss": 0.7791, "step": 550 }, { "epoch": 0.1413160052993502, "grad_norm": 0.5179520845413208, "learning_rate": 4.974253449648031e-05, "loss": 0.7854, "step": 560 }, { "epoch": 0.14383950539398146, "grad_norm": 0.4606976807117462, "learning_rate": 4.971003767161516e-05, "loss": 0.7776, "step": 570 }, { "epoch": 0.1463630054886127, "grad_norm": 0.4805031716823578, "learning_rate": 4.9675621951472584e-05, "loss": 0.7877, "step": 580 }, { "epoch": 0.14888650558324396, "grad_norm": 0.5021762847900391, "learning_rate": 4.9639290008657304e-05, "loss": 0.7923, "step": 590 }, { "epoch": 0.15141000567787521, "grad_norm": 0.5204519033432007, "learning_rate": 4.960104466458118e-05, "loss": 0.7973, "step": 600 }, { "epoch": 0.15393350577250647, "grad_norm": 0.5029839277267456, "learning_rate": 4.956088888924414e-05, "loss": 0.766, "step": 610 }, { "epoch": 0.15645700586713773, "grad_norm": 0.5078247785568237, "learning_rate": 4.951882580100353e-05, "loss": 0.7928, "step": 620 }, { "epoch": 0.15898050596176896, "grad_norm": 0.4767574369907379, "learning_rate": 4.947485866633199e-05, "loss": 0.7886, "step": 630 }, { "epoch": 0.16150400605640022, "grad_norm": 0.4844212234020233, "learning_rate": 4.94289908995637e-05, "loss": 0.7187, "step": 640 }, { "epoch": 0.16402750615103148, "grad_norm": 0.4916108548641205, "learning_rate": 4.938122606262936e-05, "loss": 0.8031, "step": 650 }, { "epoch": 0.16655100624566274, "grad_norm": 0.8660295009613037, "learning_rate": 4.9331567864779457e-05, "loss": 0.8129, "step": 660 }, { "epoch": 0.169074506340294, "grad_norm": 0.4912508428096771, "learning_rate": 4.928002016229634e-05, "loss": 0.7973, "step": 670 }, { "epoch": 0.17159800643492523, "grad_norm": 0.5532618165016174, "learning_rate": 4.9226586958194647e-05, "loss": 0.7988, "step": 680 }, { "epoch": 0.1741215065295565, "grad_norm": 0.4351136386394501, "learning_rate": 4.9171272401910504e-05, "loss": 0.7453, "step": 690 }, { "epoch": 0.17664500662418775, "grad_norm": 0.4713657796382904, "learning_rate": 4.9114080788979284e-05, "loss": 0.793, "step": 700 }, { "epoch": 0.179168506718819, "grad_norm": 0.4904249608516693, "learning_rate": 4.905501656070202e-05, "loss": 0.791, "step": 710 }, { "epoch": 0.18169200681345027, "grad_norm": 0.5009727478027344, "learning_rate": 4.8994084303800525e-05, "loss": 0.7785, "step": 720 }, { "epoch": 0.1842155069080815, "grad_norm": 0.4555422365665436, "learning_rate": 4.89312887500612e-05, "loss": 0.7772, "step": 730 }, { "epoch": 0.18673900700271276, "grad_norm": 0.4672720432281494, "learning_rate": 4.8866634775967544e-05, "loss": 0.7796, "step": 740 }, { "epoch": 0.18926250709734402, "grad_norm": 0.42111048102378845, "learning_rate": 4.880012740232154e-05, "loss": 0.7741, "step": 750 }, { "epoch": 0.19178600719197528, "grad_norm": 0.5067655444145203, "learning_rate": 4.873177179385368e-05, "loss": 0.7603, "step": 760 }, { "epoch": 0.19430950728660654, "grad_norm": 0.4543984532356262, "learning_rate": 4.866157325882192e-05, "loss": 0.7766, "step": 770 }, { "epoch": 0.19683300738123777, "grad_norm": 0.46171534061431885, "learning_rate": 4.858953724859948e-05, "loss": 0.7651, "step": 780 }, { "epoch": 0.19935650747586903, "grad_norm": 1.1524691581726074, "learning_rate": 4.851566935725147e-05, "loss": 0.7538, "step": 790 }, { "epoch": 0.20188000757050029, "grad_norm": 0.4468795657157898, "learning_rate": 4.843997532110051e-05, "loss": 0.7735, "step": 800 }, { "epoch": 0.20440350766513155, "grad_norm": 0.6003584861755371, "learning_rate": 4.836246101828124e-05, "loss": 0.7862, "step": 810 }, { "epoch": 0.2069270077597628, "grad_norm": 0.44003739953041077, "learning_rate": 4.828313246828386e-05, "loss": 0.8073, "step": 820 }, { "epoch": 0.20945050785439404, "grad_norm": 0.4460177421569824, "learning_rate": 4.820199583148667e-05, "loss": 0.7788, "step": 830 }, { "epoch": 0.2119740079490253, "grad_norm": 0.4435465633869171, "learning_rate": 4.811905740867769e-05, "loss": 0.7963, "step": 840 }, { "epoch": 0.21449750804365655, "grad_norm": 0.5036677718162537, "learning_rate": 4.803432364056535e-05, "loss": 0.7925, "step": 850 }, { "epoch": 0.2170210081382878, "grad_norm": 0.46002766489982605, "learning_rate": 4.794780110727832e-05, "loss": 0.7683, "step": 860 }, { "epoch": 0.21954450823291907, "grad_norm": 0.4961997866630554, "learning_rate": 4.785949652785453e-05, "loss": 0.7911, "step": 870 }, { "epoch": 0.2220680083275503, "grad_norm": 0.5581184029579163, "learning_rate": 4.776941675971941e-05, "loss": 0.7739, "step": 880 }, { "epoch": 0.22459150842218156, "grad_norm": 0.4687346816062927, "learning_rate": 4.767756879815334e-05, "loss": 0.7928, "step": 890 }, { "epoch": 0.22711500851681282, "grad_norm": 0.5058118104934692, "learning_rate": 4.758395977574841e-05, "loss": 0.8065, "step": 900 }, { "epoch": 0.22963850861144408, "grad_norm": 0.49092328548431396, "learning_rate": 4.748859696185458e-05, "loss": 0.7966, "step": 910 }, { "epoch": 0.23216200870607534, "grad_norm": 0.4792310893535614, "learning_rate": 4.739148776201512e-05, "loss": 0.7383, "step": 920 }, { "epoch": 0.23468550880070657, "grad_norm": 0.4560384750366211, "learning_rate": 4.729263971739154e-05, "loss": 0.7869, "step": 930 }, { "epoch": 0.23720900889533783, "grad_norm": 0.46963706612586975, "learning_rate": 4.719206050417796e-05, "loss": 0.7748, "step": 940 }, { "epoch": 0.2397325089899691, "grad_norm": 0.4931532144546509, "learning_rate": 4.7089757933005016e-05, "loss": 0.7747, "step": 950 }, { "epoch": 0.24225600908460035, "grad_norm": 0.4691235423088074, "learning_rate": 4.698573994833332e-05, "loss": 0.7725, "step": 960 }, { "epoch": 0.2447795091792316, "grad_norm": 0.44244667887687683, "learning_rate": 4.688001462783648e-05, "loss": 0.7564, "step": 970 }, { "epoch": 0.24730300927386284, "grad_norm": 0.5181055068969727, "learning_rate": 4.6772590181773866e-05, "loss": 0.7407, "step": 980 }, { "epoch": 0.2498265093684941, "grad_norm": 0.4613230526447296, "learning_rate": 4.6663474952353004e-05, "loss": 0.7437, "step": 990 }, { "epoch": 0.25235000946312536, "grad_norm": 0.5326347351074219, "learning_rate": 4.6552677413081756e-05, "loss": 0.809, "step": 1000 }, { "epoch": 0.2548735095577566, "grad_norm": 0.516962468624115, "learning_rate": 4.644020616811029e-05, "loss": 0.7867, "step": 1010 }, { "epoch": 0.2573970096523879, "grad_norm": 0.5326249003410339, "learning_rate": 4.6326069951562924e-05, "loss": 0.8096, "step": 1020 }, { "epoch": 0.25992050974701914, "grad_norm": 0.42123785614967346, "learning_rate": 4.6210277626859856e-05, "loss": 0.7241, "step": 1030 }, { "epoch": 0.2624440098416504, "grad_norm": 0.40459808707237244, "learning_rate": 4.609283818602884e-05, "loss": 0.7797, "step": 1040 }, { "epoch": 0.2649675099362816, "grad_norm": 0.4971398115158081, "learning_rate": 4.5973760749006963e-05, "loss": 0.7898, "step": 1050 }, { "epoch": 0.26749101003091286, "grad_norm": 0.4381590485572815, "learning_rate": 4.585305456293235e-05, "loss": 0.8102, "step": 1060 }, { "epoch": 0.2700145101255441, "grad_norm": 0.6431146264076233, "learning_rate": 4.5730729001426083e-05, "loss": 0.7927, "step": 1070 }, { "epoch": 0.2725380102201754, "grad_norm": 0.46233826875686646, "learning_rate": 4.5606793563864316e-05, "loss": 0.7656, "step": 1080 }, { "epoch": 0.27506151031480663, "grad_norm": 0.5226401090621948, "learning_rate": 4.548125787464054e-05, "loss": 0.7587, "step": 1090 }, { "epoch": 0.2775850104094379, "grad_norm": 0.4273032248020172, "learning_rate": 4.535413168241821e-05, "loss": 0.7697, "step": 1100 }, { "epoch": 0.28010851050406915, "grad_norm": 0.4383103549480438, "learning_rate": 4.522542485937369e-05, "loss": 0.8039, "step": 1110 }, { "epoch": 0.2826320105987004, "grad_norm": 0.4111369848251343, "learning_rate": 4.509514740042962e-05, "loss": 0.7707, "step": 1120 }, { "epoch": 0.28515551069333167, "grad_norm": 0.5138412117958069, "learning_rate": 4.496330942247873e-05, "loss": 0.7785, "step": 1130 }, { "epoch": 0.28767901078796293, "grad_norm": 0.49675437808036804, "learning_rate": 4.482992116359824e-05, "loss": 0.763, "step": 1140 }, { "epoch": 0.29020251088259413, "grad_norm": 0.4325181841850281, "learning_rate": 4.469499298225473e-05, "loss": 0.7344, "step": 1150 }, { "epoch": 0.2927260109772254, "grad_norm": 0.45691972970962524, "learning_rate": 4.455853535649984e-05, "loss": 0.7628, "step": 1160 }, { "epoch": 0.29524951107185665, "grad_norm": 0.45635420083999634, "learning_rate": 4.442055888315646e-05, "loss": 0.7666, "step": 1170 }, { "epoch": 0.2977730111664879, "grad_norm": 0.44449037313461304, "learning_rate": 4.4281074276995936e-05, "loss": 0.7308, "step": 1180 }, { "epoch": 0.30029651126111917, "grad_norm": 0.4645417332649231, "learning_rate": 4.4140092369905914e-05, "loss": 0.7385, "step": 1190 }, { "epoch": 0.30282001135575043, "grad_norm": 0.4081282913684845, "learning_rate": 4.399762411004922e-05, "loss": 0.7754, "step": 1200 }, { "epoch": 0.3053435114503817, "grad_norm": 0.5219441652297974, "learning_rate": 4.3853680561013647e-05, "loss": 0.7903, "step": 1210 }, { "epoch": 0.30786701154501295, "grad_norm": 0.48942479491233826, "learning_rate": 4.370827290095277e-05, "loss": 0.7693, "step": 1220 }, { "epoch": 0.3103905116396442, "grad_norm": 0.4511029124259949, "learning_rate": 4.356141242171795e-05, "loss": 0.7685, "step": 1230 }, { "epoch": 0.31291401173427547, "grad_norm": 0.46076688170433044, "learning_rate": 4.3413110527981406e-05, "loss": 0.745, "step": 1240 }, { "epoch": 0.31543751182890667, "grad_norm": 0.5077107548713684, "learning_rate": 4.3263378736350566e-05, "loss": 0.7679, "step": 1250 }, { "epoch": 0.31796101192353793, "grad_norm": 0.4735719561576843, "learning_rate": 4.311222867447375e-05, "loss": 0.8021, "step": 1260 }, { "epoch": 0.3204845120181692, "grad_norm": 0.49257341027259827, "learning_rate": 4.295967208013717e-05, "loss": 0.7619, "step": 1270 }, { "epoch": 0.32300801211280045, "grad_norm": 0.45219412446022034, "learning_rate": 4.280572080035348e-05, "loss": 0.7656, "step": 1280 }, { "epoch": 0.3255315122074317, "grad_norm": 0.4800606369972229, "learning_rate": 4.2650386790441696e-05, "loss": 0.7618, "step": 1290 }, { "epoch": 0.32805501230206296, "grad_norm": 0.5180128812789917, "learning_rate": 4.2493682113098855e-05, "loss": 0.766, "step": 1300 }, { "epoch": 0.3305785123966942, "grad_norm": 0.45564642548561096, "learning_rate": 4.233561893746323e-05, "loss": 0.7944, "step": 1310 }, { "epoch": 0.3331020124913255, "grad_norm": 0.46252918243408203, "learning_rate": 4.217620953816935e-05, "loss": 0.749, "step": 1320 }, { "epoch": 0.33562551258595674, "grad_norm": 0.531288743019104, "learning_rate": 4.2015466294394756e-05, "loss": 0.7869, "step": 1330 }, { "epoch": 0.338149012680588, "grad_norm": 0.5134521722793579, "learning_rate": 4.185340168889868e-05, "loss": 0.7568, "step": 1340 }, { "epoch": 0.3406725127752192, "grad_norm": 0.43066537380218506, "learning_rate": 4.169002830705274e-05, "loss": 0.7527, "step": 1350 }, { "epoch": 0.34319601286985046, "grad_norm": 0.4531955122947693, "learning_rate": 4.152535883586352e-05, "loss": 0.738, "step": 1360 }, { "epoch": 0.3457195129644817, "grad_norm": 0.4627138376235962, "learning_rate": 4.135940606298738e-05, "loss": 0.7441, "step": 1370 }, { "epoch": 0.348243013059113, "grad_norm": 0.4333173632621765, "learning_rate": 4.119218287573743e-05, "loss": 0.7344, "step": 1380 }, { "epoch": 0.35076651315374424, "grad_norm": 0.421912282705307, "learning_rate": 4.102370226008271e-05, "loss": 0.7828, "step": 1390 }, { "epoch": 0.3532900132483755, "grad_norm": 0.4382674992084503, "learning_rate": 4.085397729963976e-05, "loss": 0.7526, "step": 1400 }, { "epoch": 0.35581351334300676, "grad_norm": 0.4683435559272766, "learning_rate": 4.06830211746566e-05, "loss": 0.7804, "step": 1410 }, { "epoch": 0.358337013437638, "grad_norm": 0.49529722332954407, "learning_rate": 4.051084716098921e-05, "loss": 0.7565, "step": 1420 }, { "epoch": 0.3608605135322693, "grad_norm": 0.454121470451355, "learning_rate": 4.0337468629070496e-05, "loss": 0.7667, "step": 1430 }, { "epoch": 0.36338401362690054, "grad_norm": 0.45073944330215454, "learning_rate": 4.016289904287212e-05, "loss": 0.7543, "step": 1440 }, { "epoch": 0.36590751372153174, "grad_norm": 0.5012348890304565, "learning_rate": 3.9987151958858794e-05, "loss": 0.7642, "step": 1450 }, { "epoch": 0.368431013816163, "grad_norm": 0.45489585399627686, "learning_rate": 3.981024102493566e-05, "loss": 0.7355, "step": 1460 }, { "epoch": 0.37095451391079426, "grad_norm": 0.49024873971939087, "learning_rate": 3.963217997938834e-05, "loss": 0.7656, "step": 1470 }, { "epoch": 0.3734780140054255, "grad_norm": 0.4967275857925415, "learning_rate": 3.945298264981614e-05, "loss": 0.7493, "step": 1480 }, { "epoch": 0.3760015141000568, "grad_norm": 0.48888593912124634, "learning_rate": 3.927266295205818e-05, "loss": 0.7741, "step": 1490 }, { "epoch": 0.37852501419468804, "grad_norm": 0.44507086277008057, "learning_rate": 3.9091234889112815e-05, "loss": 0.7632, "step": 1500 }, { "epoch": 0.3810485142893193, "grad_norm": 0.4503888785839081, "learning_rate": 3.8908712550050154e-05, "loss": 0.7605, "step": 1510 }, { "epoch": 0.38357201438395055, "grad_norm": 0.42421767115592957, "learning_rate": 3.8725110108917975e-05, "loss": 0.7325, "step": 1520 }, { "epoch": 0.3860955144785818, "grad_norm": 0.45737549662590027, "learning_rate": 3.854044182364098e-05, "loss": 0.7644, "step": 1530 }, { "epoch": 0.3886190145732131, "grad_norm": 0.46508800983428955, "learning_rate": 3.835472203491367e-05, "loss": 0.7492, "step": 1540 }, { "epoch": 0.3911425146678443, "grad_norm": 0.42231640219688416, "learning_rate": 3.816796516508658e-05, "loss": 0.7495, "step": 1550 }, { "epoch": 0.39366601476247554, "grad_norm": 0.44611889123916626, "learning_rate": 3.798018571704638e-05, "loss": 0.7468, "step": 1560 }, { "epoch": 0.3961895148571068, "grad_norm": 0.4249905049800873, "learning_rate": 3.779139827308956e-05, "loss": 0.7757, "step": 1570 }, { "epoch": 0.39871301495173805, "grad_norm": 0.46931806206703186, "learning_rate": 3.760161749379008e-05, "loss": 0.7576, "step": 1580 }, { "epoch": 0.4012365150463693, "grad_norm": 0.4227593243122101, "learning_rate": 3.7410858116860836e-05, "loss": 0.7415, "step": 1590 }, { "epoch": 0.40376001514100057, "grad_norm": 0.4338664710521698, "learning_rate": 3.721913495600923e-05, "loss": 0.7676, "step": 1600 }, { "epoch": 0.40628351523563183, "grad_norm": 0.4017735719680786, "learning_rate": 3.7026462899786726e-05, "loss": 0.7171, "step": 1610 }, { "epoch": 0.4088070153302631, "grad_norm": 0.4449094235897064, "learning_rate": 3.683285691043272e-05, "loss": 0.7617, "step": 1620 }, { "epoch": 0.41133051542489435, "grad_norm": 0.4355940818786621, "learning_rate": 3.663833202271257e-05, "loss": 0.7616, "step": 1630 }, { "epoch": 0.4138540155195256, "grad_norm": 0.4568544328212738, "learning_rate": 3.6442903342750084e-05, "loss": 0.7141, "step": 1640 }, { "epoch": 0.4163775156141568, "grad_norm": 0.46118539571762085, "learning_rate": 3.624658604685443e-05, "loss": 0.7347, "step": 1650 }, { "epoch": 0.41890101570878807, "grad_norm": 0.49249446392059326, "learning_rate": 3.604939538034158e-05, "loss": 0.7557, "step": 1660 }, { "epoch": 0.42142451580341933, "grad_norm": 0.5076271891593933, "learning_rate": 3.585134665635041e-05, "loss": 0.7512, "step": 1670 }, { "epoch": 0.4239480158980506, "grad_norm": 0.4391838312149048, "learning_rate": 3.565245525465355e-05, "loss": 0.782, "step": 1680 }, { "epoch": 0.42647151599268185, "grad_norm": 0.43777164816856384, "learning_rate": 3.5452736620463064e-05, "loss": 0.7442, "step": 1690 }, { "epoch": 0.4289950160873131, "grad_norm": 0.4787698984146118, "learning_rate": 3.525220626323097e-05, "loss": 0.7263, "step": 1700 }, { "epoch": 0.43151851618194437, "grad_norm": 0.4493311941623688, "learning_rate": 3.5050879755444877e-05, "loss": 0.7771, "step": 1710 }, { "epoch": 0.4340420162765756, "grad_norm": 0.4277406334877014, "learning_rate": 3.484877273141866e-05, "loss": 0.7533, "step": 1720 }, { "epoch": 0.4365655163712069, "grad_norm": 0.44397228956222534, "learning_rate": 3.464590088607839e-05, "loss": 0.7581, "step": 1730 }, { "epoch": 0.43908901646583814, "grad_norm": 0.4278891086578369, "learning_rate": 3.444227997374345e-05, "loss": 0.7726, "step": 1740 }, { "epoch": 0.44161251656046935, "grad_norm": 0.4884257912635803, "learning_rate": 3.4237925806903184e-05, "loss": 0.7537, "step": 1750 }, { "epoch": 0.4441360166551006, "grad_norm": 0.39494335651397705, "learning_rate": 3.403285425498889e-05, "loss": 0.7513, "step": 1760 }, { "epoch": 0.44665951674973187, "grad_norm": 0.4903046488761902, "learning_rate": 3.3827081243141534e-05, "loss": 0.7259, "step": 1770 }, { "epoch": 0.4491830168443631, "grad_norm": 0.47485992312431335, "learning_rate": 3.362062275097496e-05, "loss": 0.7674, "step": 1780 }, { "epoch": 0.4517065169389944, "grad_norm": 0.47219303250312805, "learning_rate": 3.341349481133507e-05, "loss": 0.7234, "step": 1790 }, { "epoch": 0.45423001703362564, "grad_norm": 0.4186359643936157, "learning_rate": 3.320571350905466e-05, "loss": 0.7587, "step": 1800 }, { "epoch": 0.4567535171282569, "grad_norm": 0.4978404641151428, "learning_rate": 3.299729497970444e-05, "loss": 0.7754, "step": 1810 }, { "epoch": 0.45927701722288816, "grad_norm": 0.45416682958602905, "learning_rate": 3.278825540833995e-05, "loss": 0.7464, "step": 1820 }, { "epoch": 0.4618005173175194, "grad_norm": 0.433159202337265, "learning_rate": 3.2578611028244656e-05, "loss": 0.7355, "step": 1830 }, { "epoch": 0.4643240174121507, "grad_norm": 0.47343000769615173, "learning_rate": 3.2368378119669363e-05, "loss": 0.7115, "step": 1840 }, { "epoch": 0.4668475175067819, "grad_norm": 0.39492008090019226, "learning_rate": 3.215757300856796e-05, "loss": 0.731, "step": 1850 }, { "epoch": 0.46937101760141314, "grad_norm": 0.4586881697177887, "learning_rate": 3.194621206532957e-05, "loss": 0.7729, "step": 1860 }, { "epoch": 0.4718945176960444, "grad_norm": 0.5566070675849915, "learning_rate": 3.173431170350732e-05, "loss": 0.7454, "step": 1870 }, { "epoch": 0.47441801779067566, "grad_norm": 0.44794589281082153, "learning_rate": 3.152188837854369e-05, "loss": 0.7682, "step": 1880 }, { "epoch": 0.4769415178853069, "grad_norm": 0.45499035716056824, "learning_rate": 3.130895858649264e-05, "loss": 0.7243, "step": 1890 }, { "epoch": 0.4794650179799382, "grad_norm": 0.4766923487186432, "learning_rate": 3.109553886273863e-05, "loss": 0.7582, "step": 1900 }, { "epoch": 0.48198851807456944, "grad_norm": 0.4546712040901184, "learning_rate": 3.088164578071246e-05, "loss": 0.7318, "step": 1910 }, { "epoch": 0.4845120181692007, "grad_norm": 0.46813201904296875, "learning_rate": 3.066729595060431e-05, "loss": 0.7199, "step": 1920 }, { "epoch": 0.48703551826383196, "grad_norm": 0.5136599540710449, "learning_rate": 3.0452506018073833e-05, "loss": 0.7531, "step": 1930 }, { "epoch": 0.4895590183584632, "grad_norm": 0.48750588297843933, "learning_rate": 3.0237292662957473e-05, "loss": 0.7373, "step": 1940 }, { "epoch": 0.4920825184530944, "grad_norm": 0.42763668298721313, "learning_rate": 3.0021672597973207e-05, "loss": 0.7256, "step": 1950 }, { "epoch": 0.4946060185477257, "grad_norm": 0.44293949007987976, "learning_rate": 2.9805662567422676e-05, "loss": 0.7644, "step": 1960 }, { "epoch": 0.49712951864235694, "grad_norm": 0.5033817887306213, "learning_rate": 2.9589279345890895e-05, "loss": 0.7521, "step": 1970 }, { "epoch": 0.4996530187369882, "grad_norm": 0.3790980577468872, "learning_rate": 2.9372539736943577e-05, "loss": 0.7437, "step": 1980 }, { "epoch": 0.5021765188316195, "grad_norm": 0.3884977400302887, "learning_rate": 2.9155460571822245e-05, "loss": 0.7309, "step": 1990 }, { "epoch": 0.5047000189262507, "grad_norm": 0.4705371558666229, "learning_rate": 2.893805870813717e-05, "loss": 0.7223, "step": 2000 }, { "epoch": 0.5072235190208819, "grad_norm": 0.42519432306289673, "learning_rate": 2.872035102855826e-05, "loss": 0.7247, "step": 2010 }, { "epoch": 0.5097470191155132, "grad_norm": 0.4545595347881317, "learning_rate": 2.850235443950402e-05, "loss": 0.7047, "step": 2020 }, { "epoch": 0.5122705192101444, "grad_norm": 0.4008665382862091, "learning_rate": 2.8284085869828665e-05, "loss": 0.7417, "step": 2030 }, { "epoch": 0.5147940193047758, "grad_norm": 0.4608813226222992, "learning_rate": 2.8065562269507463e-05, "loss": 0.7508, "step": 2040 }, { "epoch": 0.517317519399407, "grad_norm": 0.5732349157333374, "learning_rate": 2.7846800608320485e-05, "loss": 0.7352, "step": 2050 }, { "epoch": 0.5198410194940383, "grad_norm": 0.39458563923835754, "learning_rate": 2.7627817874534762e-05, "loss": 0.7205, "step": 2060 }, { "epoch": 0.5223645195886695, "grad_norm": 0.473949670791626, "learning_rate": 2.7408631073585068e-05, "loss": 0.7096, "step": 2070 }, { "epoch": 0.5248880196833008, "grad_norm": 0.45874857902526855, "learning_rate": 2.7189257226753305e-05, "loss": 0.7349, "step": 2080 }, { "epoch": 0.527411519777932, "grad_norm": 0.41197165846824646, "learning_rate": 2.696971336984672e-05, "loss": 0.7522, "step": 2090 }, { "epoch": 0.5299350198725632, "grad_norm": 0.5001893639564514, "learning_rate": 2.6750016551874945e-05, "loss": 0.7249, "step": 2100 }, { "epoch": 0.5324585199671945, "grad_norm": 0.4561350345611572, "learning_rate": 2.6530183833726025e-05, "loss": 0.732, "step": 2110 }, { "epoch": 0.5349820200618257, "grad_norm": 0.4732155203819275, "learning_rate": 2.6310232286841546e-05, "loss": 0.7581, "step": 2120 }, { "epoch": 0.537505520156457, "grad_norm": 0.44940948486328125, "learning_rate": 2.609017899189092e-05, "loss": 0.7439, "step": 2130 }, { "epoch": 0.5400290202510882, "grad_norm": 0.4435223639011383, "learning_rate": 2.587004103744495e-05, "loss": 0.7016, "step": 2140 }, { "epoch": 0.5425525203457195, "grad_norm": 0.4022032618522644, "learning_rate": 2.564983551864882e-05, "loss": 0.7124, "step": 2150 }, { "epoch": 0.5450760204403508, "grad_norm": 0.5092436671257019, "learning_rate": 2.54295795358945e-05, "loss": 0.7355, "step": 2160 }, { "epoch": 0.5475995205349821, "grad_norm": 0.43439218401908875, "learning_rate": 2.5209290193492834e-05, "loss": 0.7301, "step": 2170 }, { "epoch": 0.5501230206296133, "grad_norm": 0.4191984236240387, "learning_rate": 2.4988984598345247e-05, "loss": 0.6951, "step": 2180 }, { "epoch": 0.5526465207242445, "grad_norm": 0.42209818959236145, "learning_rate": 2.4768679858615304e-05, "loss": 0.746, "step": 2190 }, { "epoch": 0.5551700208188758, "grad_norm": 0.49302446842193604, "learning_rate": 2.454839308240014e-05, "loss": 0.7687, "step": 2200 }, { "epoch": 0.557693520913507, "grad_norm": 0.46349719166755676, "learning_rate": 2.4328141376401903e-05, "loss": 0.7203, "step": 2210 }, { "epoch": 0.5602170210081383, "grad_norm": 0.4288594722747803, "learning_rate": 2.4107941844599312e-05, "loss": 0.709, "step": 2220 }, { "epoch": 0.5627405211027695, "grad_norm": 0.42657721042633057, "learning_rate": 2.3887811586919424e-05, "loss": 0.7385, "step": 2230 }, { "epoch": 0.5652640211974008, "grad_norm": 0.447419673204422, "learning_rate": 2.3667767697909694e-05, "loss": 0.7199, "step": 2240 }, { "epoch": 0.567787521292032, "grad_norm": 0.460227906703949, "learning_rate": 2.3447827265410517e-05, "loss": 0.76, "step": 2250 }, { "epoch": 0.5703110213866633, "grad_norm": 0.48693689703941345, "learning_rate": 2.3228007369228178e-05, "loss": 0.7893, "step": 2260 }, { "epoch": 0.5728345214812945, "grad_norm": 0.4543781578540802, "learning_rate": 2.3008325079808576e-05, "loss": 0.7431, "step": 2270 }, { "epoch": 0.5753580215759259, "grad_norm": 0.44196173548698425, "learning_rate": 2.2788797456911503e-05, "loss": 0.7024, "step": 2280 }, { "epoch": 0.5778815216705571, "grad_norm": 0.4773748517036438, "learning_rate": 2.2569441548285934e-05, "loss": 0.7366, "step": 2290 }, { "epoch": 0.5804050217651883, "grad_norm": 0.35127493739128113, "learning_rate": 2.2350274388346064e-05, "loss": 0.6932, "step": 2300 }, { "epoch": 0.5829285218598196, "grad_norm": 0.4260622262954712, "learning_rate": 2.213131299684858e-05, "loss": 0.7481, "step": 2310 }, { "epoch": 0.5854520219544508, "grad_norm": 0.49954676628112793, "learning_rate": 2.191257437757086e-05, "loss": 0.7196, "step": 2320 }, { "epoch": 0.5879755220490821, "grad_norm": 0.41478589177131653, "learning_rate": 2.16940755169906e-05, "loss": 0.744, "step": 2330 }, { "epoch": 0.5904990221437133, "grad_norm": 0.4980543255805969, "learning_rate": 2.1475833382966647e-05, "loss": 0.7798, "step": 2340 }, { "epoch": 0.5930225222383446, "grad_norm": 0.4331112802028656, "learning_rate": 2.1257864923421404e-05, "loss": 0.7616, "step": 2350 }, { "epoch": 0.5955460223329758, "grad_norm": 0.38317206501960754, "learning_rate": 2.1040187065024605e-05, "loss": 0.6863, "step": 2360 }, { "epoch": 0.5980695224276071, "grad_norm": 0.44444334506988525, "learning_rate": 2.0822816711878978e-05, "loss": 0.7482, "step": 2370 }, { "epoch": 0.6005930225222383, "grad_norm": 0.47149887681007385, "learning_rate": 2.0605770744207413e-05, "loss": 0.7289, "step": 2380 }, { "epoch": 0.6031165226168695, "grad_norm": 0.4344223737716675, "learning_rate": 2.0389066017042192e-05, "loss": 0.708, "step": 2390 }, { "epoch": 0.6056400227115009, "grad_norm": 0.8328418135643005, "learning_rate": 2.0172719358916042e-05, "loss": 0.709, "step": 2400 }, { "epoch": 0.6081635228061321, "grad_norm": 0.4812801480293274, "learning_rate": 1.9956747570555288e-05, "loss": 0.7526, "step": 2410 }, { "epoch": 0.6106870229007634, "grad_norm": 0.43277397751808167, "learning_rate": 1.9741167423575186e-05, "loss": 0.6928, "step": 2420 }, { "epoch": 0.6132105229953946, "grad_norm": 0.41379523277282715, "learning_rate": 1.9525995659177484e-05, "loss": 0.7236, "step": 2430 }, { "epoch": 0.6157340230900259, "grad_norm": 0.4115571081638336, "learning_rate": 1.9311248986850365e-05, "loss": 0.7191, "step": 2440 }, { "epoch": 0.6182575231846571, "grad_norm": 0.4394511282444, "learning_rate": 1.9096944083070866e-05, "loss": 0.7345, "step": 2450 }, { "epoch": 0.6207810232792884, "grad_norm": 0.46809184551239014, "learning_rate": 1.8883097590009775e-05, "loss": 0.7342, "step": 2460 }, { "epoch": 0.6233045233739196, "grad_norm": 0.4622188210487366, "learning_rate": 1.866972611423936e-05, "loss": 0.7239, "step": 2470 }, { "epoch": 0.6258280234685509, "grad_norm": 0.5006013512611389, "learning_rate": 1.8456846225443648e-05, "loss": 0.6781, "step": 2480 }, { "epoch": 0.6283515235631821, "grad_norm": 0.43140849471092224, "learning_rate": 1.8244474455131792e-05, "loss": 0.7201, "step": 2490 }, { "epoch": 0.6308750236578133, "grad_norm": 0.44335806369781494, "learning_rate": 1.8032627295354183e-05, "loss": 0.7117, "step": 2500 }, { "epoch": 0.6333985237524447, "grad_norm": 0.5342260003089905, "learning_rate": 1.7821321197421837e-05, "loss": 0.7597, "step": 2510 }, { "epoch": 0.6359220238470759, "grad_norm": 0.46154212951660156, "learning_rate": 1.761057257062876e-05, "loss": 0.7391, "step": 2520 }, { "epoch": 0.6384455239417072, "grad_norm": 0.4077001214027405, "learning_rate": 1.740039778097772e-05, "loss": 0.6893, "step": 2530 }, { "epoch": 0.6409690240363384, "grad_norm": 0.4284549057483673, "learning_rate": 1.7190813149909274e-05, "loss": 0.7196, "step": 2540 }, { "epoch": 0.6434925241309697, "grad_norm": 0.4528105556964874, "learning_rate": 1.6981834953034344e-05, "loss": 0.7245, "step": 2550 }, { "epoch": 0.6460160242256009, "grad_norm": 0.46201303601264954, "learning_rate": 1.677347941887028e-05, "loss": 0.6907, "step": 2560 }, { "epoch": 0.6485395243202322, "grad_norm": 0.4665580987930298, "learning_rate": 1.656576272758061e-05, "loss": 0.7494, "step": 2570 }, { "epoch": 0.6510630244148634, "grad_norm": 0.44148868322372437, "learning_rate": 1.6358701009718577e-05, "loss": 0.7224, "step": 2580 }, { "epoch": 0.6535865245094946, "grad_norm": 0.41278114914894104, "learning_rate": 1.615231034497444e-05, "loss": 0.7381, "step": 2590 }, { "epoch": 0.6561100246041259, "grad_norm": 0.4944005012512207, "learning_rate": 1.5946606760926865e-05, "loss": 0.7103, "step": 2600 }, { "epoch": 0.6586335246987571, "grad_norm": 0.5193942785263062, "learning_rate": 1.574160623179816e-05, "loss": 0.7289, "step": 2610 }, { "epoch": 0.6611570247933884, "grad_norm": 0.47881239652633667, "learning_rate": 1.553732467721392e-05, "loss": 0.6884, "step": 2620 }, { "epoch": 0.6636805248880197, "grad_norm": 0.3865389823913574, "learning_rate": 1.5333777960966616e-05, "loss": 0.6947, "step": 2630 }, { "epoch": 0.666204024982651, "grad_norm": 0.41780009865760803, "learning_rate": 1.5130981889783795e-05, "loss": 0.7063, "step": 2640 }, { "epoch": 0.6687275250772822, "grad_norm": 0.4426202178001404, "learning_rate": 1.4928952212100483e-05, "loss": 0.7349, "step": 2650 }, { "epoch": 0.6712510251719135, "grad_norm": 0.4114045202732086, "learning_rate": 1.4727704616836296e-05, "loss": 0.7294, "step": 2660 }, { "epoch": 0.6737745252665447, "grad_norm": 0.44464120268821716, "learning_rate": 1.4527254732177043e-05, "loss": 0.717, "step": 2670 }, { "epoch": 0.676298025361176, "grad_norm": 0.4169268012046814, "learning_rate": 1.4327618124361114e-05, "loss": 0.722, "step": 2680 }, { "epoch": 0.6788215254558072, "grad_norm": 0.42548105120658875, "learning_rate": 1.412881029647065e-05, "loss": 0.687, "step": 2690 }, { "epoch": 0.6813450255504384, "grad_norm": 0.4200035631656647, "learning_rate": 1.3930846687227664e-05, "loss": 0.7011, "step": 2700 }, { "epoch": 0.6838685256450697, "grad_norm": 0.43831759691238403, "learning_rate": 1.3733742669795049e-05, "loss": 0.7447, "step": 2710 }, { "epoch": 0.6863920257397009, "grad_norm": 0.4923849403858185, "learning_rate": 1.3537513550582853e-05, "loss": 0.7263, "step": 2720 }, { "epoch": 0.6889155258343322, "grad_norm": 0.4158725142478943, "learning_rate": 1.3342174568059527e-05, "loss": 0.7009, "step": 2730 }, { "epoch": 0.6914390259289634, "grad_norm": 0.4092692732810974, "learning_rate": 1.3147740891568661e-05, "loss": 0.6921, "step": 2740 }, { "epoch": 0.6939625260235948, "grad_norm": 0.43437299132347107, "learning_rate": 1.2954227620150904e-05, "loss": 0.7377, "step": 2750 }, { "epoch": 0.696486026118226, "grad_norm": 0.45701101422309875, "learning_rate": 1.2761649781371479e-05, "loss": 0.7088, "step": 2760 }, { "epoch": 0.6990095262128573, "grad_norm": 0.42999330163002014, "learning_rate": 1.257002233015318e-05, "loss": 0.7286, "step": 2770 }, { "epoch": 0.7015330263074885, "grad_norm": 0.4685601592063904, "learning_rate": 1.2379360147614994e-05, "loss": 0.7034, "step": 2780 }, { "epoch": 0.7040565264021197, "grad_norm": 0.4414807856082916, "learning_rate": 1.2189678039916532e-05, "loss": 0.6826, "step": 2790 }, { "epoch": 0.706580026496751, "grad_norm": 0.4224618673324585, "learning_rate": 1.2000990737108225e-05, "loss": 0.6999, "step": 2800 }, { "epoch": 0.7091035265913822, "grad_norm": 0.5020971298217773, "learning_rate": 1.1813312891987392e-05, "loss": 0.734, "step": 2810 }, { "epoch": 0.7116270266860135, "grad_norm": 0.4721106290817261, "learning_rate": 1.1626659078960424e-05, "loss": 0.7062, "step": 2820 }, { "epoch": 0.7141505267806447, "grad_norm": 0.43608254194259644, "learning_rate": 1.1441043792910936e-05, "loss": 0.7132, "step": 2830 }, { "epoch": 0.716674026875276, "grad_norm": 0.47143277525901794, "learning_rate": 1.1256481448074179e-05, "loss": 0.7037, "step": 2840 }, { "epoch": 0.7191975269699072, "grad_norm": 0.44380879402160645, "learning_rate": 1.1072986376917638e-05, "loss": 0.7114, "step": 2850 }, { "epoch": 0.7217210270645386, "grad_norm": 0.41831108927726746, "learning_rate": 1.0890572829028087e-05, "loss": 0.7452, "step": 2860 }, { "epoch": 0.7242445271591698, "grad_norm": 0.4085826575756073, "learning_rate": 1.0709254970004937e-05, "loss": 0.6973, "step": 2870 }, { "epoch": 0.7267680272538011, "grad_norm": 0.39444825053215027, "learning_rate": 1.0529046880360263e-05, "loss": 0.6945, "step": 2880 }, { "epoch": 0.7292915273484323, "grad_norm": 0.3804919421672821, "learning_rate": 1.034996255442529e-05, "loss": 0.7419, "step": 2890 }, { "epoch": 0.7318150274430635, "grad_norm": 0.3947986364364624, "learning_rate": 1.0172015899263712e-05, "loss": 0.7092, "step": 2900 }, { "epoch": 0.7343385275376948, "grad_norm": 0.4076451361179352, "learning_rate": 9.995220733591639e-06, "loss": 0.6974, "step": 2910 }, { "epoch": 0.736862027632326, "grad_norm": 0.39703309535980225, "learning_rate": 9.819590786704572e-06, "loss": 0.7296, "step": 2920 }, { "epoch": 0.7393855277269573, "grad_norm": 0.4251289665699005, "learning_rate": 9.645139697411149e-06, "loss": 0.7106, "step": 2930 }, { "epoch": 0.7419090278215885, "grad_norm": 0.44590967893600464, "learning_rate": 9.471881012974071e-06, "loss": 0.7285, "step": 2940 }, { "epoch": 0.7444325279162198, "grad_norm": 0.46388325095176697, "learning_rate": 9.299828188058013e-06, "loss": 0.7135, "step": 2950 }, { "epoch": 0.746956028010851, "grad_norm": 0.431956946849823, "learning_rate": 9.128994583684838e-06, "loss": 0.695, "step": 2960 }, { "epoch": 0.7494795281054824, "grad_norm": 0.41202619671821594, "learning_rate": 8.959393466195972e-06, "loss": 0.7424, "step": 2970 }, { "epoch": 0.7520030282001136, "grad_norm": 0.41577231884002686, "learning_rate": 8.791038006222233e-06, "loss": 0.7222, "step": 2980 }, { "epoch": 0.7545265282947448, "grad_norm": 0.4451634883880615, "learning_rate": 8.623941277660994e-06, "loss": 0.7017, "step": 2990 }, { "epoch": 0.7570500283893761, "grad_norm": 0.4805489182472229, "learning_rate": 8.458116256660981e-06, "loss": 0.7407, "step": 3000 }, { "epoch": 0.7595735284840073, "grad_norm": 0.415554404258728, "learning_rate": 8.293575820614508e-06, "loss": 0.7122, "step": 3010 }, { "epoch": 0.7620970285786386, "grad_norm": 0.42586177587509155, "learning_rate": 8.130332747157542e-06, "loss": 0.6742, "step": 3020 }, { "epoch": 0.7646205286732698, "grad_norm": 0.5801184177398682, "learning_rate": 7.968399713177366e-06, "loss": 0.7278, "step": 3030 }, { "epoch": 0.7671440287679011, "grad_norm": 0.4051037132740021, "learning_rate": 7.807789293828204e-06, "loss": 0.7334, "step": 3040 }, { "epoch": 0.7696675288625323, "grad_norm": 0.44045910239219666, "learning_rate": 7.648513961554607e-06, "loss": 0.6804, "step": 3050 }, { "epoch": 0.7721910289571636, "grad_norm": 0.3936316967010498, "learning_rate": 7.4905860851229605e-06, "loss": 0.7251, "step": 3060 }, { "epoch": 0.7747145290517948, "grad_norm": 0.4889877736568451, "learning_rate": 7.334017928660902e-06, "loss": 0.7208, "step": 3070 }, { "epoch": 0.7772380291464261, "grad_norm": 0.4045318067073822, "learning_rate": 7.1788216507049865e-06, "loss": 0.7035, "step": 3080 }, { "epoch": 0.7797615292410573, "grad_norm": 0.45931798219680786, "learning_rate": 7.0250093032564494e-06, "loss": 0.672, "step": 3090 }, { "epoch": 0.7822850293356886, "grad_norm": 0.42762941122055054, "learning_rate": 6.872592830845339e-06, "loss": 0.7027, "step": 3100 }, { "epoch": 0.7848085294303199, "grad_norm": 0.4001401662826538, "learning_rate": 6.72158406960289e-06, "loss": 0.7308, "step": 3110 }, { "epoch": 0.7873320295249511, "grad_norm": 0.3829699158668518, "learning_rate": 6.571994746342439e-06, "loss": 0.7069, "step": 3120 }, { "epoch": 0.7898555296195824, "grad_norm": 0.4204135537147522, "learning_rate": 6.4238364776486785e-06, "loss": 0.6928, "step": 3130 }, { "epoch": 0.7923790297142136, "grad_norm": 0.4242267608642578, "learning_rate": 6.277120768975644e-06, "loss": 0.7346, "step": 3140 }, { "epoch": 0.7949025298088449, "grad_norm": 0.49215856194496155, "learning_rate": 6.131859013753155e-06, "loss": 0.6918, "step": 3150 }, { "epoch": 0.7974260299034761, "grad_norm": 0.45208001136779785, "learning_rate": 5.988062492502117e-06, "loss": 0.6806, "step": 3160 }, { "epoch": 0.7999495299981074, "grad_norm": 0.39334243535995483, "learning_rate": 5.8457423719584435e-06, "loss": 0.6936, "step": 3170 }, { "epoch": 0.8024730300927386, "grad_norm": 0.4453311860561371, "learning_rate": 5.704909704205949e-06, "loss": 0.7286, "step": 3180 }, { "epoch": 0.8049965301873698, "grad_norm": 0.446539968252182, "learning_rate": 5.565575425818054e-06, "loss": 0.7129, "step": 3190 }, { "epoch": 0.8075200302820011, "grad_norm": 0.4576982259750366, "learning_rate": 5.427750357008468e-06, "loss": 0.6787, "step": 3200 }, { "epoch": 0.8100435303766323, "grad_norm": 0.43956705927848816, "learning_rate": 5.291445200790982e-06, "loss": 0.7249, "step": 3210 }, { "epoch": 0.8125670304712637, "grad_norm": 0.3920847773551941, "learning_rate": 5.156670542148267e-06, "loss": 0.7237, "step": 3220 }, { "epoch": 0.8150905305658949, "grad_norm": 0.4733986556529999, "learning_rate": 5.023436847209887e-06, "loss": 0.7165, "step": 3230 }, { "epoch": 0.8176140306605262, "grad_norm": 0.39961108565330505, "learning_rate": 4.891754462439557e-06, "loss": 0.6806, "step": 3240 }, { "epoch": 0.8201375307551574, "grad_norm": 0.472107470035553, "learning_rate": 4.761633613831645e-06, "loss": 0.7165, "step": 3250 }, { "epoch": 0.8226610308497887, "grad_norm": 0.48432740569114685, "learning_rate": 4.6330844061170914e-06, "loss": 0.6869, "step": 3260 }, { "epoch": 0.8251845309444199, "grad_norm": 0.4450688064098358, "learning_rate": 4.506116821978662e-06, "loss": 0.6968, "step": 3270 }, { "epoch": 0.8277080310390512, "grad_norm": 0.3828079402446747, "learning_rate": 4.380740721275786e-06, "loss": 0.7258, "step": 3280 }, { "epoch": 0.8302315311336824, "grad_norm": 0.5289183259010315, "learning_rate": 4.25696584027882e-06, "loss": 0.7084, "step": 3290 }, { "epoch": 0.8327550312283136, "grad_norm": 0.4827065169811249, "learning_rate": 4.134801790913006e-06, "loss": 0.6704, "step": 3300 }, { "epoch": 0.8352785313229449, "grad_norm": 0.5083085298538208, "learning_rate": 4.014258060012005e-06, "loss": 0.7284, "step": 3310 }, { "epoch": 0.8378020314175761, "grad_norm": 0.38500750064849854, "learning_rate": 3.895344008581222e-06, "loss": 0.6969, "step": 3320 }, { "epoch": 0.8403255315122075, "grad_norm": 0.4650344252586365, "learning_rate": 3.7780688710708223e-06, "loss": 0.6856, "step": 3330 }, { "epoch": 0.8428490316068387, "grad_norm": 0.4428897202014923, "learning_rate": 3.6624417546586574e-06, "loss": 0.6628, "step": 3340 }, { "epoch": 0.84537253170147, "grad_norm": 0.4858897626399994, "learning_rate": 3.548471638542991e-06, "loss": 0.726, "step": 3350 }, { "epoch": 0.8478960317961012, "grad_norm": 0.4039098024368286, "learning_rate": 3.436167373245247e-06, "loss": 0.7078, "step": 3360 }, { "epoch": 0.8504195318907325, "grad_norm": 0.43264418840408325, "learning_rate": 3.325537679922672e-06, "loss": 0.7173, "step": 3370 }, { "epoch": 0.8529430319853637, "grad_norm": 0.48531806468963623, "learning_rate": 3.2165911496911173e-06, "loss": 0.6934, "step": 3380 }, { "epoch": 0.8554665320799949, "grad_norm": 0.4327000081539154, "learning_rate": 3.1093362429578414e-06, "loss": 0.7103, "step": 3390 }, { "epoch": 0.8579900321746262, "grad_norm": 0.3971143066883087, "learning_rate": 3.0037812887645483e-06, "loss": 0.728, "step": 3400 }, { "epoch": 0.8605135322692574, "grad_norm": 0.3819441497325897, "learning_rate": 2.8999344841405373e-06, "loss": 0.6971, "step": 3410 }, { "epoch": 0.8630370323638887, "grad_norm": 0.4157625734806061, "learning_rate": 2.7978038934662024e-06, "loss": 0.6919, "step": 3420 }, { "epoch": 0.8655605324585199, "grad_norm": 0.43926429748535156, "learning_rate": 2.697397447846725e-06, "loss": 0.7088, "step": 3430 }, { "epoch": 0.8680840325531513, "grad_norm": 0.4086776673793793, "learning_rate": 2.5987229444962237e-06, "loss": 0.723, "step": 3440 }, { "epoch": 0.8706075326477825, "grad_norm": 0.4299987256526947, "learning_rate": 2.501788046132203e-06, "loss": 0.7267, "step": 3450 }, { "epoch": 0.8731310327424138, "grad_norm": 0.4047602713108063, "learning_rate": 2.4066002803805386e-06, "loss": 0.7044, "step": 3460 }, { "epoch": 0.875654532837045, "grad_norm": 0.47258859872817993, "learning_rate": 2.313167039190861e-06, "loss": 0.7114, "step": 3470 }, { "epoch": 0.8781780329316763, "grad_norm": 0.4289475977420807, "learning_rate": 2.2214955782625752e-06, "loss": 0.7081, "step": 3480 }, { "epoch": 0.8807015330263075, "grad_norm": 0.4776837229728699, "learning_rate": 2.1315930164813507e-06, "loss": 0.7302, "step": 3490 }, { "epoch": 0.8832250331209387, "grad_norm": 0.4333973228931427, "learning_rate": 2.0434663353663536e-06, "loss": 0.7042, "step": 3500 }, { "epoch": 0.88574853321557, "grad_norm": 0.46293777227401733, "learning_rate": 1.9571223785280314e-06, "loss": 0.7028, "step": 3510 }, { "epoch": 0.8882720333102012, "grad_norm": 0.4550043046474457, "learning_rate": 1.8725678511367001e-06, "loss": 0.7129, "step": 3520 }, { "epoch": 0.8907955334048325, "grad_norm": 0.43017250299453735, "learning_rate": 1.789809319401825e-06, "loss": 0.7154, "step": 3530 }, { "epoch": 0.8933190334994637, "grad_norm": 0.4087255895137787, "learning_rate": 1.7088532100621224e-06, "loss": 0.6848, "step": 3540 }, { "epoch": 0.895842533594095, "grad_norm": 0.40603166818618774, "learning_rate": 1.629705809886467e-06, "loss": 0.688, "step": 3550 }, { "epoch": 0.8983660336887263, "grad_norm": 0.3984641432762146, "learning_rate": 1.5523732651857082e-06, "loss": 0.708, "step": 3560 }, { "epoch": 0.9008895337833576, "grad_norm": 0.457427054643631, "learning_rate": 1.4768615813353398e-06, "loss": 0.7051, "step": 3570 }, { "epoch": 0.9034130338779888, "grad_norm": 0.3801426887512207, "learning_rate": 1.4031766223091603e-06, "loss": 0.6992, "step": 3580 }, { "epoch": 0.90593653397262, "grad_norm": 0.39176779985427856, "learning_rate": 1.3313241102239054e-06, "loss": 0.6811, "step": 3590 }, { "epoch": 0.9084600340672513, "grad_norm": 0.41780802607536316, "learning_rate": 1.261309624894863e-06, "loss": 0.7203, "step": 3600 }, { "epoch": 0.9109835341618825, "grad_norm": 0.3884226679801941, "learning_rate": 1.1931386034025882e-06, "loss": 0.7038, "step": 3610 }, { "epoch": 0.9135070342565138, "grad_norm": 0.3755139410495758, "learning_rate": 1.1268163396706583e-06, "loss": 0.7193, "step": 3620 }, { "epoch": 0.916030534351145, "grad_norm": 0.3906721770763397, "learning_rate": 1.0623479840545874e-06, "loss": 0.6594, "step": 3630 }, { "epoch": 0.9185540344457763, "grad_norm": 0.4476547837257385, "learning_rate": 9.997385429418555e-07, "loss": 0.7282, "step": 3640 }, { "epoch": 0.9210775345404075, "grad_norm": 0.4512433707714081, "learning_rate": 9.389928783631207e-07, "loss": 0.7275, "step": 3650 }, { "epoch": 0.9236010346350388, "grad_norm": 0.4227374494075775, "learning_rate": 8.801157076146705e-07, "loss": 0.6884, "step": 3660 }, { "epoch": 0.92612453472967, "grad_norm": 0.46542125940322876, "learning_rate": 8.231116028920765e-07, "loss": 0.6995, "step": 3670 }, { "epoch": 0.9286480348243014, "grad_norm": 0.4018099308013916, "learning_rate": 7.679849909351472e-07, "loss": 0.6846, "step": 3680 }, { "epoch": 0.9311715349189326, "grad_norm": 0.46164804697036743, "learning_rate": 7.147401526841485e-07, "loss": 0.6821, "step": 3690 }, { "epoch": 0.9336950350135638, "grad_norm": 0.44453224539756775, "learning_rate": 6.633812229473791e-07, "loss": 0.7343, "step": 3700 }, { "epoch": 0.9362185351081951, "grad_norm": 0.4445902109146118, "learning_rate": 6.139121900800515e-07, "loss": 0.683, "step": 3710 }, { "epoch": 0.9387420352028263, "grad_norm": 0.4326709508895874, "learning_rate": 5.663368956745963e-07, "loss": 0.6887, "step": 3720 }, { "epoch": 0.9412655352974576, "grad_norm": 0.4491145610809326, "learning_rate": 5.206590342623164e-07, "loss": 0.6878, "step": 3730 }, { "epoch": 0.9437890353920888, "grad_norm": 0.42220959067344666, "learning_rate": 4.768821530264977e-07, "loss": 0.7293, "step": 3740 }, { "epoch": 0.9463125354867201, "grad_norm": 0.43391677737236023, "learning_rate": 4.350096515269325e-07, "loss": 0.7175, "step": 3750 }, { "epoch": 0.9488360355813513, "grad_norm": 0.43437376618385315, "learning_rate": 3.950447814359409e-07, "loss": 0.7162, "step": 3760 }, { "epoch": 0.9513595356759826, "grad_norm": 0.4475696086883545, "learning_rate": 3.5699064628583745e-07, "loss": 0.6984, "step": 3770 }, { "epoch": 0.9538830357706138, "grad_norm": 0.38860946893692017, "learning_rate": 3.2085020122793186e-07, "loss": 0.701, "step": 3780 }, { "epoch": 0.956406535865245, "grad_norm": 0.41958314180374146, "learning_rate": 2.8662625280304613e-07, "loss": 0.6944, "step": 3790 }, { "epoch": 0.9589300359598764, "grad_norm": 0.4484213590621948, "learning_rate": 2.5432145872355816e-07, "loss": 0.6876, "step": 3800 }, { "epoch": 0.9614535360545076, "grad_norm": 0.4424046277999878, "learning_rate": 2.2393832766701706e-07, "loss": 0.7071, "step": 3810 }, { "epoch": 0.9639770361491389, "grad_norm": 0.38398098945617676, "learning_rate": 1.9547921908133483e-07, "loss": 0.7069, "step": 3820 }, { "epoch": 0.9665005362437701, "grad_norm": 0.46120497584342957, "learning_rate": 1.689463430015442e-07, "loss": 0.6891, "step": 3830 }, { "epoch": 0.9690240363384014, "grad_norm": 0.4390755295753479, "learning_rate": 1.443417598781971e-07, "loss": 0.7045, "step": 3840 }, { "epoch": 0.9715475364330326, "grad_norm": 0.45651644468307495, "learning_rate": 1.2166738041733684e-07, "loss": 0.7043, "step": 3850 }, { "epoch": 0.9740710365276639, "grad_norm": 0.4153136610984802, "learning_rate": 1.0092496543212814e-07, "loss": 0.6998, "step": 3860 }, { "epoch": 0.9765945366222951, "grad_norm": 0.40442293882369995, "learning_rate": 8.211612570611926e-08, "loss": 0.6918, "step": 3870 }, { "epoch": 0.9791180367169264, "grad_norm": 0.4135897159576416, "learning_rate": 6.524232186815305e-08, "loss": 0.6865, "step": 3880 }, { "epoch": 0.9816415368115576, "grad_norm": 0.43170320987701416, "learning_rate": 5.03048642789411e-08, "loss": 0.6825, "step": 3890 }, { "epoch": 0.9841650369061888, "grad_norm": 0.5247434973716736, "learning_rate": 3.730491292930072e-08, "loss": 0.6978, "step": 3900 }, { "epoch": 0.9866885370008202, "grad_norm": 0.5228101015090942, "learning_rate": 2.624347735007693e-08, "loss": 0.7135, "step": 3910 }, { "epoch": 0.9892120370954514, "grad_norm": 0.4438874125480652, "learning_rate": 1.7121416533749658e-08, "loss": 0.7048, "step": 3920 }, { "epoch": 0.9917355371900827, "grad_norm": 0.40782466530799866, "learning_rate": 9.939438867723194e-09, "loss": 0.705, "step": 3930 }, { "epoch": 0.9942590372847139, "grad_norm": 0.4269384443759918, "learning_rate": 4.6981020793118725e-09, "loss": 0.7149, "step": 3940 }, { "epoch": 0.9967825373793452, "grad_norm": 0.4239284098148346, "learning_rate": 1.3978131924385906e-09, "loss": 0.6901, "step": 3950 }, { "epoch": 0.9993060374739764, "grad_norm": 0.4434085488319397, "learning_rate": 3.88284960184393e-11, "loss": 0.7212, "step": 3960 } ], "logging_steps": 10, "max_steps": 3962, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.1480232846728102e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }