Muedi's picture
Training in progress, step 21500, checkpoint
b40746f verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.2100942981384668,
"eval_steps": 500,
"global_step": 21500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0009771827820393806,
"grad_norm": 0.5417118072509766,
"learning_rate": 4.995602247740044e-05,
"loss": 1.378,
"step": 100
},
{
"epoch": 0.001954365564078761,
"grad_norm": 0.6493918895721436,
"learning_rate": 4.990715856340093e-05,
"loss": 1.3304,
"step": 200
},
{
"epoch": 0.0029315483461181415,
"grad_norm": 0.9062462449073792,
"learning_rate": 4.9858294649401425e-05,
"loss": 1.3284,
"step": 300
},
{
"epoch": 0.003908731128157522,
"grad_norm": 0.750052273273468,
"learning_rate": 4.9809430735401906e-05,
"loss": 1.3166,
"step": 400
},
{
"epoch": 0.004885913910196903,
"grad_norm": 0.6602022051811218,
"learning_rate": 4.97605668214024e-05,
"loss": 1.3166,
"step": 500
},
{
"epoch": 0.005863096692236283,
"grad_norm": 0.4193927049636841,
"learning_rate": 4.971170290740288e-05,
"loss": 1.3098,
"step": 600
},
{
"epoch": 0.006840279474275663,
"grad_norm": 0.6095415949821472,
"learning_rate": 4.966283899340338e-05,
"loss": 1.3103,
"step": 700
},
{
"epoch": 0.007817462256315045,
"grad_norm": 0.9943467378616333,
"learning_rate": 4.9613975079403865e-05,
"loss": 1.3096,
"step": 800
},
{
"epoch": 0.008794645038354424,
"grad_norm": 1.2263585329055786,
"learning_rate": 4.9565111165404346e-05,
"loss": 1.3067,
"step": 900
},
{
"epoch": 0.009771827820393805,
"grad_norm": 0.7198677659034729,
"learning_rate": 4.951624725140484e-05,
"loss": 1.3041,
"step": 1000
},
{
"epoch": 0.010749010602433185,
"grad_norm": 0.7370775938034058,
"learning_rate": 4.946738333740533e-05,
"loss": 1.302,
"step": 1100
},
{
"epoch": 0.011726193384472566,
"grad_norm": 0.5109437704086304,
"learning_rate": 4.941851942340582e-05,
"loss": 1.3089,
"step": 1200
},
{
"epoch": 0.012703376166511945,
"grad_norm": 0.1879555583000183,
"learning_rate": 4.9369655509406305e-05,
"loss": 1.3043,
"step": 1300
},
{
"epoch": 0.013680558948551327,
"grad_norm": 0.951046884059906,
"learning_rate": 4.932079159540679e-05,
"loss": 1.3098,
"step": 1400
},
{
"epoch": 0.014657741730590706,
"grad_norm": 0.2478829026222229,
"learning_rate": 4.927192768140728e-05,
"loss": 1.3026,
"step": 1500
},
{
"epoch": 0.01563492451263009,
"grad_norm": 0.5585843324661255,
"learning_rate": 4.9223063767407776e-05,
"loss": 1.3014,
"step": 1600
},
{
"epoch": 0.016612107294669467,
"grad_norm": 0.48532453179359436,
"learning_rate": 4.917419985340826e-05,
"loss": 1.2981,
"step": 1700
},
{
"epoch": 0.017589290076708848,
"grad_norm": 0.4233573079109192,
"learning_rate": 4.912533593940875e-05,
"loss": 1.2992,
"step": 1800
},
{
"epoch": 0.01856647285874823,
"grad_norm": 0.3272475600242615,
"learning_rate": 4.9076472025409234e-05,
"loss": 1.292,
"step": 1900
},
{
"epoch": 0.01954365564078761,
"grad_norm": 0.5299385786056519,
"learning_rate": 4.902760811140973e-05,
"loss": 1.2963,
"step": 2000
},
{
"epoch": 0.02052083842282699,
"grad_norm": 0.1614024043083191,
"learning_rate": 4.8978744197410216e-05,
"loss": 1.2945,
"step": 2100
},
{
"epoch": 0.02149802120486637,
"grad_norm": 0.6039963960647583,
"learning_rate": 4.8929880283410705e-05,
"loss": 1.2913,
"step": 2200
},
{
"epoch": 0.02247520398690575,
"grad_norm": 0.5772804021835327,
"learning_rate": 4.888101636941119e-05,
"loss": 1.2895,
"step": 2300
},
{
"epoch": 0.023452386768945132,
"grad_norm": 0.7489622235298157,
"learning_rate": 4.883215245541168e-05,
"loss": 1.2847,
"step": 2400
},
{
"epoch": 0.024429569550984513,
"grad_norm": 0.30208253860473633,
"learning_rate": 4.878328854141217e-05,
"loss": 1.2924,
"step": 2500
},
{
"epoch": 0.02540675233302389,
"grad_norm": 0.36944472789764404,
"learning_rate": 4.873442462741266e-05,
"loss": 1.2916,
"step": 2600
},
{
"epoch": 0.026383935115063272,
"grad_norm": 0.3268676698207855,
"learning_rate": 4.8685560713413145e-05,
"loss": 1.2893,
"step": 2700
},
{
"epoch": 0.027361117897102653,
"grad_norm": 0.2795974910259247,
"learning_rate": 4.863669679941363e-05,
"loss": 1.282,
"step": 2800
},
{
"epoch": 0.028338300679142035,
"grad_norm": 0.36298853158950806,
"learning_rate": 4.858783288541413e-05,
"loss": 1.2832,
"step": 2900
},
{
"epoch": 0.029315483461181412,
"grad_norm": 0.5242423415184021,
"learning_rate": 4.853896897141461e-05,
"loss": 1.2819,
"step": 3000
},
{
"epoch": 0.030292666243220794,
"grad_norm": 0.25340864062309265,
"learning_rate": 4.8490105057415104e-05,
"loss": 1.2809,
"step": 3100
},
{
"epoch": 0.03126984902526018,
"grad_norm": 0.7241976261138916,
"learning_rate": 4.844124114341559e-05,
"loss": 1.2802,
"step": 3200
},
{
"epoch": 0.032247031807299556,
"grad_norm": 0.5154001712799072,
"learning_rate": 4.839237722941608e-05,
"loss": 1.2748,
"step": 3300
},
{
"epoch": 0.033224214589338934,
"grad_norm": 0.5323473811149597,
"learning_rate": 4.834351331541657e-05,
"loss": 1.284,
"step": 3400
},
{
"epoch": 0.03420139737137832,
"grad_norm": 0.3947168290615082,
"learning_rate": 4.8294649401417056e-05,
"loss": 1.276,
"step": 3500
},
{
"epoch": 0.035178580153417696,
"grad_norm": 0.4776057302951813,
"learning_rate": 4.8245785487417544e-05,
"loss": 1.2783,
"step": 3600
},
{
"epoch": 0.036155762935457074,
"grad_norm": 0.4884164035320282,
"learning_rate": 4.819692157341804e-05,
"loss": 1.2745,
"step": 3700
},
{
"epoch": 0.03713294571749646,
"grad_norm": 0.5210428833961487,
"learning_rate": 4.814805765941852e-05,
"loss": 1.2707,
"step": 3800
},
{
"epoch": 0.038110128499535836,
"grad_norm": 0.46214359998703003,
"learning_rate": 4.809919374541901e-05,
"loss": 1.2727,
"step": 3900
},
{
"epoch": 0.03908731128157522,
"grad_norm": 0.2656782865524292,
"learning_rate": 4.8050329831419496e-05,
"loss": 1.2694,
"step": 4000
},
{
"epoch": 0.0400644940636146,
"grad_norm": 0.4923059940338135,
"learning_rate": 4.8001465917419985e-05,
"loss": 1.2665,
"step": 4100
},
{
"epoch": 0.04104167684565398,
"grad_norm": 0.92928147315979,
"learning_rate": 4.795260200342048e-05,
"loss": 1.2627,
"step": 4200
},
{
"epoch": 0.04201885962769336,
"grad_norm": 1.0651229619979858,
"learning_rate": 4.790373808942096e-05,
"loss": 1.2623,
"step": 4300
},
{
"epoch": 0.04299604240973274,
"grad_norm": 0.9612557888031006,
"learning_rate": 4.7854874175421456e-05,
"loss": 1.2482,
"step": 4400
},
{
"epoch": 0.043973225191772124,
"grad_norm": 1.0120874643325806,
"learning_rate": 4.7806010261421944e-05,
"loss": 1.2589,
"step": 4500
},
{
"epoch": 0.0449504079738115,
"grad_norm": 0.6250020861625671,
"learning_rate": 4.775714634742243e-05,
"loss": 1.2499,
"step": 4600
},
{
"epoch": 0.04592759075585088,
"grad_norm": 0.2850038707256317,
"learning_rate": 4.770828243342292e-05,
"loss": 1.2446,
"step": 4700
},
{
"epoch": 0.046904773537890264,
"grad_norm": 1.2032625675201416,
"learning_rate": 4.765941851942341e-05,
"loss": 1.2238,
"step": 4800
},
{
"epoch": 0.04788195631992964,
"grad_norm": 0.42024949193000793,
"learning_rate": 4.7610554605423896e-05,
"loss": 1.2255,
"step": 4900
},
{
"epoch": 0.048859139101969026,
"grad_norm": 0.7451406121253967,
"learning_rate": 4.756169069142439e-05,
"loss": 1.2071,
"step": 5000
},
{
"epoch": 0.049836321884008404,
"grad_norm": 0.8735096454620361,
"learning_rate": 4.751282677742487e-05,
"loss": 1.2126,
"step": 5100
},
{
"epoch": 0.05081350466604778,
"grad_norm": 0.73675137758255,
"learning_rate": 4.746396286342537e-05,
"loss": 1.2036,
"step": 5200
},
{
"epoch": 0.051790687448087167,
"grad_norm": 0.6540606617927551,
"learning_rate": 4.741509894942585e-05,
"loss": 1.1825,
"step": 5300
},
{
"epoch": 0.052767870230126544,
"grad_norm": 0.825066864490509,
"learning_rate": 4.7366235035426336e-05,
"loss": 1.1655,
"step": 5400
},
{
"epoch": 0.05374505301216593,
"grad_norm": 1.6421219110488892,
"learning_rate": 4.731737112142683e-05,
"loss": 1.1716,
"step": 5500
},
{
"epoch": 0.05472223579420531,
"grad_norm": 1.0644057989120483,
"learning_rate": 4.726850720742731e-05,
"loss": 1.1384,
"step": 5600
},
{
"epoch": 0.055699418576244684,
"grad_norm": 1.1611616611480713,
"learning_rate": 4.721964329342781e-05,
"loss": 1.1499,
"step": 5700
},
{
"epoch": 0.05667660135828407,
"grad_norm": 2.0900723934173584,
"learning_rate": 4.7170779379428295e-05,
"loss": 1.1323,
"step": 5800
},
{
"epoch": 0.05765378414032345,
"grad_norm": 1.0580404996871948,
"learning_rate": 4.712191546542878e-05,
"loss": 1.112,
"step": 5900
},
{
"epoch": 0.058630966922362825,
"grad_norm": 0.6299407482147217,
"learning_rate": 4.707305155142927e-05,
"loss": 1.104,
"step": 6000
},
{
"epoch": 0.05960814970440221,
"grad_norm": 0.6816271543502808,
"learning_rate": 4.702418763742976e-05,
"loss": 1.1128,
"step": 6100
},
{
"epoch": 0.06058533248644159,
"grad_norm": 0.654796302318573,
"learning_rate": 4.697532372343025e-05,
"loss": 1.0942,
"step": 6200
},
{
"epoch": 0.06156251526848097,
"grad_norm": 1.0433884859085083,
"learning_rate": 4.692645980943074e-05,
"loss": 1.0862,
"step": 6300
},
{
"epoch": 0.06253969805052036,
"grad_norm": 0.6256537437438965,
"learning_rate": 4.6877595895431224e-05,
"loss": 1.081,
"step": 6400
},
{
"epoch": 0.06351688083255973,
"grad_norm": 0.8173975348472595,
"learning_rate": 4.682873198143172e-05,
"loss": 1.0767,
"step": 6500
},
{
"epoch": 0.06449406361459911,
"grad_norm": 0.7856473922729492,
"learning_rate": 4.6779868067432206e-05,
"loss": 1.0767,
"step": 6600
},
{
"epoch": 0.0654712463966385,
"grad_norm": 0.6337741017341614,
"learning_rate": 4.6731004153432695e-05,
"loss": 1.0829,
"step": 6700
},
{
"epoch": 0.06644842917867787,
"grad_norm": 0.5813809037208557,
"learning_rate": 4.668214023943318e-05,
"loss": 1.0571,
"step": 6800
},
{
"epoch": 0.06742561196071725,
"grad_norm": 0.4155445992946625,
"learning_rate": 4.6633276325433664e-05,
"loss": 1.0707,
"step": 6900
},
{
"epoch": 0.06840279474275664,
"grad_norm": 0.6730567812919617,
"learning_rate": 4.658441241143416e-05,
"loss": 1.0477,
"step": 7000
},
{
"epoch": 0.06937997752479601,
"grad_norm": 0.8348300457000732,
"learning_rate": 4.653554849743465e-05,
"loss": 1.0644,
"step": 7100
},
{
"epoch": 0.07035716030683539,
"grad_norm": 2.2414326667785645,
"learning_rate": 4.6486684583435135e-05,
"loss": 1.0577,
"step": 7200
},
{
"epoch": 0.07133434308887478,
"grad_norm": 1.6573911905288696,
"learning_rate": 4.643782066943562e-05,
"loss": 1.0836,
"step": 7300
},
{
"epoch": 0.07231152587091415,
"grad_norm": 0.5690039396286011,
"learning_rate": 4.638895675543611e-05,
"loss": 1.0541,
"step": 7400
},
{
"epoch": 0.07328870865295353,
"grad_norm": 0.527215301990509,
"learning_rate": 4.63400928414366e-05,
"loss": 1.0164,
"step": 7500
},
{
"epoch": 0.07426589143499292,
"grad_norm": 0.7997362613677979,
"learning_rate": 4.6291228927437094e-05,
"loss": 1.0447,
"step": 7600
},
{
"epoch": 0.0752430742170323,
"grad_norm": 2.257143259048462,
"learning_rate": 4.6242365013437575e-05,
"loss": 1.0365,
"step": 7700
},
{
"epoch": 0.07622025699907167,
"grad_norm": 0.9132490158081055,
"learning_rate": 4.619350109943807e-05,
"loss": 1.0498,
"step": 7800
},
{
"epoch": 0.07719743978111106,
"grad_norm": 0.5229859948158264,
"learning_rate": 4.614463718543856e-05,
"loss": 1.0342,
"step": 7900
},
{
"epoch": 0.07817462256315044,
"grad_norm": 0.6948792338371277,
"learning_rate": 4.6095773271439046e-05,
"loss": 1.0325,
"step": 8000
},
{
"epoch": 0.07915180534518981,
"grad_norm": 0.8526360988616943,
"learning_rate": 4.6046909357439534e-05,
"loss": 1.0183,
"step": 8100
},
{
"epoch": 0.0801289881272292,
"grad_norm": 1.1457374095916748,
"learning_rate": 4.599804544344002e-05,
"loss": 1.0243,
"step": 8200
},
{
"epoch": 0.08110617090926858,
"grad_norm": 0.9335997700691223,
"learning_rate": 4.594918152944051e-05,
"loss": 1.046,
"step": 8300
},
{
"epoch": 0.08208335369130795,
"grad_norm": 0.8367229700088501,
"learning_rate": 4.5900317615441e-05,
"loss": 1.0176,
"step": 8400
},
{
"epoch": 0.08306053647334734,
"grad_norm": 3.7648801803588867,
"learning_rate": 4.5851453701441486e-05,
"loss": 1.0047,
"step": 8500
},
{
"epoch": 0.08403771925538672,
"grad_norm": 0.5877612829208374,
"learning_rate": 4.5802589787441975e-05,
"loss": 1.0346,
"step": 8600
},
{
"epoch": 0.08501490203742611,
"grad_norm": 0.5145990252494812,
"learning_rate": 4.575372587344246e-05,
"loss": 1.0268,
"step": 8700
},
{
"epoch": 0.08599208481946548,
"grad_norm": 0.9310688376426697,
"learning_rate": 4.570486195944295e-05,
"loss": 1.0109,
"step": 8800
},
{
"epoch": 0.08696926760150486,
"grad_norm": 0.5182886719703674,
"learning_rate": 4.5655998045443445e-05,
"loss": 1.0117,
"step": 8900
},
{
"epoch": 0.08794645038354425,
"grad_norm": 0.4319695234298706,
"learning_rate": 4.560713413144393e-05,
"loss": 1.0053,
"step": 9000
},
{
"epoch": 0.08892363316558362,
"grad_norm": 4.307732582092285,
"learning_rate": 4.555827021744442e-05,
"loss": 1.0151,
"step": 9100
},
{
"epoch": 0.089900815947623,
"grad_norm": 0.46516236662864685,
"learning_rate": 4.550940630344491e-05,
"loss": 0.9945,
"step": 9200
},
{
"epoch": 0.09087799872966239,
"grad_norm": 1.2372952699661255,
"learning_rate": 4.54605423894454e-05,
"loss": 0.9865,
"step": 9300
},
{
"epoch": 0.09185518151170176,
"grad_norm": 0.7494595646858215,
"learning_rate": 4.5411678475445886e-05,
"loss": 0.9824,
"step": 9400
},
{
"epoch": 0.09283236429374114,
"grad_norm": 0.5540333390235901,
"learning_rate": 4.5362814561446374e-05,
"loss": 1.0132,
"step": 9500
},
{
"epoch": 0.09380954707578053,
"grad_norm": 0.48533427715301514,
"learning_rate": 4.531395064744686e-05,
"loss": 1.0173,
"step": 9600
},
{
"epoch": 0.0947867298578199,
"grad_norm": 0.4972572922706604,
"learning_rate": 4.526508673344736e-05,
"loss": 1.0078,
"step": 9700
},
{
"epoch": 0.09576391263985928,
"grad_norm": 0.6748878955841064,
"learning_rate": 4.521622281944784e-05,
"loss": 1.0172,
"step": 9800
},
{
"epoch": 0.09674109542189867,
"grad_norm": 0.5261876583099365,
"learning_rate": 4.5167358905448326e-05,
"loss": 1.0189,
"step": 9900
},
{
"epoch": 0.09771827820393805,
"grad_norm": 0.4164600670337677,
"learning_rate": 4.5118494991448814e-05,
"loss": 0.9978,
"step": 10000
},
{
"epoch": 0.09869546098597742,
"grad_norm": 0.40417763590812683,
"learning_rate": 4.50696310774493e-05,
"loss": 1.0103,
"step": 10100
},
{
"epoch": 0.09967264376801681,
"grad_norm": 0.8591890931129456,
"learning_rate": 4.50207671634498e-05,
"loss": 1.0065,
"step": 10200
},
{
"epoch": 0.10064982655005619,
"grad_norm": 0.5676371455192566,
"learning_rate": 4.497190324945028e-05,
"loss": 1.0089,
"step": 10300
},
{
"epoch": 0.10162700933209556,
"grad_norm": 0.616646945476532,
"learning_rate": 4.492303933545077e-05,
"loss": 0.9897,
"step": 10400
},
{
"epoch": 0.10260419211413495,
"grad_norm": 0.37536484003067017,
"learning_rate": 4.487417542145126e-05,
"loss": 0.9989,
"step": 10500
},
{
"epoch": 0.10358137489617433,
"grad_norm": 0.6801789402961731,
"learning_rate": 4.482531150745175e-05,
"loss": 0.9923,
"step": 10600
},
{
"epoch": 0.1045585576782137,
"grad_norm": 0.5848776698112488,
"learning_rate": 4.477644759345224e-05,
"loss": 0.9919,
"step": 10700
},
{
"epoch": 0.10553574046025309,
"grad_norm": 0.7715157866477966,
"learning_rate": 4.4727583679452725e-05,
"loss": 0.9814,
"step": 10800
},
{
"epoch": 0.10651292324229247,
"grad_norm": 0.8080986142158508,
"learning_rate": 4.4678719765453214e-05,
"loss": 0.9935,
"step": 10900
},
{
"epoch": 0.10749010602433186,
"grad_norm": 0.4375016391277313,
"learning_rate": 4.462985585145371e-05,
"loss": 0.988,
"step": 11000
},
{
"epoch": 0.10846728880637123,
"grad_norm": 0.8055805563926697,
"learning_rate": 4.458099193745419e-05,
"loss": 0.9861,
"step": 11100
},
{
"epoch": 0.10944447158841061,
"grad_norm": 1.1914618015289307,
"learning_rate": 4.4532128023454685e-05,
"loss": 0.9622,
"step": 11200
},
{
"epoch": 0.11042165437045,
"grad_norm": 0.4247540533542633,
"learning_rate": 4.448326410945517e-05,
"loss": 0.9602,
"step": 11300
},
{
"epoch": 0.11139883715248937,
"grad_norm": 0.5454650521278381,
"learning_rate": 4.4434400195455654e-05,
"loss": 0.9696,
"step": 11400
},
{
"epoch": 0.11237601993452875,
"grad_norm": 0.5259748697280884,
"learning_rate": 4.438553628145615e-05,
"loss": 1.0021,
"step": 11500
},
{
"epoch": 0.11335320271656814,
"grad_norm": 0.5165246725082397,
"learning_rate": 4.433667236745663e-05,
"loss": 0.982,
"step": 11600
},
{
"epoch": 0.11433038549860751,
"grad_norm": 0.6768147945404053,
"learning_rate": 4.4287808453457125e-05,
"loss": 0.9398,
"step": 11700
},
{
"epoch": 0.1153075682806469,
"grad_norm": 1.0245041847229004,
"learning_rate": 4.423894453945761e-05,
"loss": 0.9934,
"step": 11800
},
{
"epoch": 0.11628475106268628,
"grad_norm": 0.6241583228111267,
"learning_rate": 4.41900806254581e-05,
"loss": 0.9697,
"step": 11900
},
{
"epoch": 0.11726193384472565,
"grad_norm": 0.4234873652458191,
"learning_rate": 4.414121671145859e-05,
"loss": 0.9723,
"step": 12000
},
{
"epoch": 0.11823911662676503,
"grad_norm": 0.3932545781135559,
"learning_rate": 4.409235279745908e-05,
"loss": 0.9826,
"step": 12100
},
{
"epoch": 0.11921629940880442,
"grad_norm": 1.5067880153656006,
"learning_rate": 4.4043488883459565e-05,
"loss": 0.9581,
"step": 12200
},
{
"epoch": 0.1201934821908438,
"grad_norm": 0.41707366704940796,
"learning_rate": 4.399462496946006e-05,
"loss": 0.9666,
"step": 12300
},
{
"epoch": 0.12117066497288317,
"grad_norm": 1.1278653144836426,
"learning_rate": 4.394576105546054e-05,
"loss": 0.9553,
"step": 12400
},
{
"epoch": 0.12214784775492256,
"grad_norm": 0.350543737411499,
"learning_rate": 4.3896897141461036e-05,
"loss": 0.9422,
"step": 12500
},
{
"epoch": 0.12312503053696194,
"grad_norm": 0.3775838315486908,
"learning_rate": 4.3848033227461524e-05,
"loss": 0.9626,
"step": 12600
},
{
"epoch": 0.12410221331900131,
"grad_norm": 0.8341017365455627,
"learning_rate": 4.379916931346201e-05,
"loss": 0.9289,
"step": 12700
},
{
"epoch": 0.1250793961010407,
"grad_norm": 0.805614173412323,
"learning_rate": 4.37503053994625e-05,
"loss": 0.9474,
"step": 12800
},
{
"epoch": 0.12605657888308008,
"grad_norm": 0.8439397215843201,
"learning_rate": 4.370144148546299e-05,
"loss": 0.9661,
"step": 12900
},
{
"epoch": 0.12703376166511945,
"grad_norm": 1.1272892951965332,
"learning_rate": 4.3652577571463476e-05,
"loss": 0.9514,
"step": 13000
},
{
"epoch": 0.12801094444715885,
"grad_norm": 0.6426375508308411,
"learning_rate": 4.3603713657463965e-05,
"loss": 0.9448,
"step": 13100
},
{
"epoch": 0.12898812722919822,
"grad_norm": 1.3205431699752808,
"learning_rate": 4.355484974346445e-05,
"loss": 0.9511,
"step": 13200
},
{
"epoch": 0.1299653100112376,
"grad_norm": 0.3671954870223999,
"learning_rate": 4.350598582946494e-05,
"loss": 0.9506,
"step": 13300
},
{
"epoch": 0.130942492793277,
"grad_norm": 0.7566332817077637,
"learning_rate": 4.345712191546543e-05,
"loss": 0.9363,
"step": 13400
},
{
"epoch": 0.13191967557531636,
"grad_norm": 0.8800159692764282,
"learning_rate": 4.340825800146592e-05,
"loss": 0.9388,
"step": 13500
},
{
"epoch": 0.13289685835735573,
"grad_norm": 0.7134628891944885,
"learning_rate": 4.335939408746641e-05,
"loss": 0.9162,
"step": 13600
},
{
"epoch": 0.13387404113939513,
"grad_norm": 0.5555543899536133,
"learning_rate": 4.331053017346689e-05,
"loss": 0.9366,
"step": 13700
},
{
"epoch": 0.1348512239214345,
"grad_norm": 0.4485512375831604,
"learning_rate": 4.326166625946739e-05,
"loss": 0.9286,
"step": 13800
},
{
"epoch": 0.13582840670347388,
"grad_norm": 0.8888948559761047,
"learning_rate": 4.3212802345467876e-05,
"loss": 0.943,
"step": 13900
},
{
"epoch": 0.13680558948551327,
"grad_norm": 0.6719749569892883,
"learning_rate": 4.3163938431468364e-05,
"loss": 0.9217,
"step": 14000
},
{
"epoch": 0.13778277226755264,
"grad_norm": 0.695377767086029,
"learning_rate": 4.311507451746885e-05,
"loss": 0.9093,
"step": 14100
},
{
"epoch": 0.13875995504959202,
"grad_norm": 0.5966312885284424,
"learning_rate": 4.306621060346934e-05,
"loss": 0.9195,
"step": 14200
},
{
"epoch": 0.13973713783163141,
"grad_norm": 0.8073310256004333,
"learning_rate": 4.301734668946983e-05,
"loss": 0.9309,
"step": 14300
},
{
"epoch": 0.14071432061367078,
"grad_norm": 0.6303800940513611,
"learning_rate": 4.2968482775470316e-05,
"loss": 0.9458,
"step": 14400
},
{
"epoch": 0.14169150339571016,
"grad_norm": 0.7043970823287964,
"learning_rate": 4.2919618861470804e-05,
"loss": 0.9132,
"step": 14500
},
{
"epoch": 0.14266868617774955,
"grad_norm": 0.9100736379623413,
"learning_rate": 4.287075494747129e-05,
"loss": 0.9296,
"step": 14600
},
{
"epoch": 0.14364586895978892,
"grad_norm": 0.787862241268158,
"learning_rate": 4.282189103347179e-05,
"loss": 0.9643,
"step": 14700
},
{
"epoch": 0.1446230517418283,
"grad_norm": 0.8169028162956238,
"learning_rate": 4.277302711947227e-05,
"loss": 0.9244,
"step": 14800
},
{
"epoch": 0.1456002345238677,
"grad_norm": 0.9544184803962708,
"learning_rate": 4.272416320547276e-05,
"loss": 0.918,
"step": 14900
},
{
"epoch": 0.14657741730590707,
"grad_norm": 0.5325574278831482,
"learning_rate": 4.2675299291473245e-05,
"loss": 0.9273,
"step": 15000
},
{
"epoch": 0.14755460008794646,
"grad_norm": 1.1403323411941528,
"learning_rate": 4.262643537747374e-05,
"loss": 0.9095,
"step": 15100
},
{
"epoch": 0.14853178286998583,
"grad_norm": 1.0411937236785889,
"learning_rate": 4.257757146347423e-05,
"loss": 0.8967,
"step": 15200
},
{
"epoch": 0.1495089656520252,
"grad_norm": 0.630393922328949,
"learning_rate": 4.2528707549474715e-05,
"loss": 0.8883,
"step": 15300
},
{
"epoch": 0.1504861484340646,
"grad_norm": 0.9445775747299194,
"learning_rate": 4.2479843635475204e-05,
"loss": 0.9253,
"step": 15400
},
{
"epoch": 0.15146333121610397,
"grad_norm": 0.5689444541931152,
"learning_rate": 4.243097972147569e-05,
"loss": 0.8983,
"step": 15500
},
{
"epoch": 0.15244051399814335,
"grad_norm": 0.7726677656173706,
"learning_rate": 4.238211580747618e-05,
"loss": 0.9228,
"step": 15600
},
{
"epoch": 0.15341769678018274,
"grad_norm": 0.8260165452957153,
"learning_rate": 4.2333251893476675e-05,
"loss": 0.9202,
"step": 15700
},
{
"epoch": 0.15439487956222211,
"grad_norm": 0.4869302809238434,
"learning_rate": 4.2284387979477156e-05,
"loss": 0.9283,
"step": 15800
},
{
"epoch": 0.15537206234426149,
"grad_norm": 0.5768991708755493,
"learning_rate": 4.2235524065477644e-05,
"loss": 0.9233,
"step": 15900
},
{
"epoch": 0.15634924512630088,
"grad_norm": 0.8856435418128967,
"learning_rate": 4.218666015147814e-05,
"loss": 0.8825,
"step": 16000
},
{
"epoch": 0.15732642790834026,
"grad_norm": 0.5258185267448425,
"learning_rate": 4.213779623747862e-05,
"loss": 0.8834,
"step": 16100
},
{
"epoch": 0.15830361069037963,
"grad_norm": 0.8340526223182678,
"learning_rate": 4.2088932323479115e-05,
"loss": 0.8856,
"step": 16200
},
{
"epoch": 0.15928079347241902,
"grad_norm": 0.4123723804950714,
"learning_rate": 4.2040068409479596e-05,
"loss": 0.8957,
"step": 16300
},
{
"epoch": 0.1602579762544584,
"grad_norm": 0.8336274027824402,
"learning_rate": 4.199120449548009e-05,
"loss": 0.9053,
"step": 16400
},
{
"epoch": 0.16123515903649777,
"grad_norm": 0.7977516055107117,
"learning_rate": 4.194234058148058e-05,
"loss": 0.8698,
"step": 16500
},
{
"epoch": 0.16221234181853716,
"grad_norm": 0.5064985156059265,
"learning_rate": 4.189347666748107e-05,
"loss": 0.8945,
"step": 16600
},
{
"epoch": 0.16318952460057654,
"grad_norm": 0.8241267204284668,
"learning_rate": 4.1844612753481555e-05,
"loss": 0.8875,
"step": 16700
},
{
"epoch": 0.1641667073826159,
"grad_norm": 0.7517113089561462,
"learning_rate": 4.179574883948204e-05,
"loss": 0.8845,
"step": 16800
},
{
"epoch": 0.1651438901646553,
"grad_norm": 0.6297169923782349,
"learning_rate": 4.174688492548253e-05,
"loss": 0.9303,
"step": 16900
},
{
"epoch": 0.16612107294669468,
"grad_norm": 0.5828490257263184,
"learning_rate": 4.1698021011483026e-05,
"loss": 0.8654,
"step": 17000
},
{
"epoch": 0.16709825572873405,
"grad_norm": 0.3038561940193176,
"learning_rate": 4.164915709748351e-05,
"loss": 0.8933,
"step": 17100
},
{
"epoch": 0.16807543851077344,
"grad_norm": 0.8928827047348022,
"learning_rate": 4.1600293183484e-05,
"loss": 0.8509,
"step": 17200
},
{
"epoch": 0.16905262129281282,
"grad_norm": 0.7055086493492126,
"learning_rate": 4.155142926948449e-05,
"loss": 0.8814,
"step": 17300
},
{
"epoch": 0.17002980407485221,
"grad_norm": 0.5377823710441589,
"learning_rate": 4.150256535548497e-05,
"loss": 0.888,
"step": 17400
},
{
"epoch": 0.17100698685689159,
"grad_norm": 0.6319778561592102,
"learning_rate": 4.1453701441485466e-05,
"loss": 0.8575,
"step": 17500
},
{
"epoch": 0.17198416963893096,
"grad_norm": 0.8756042122840881,
"learning_rate": 4.1404837527485954e-05,
"loss": 0.8805,
"step": 17600
},
{
"epoch": 0.17296135242097035,
"grad_norm": 0.5293178558349609,
"learning_rate": 4.135597361348644e-05,
"loss": 0.8471,
"step": 17700
},
{
"epoch": 0.17393853520300973,
"grad_norm": 0.9118284583091736,
"learning_rate": 4.130710969948693e-05,
"loss": 0.8426,
"step": 17800
},
{
"epoch": 0.1749157179850491,
"grad_norm": 1.0211195945739746,
"learning_rate": 4.125824578548742e-05,
"loss": 0.8877,
"step": 17900
},
{
"epoch": 0.1758929007670885,
"grad_norm": 1.4174985885620117,
"learning_rate": 4.120938187148791e-05,
"loss": 0.8731,
"step": 18000
},
{
"epoch": 0.17687008354912787,
"grad_norm": 0.8243415951728821,
"learning_rate": 4.1160517957488395e-05,
"loss": 0.8852,
"step": 18100
},
{
"epoch": 0.17784726633116724,
"grad_norm": 0.8385602235794067,
"learning_rate": 4.111165404348888e-05,
"loss": 0.8361,
"step": 18200
},
{
"epoch": 0.17882444911320663,
"grad_norm": 1.003968358039856,
"learning_rate": 4.106279012948938e-05,
"loss": 0.8738,
"step": 18300
},
{
"epoch": 0.179801631895246,
"grad_norm": 0.7428449988365173,
"learning_rate": 4.101392621548986e-05,
"loss": 0.8563,
"step": 18400
},
{
"epoch": 0.18077881467728538,
"grad_norm": 1.8963735103607178,
"learning_rate": 4.0965062301490354e-05,
"loss": 0.8428,
"step": 18500
},
{
"epoch": 0.18175599745932478,
"grad_norm": 0.6868895888328552,
"learning_rate": 4.091619838749084e-05,
"loss": 0.8727,
"step": 18600
},
{
"epoch": 0.18273318024136415,
"grad_norm": 1.8936256170272827,
"learning_rate": 4.086733447349133e-05,
"loss": 0.9211,
"step": 18700
},
{
"epoch": 0.18371036302340352,
"grad_norm": 1.004941463470459,
"learning_rate": 4.081847055949182e-05,
"loss": 0.8404,
"step": 18800
},
{
"epoch": 0.18468754580544292,
"grad_norm": 1.4084818363189697,
"learning_rate": 4.0769606645492306e-05,
"loss": 0.868,
"step": 18900
},
{
"epoch": 0.1856647285874823,
"grad_norm": 0.6459541320800781,
"learning_rate": 4.0720742731492794e-05,
"loss": 0.8583,
"step": 19000
},
{
"epoch": 0.18664191136952166,
"grad_norm": 0.7335548996925354,
"learning_rate": 4.067187881749328e-05,
"loss": 0.8622,
"step": 19100
},
{
"epoch": 0.18761909415156106,
"grad_norm": 0.6783348321914673,
"learning_rate": 4.062301490349377e-05,
"loss": 0.8572,
"step": 19200
},
{
"epoch": 0.18859627693360043,
"grad_norm": 0.6323419809341431,
"learning_rate": 4.057415098949426e-05,
"loss": 0.8763,
"step": 19300
},
{
"epoch": 0.1895734597156398,
"grad_norm": 0.963927686214447,
"learning_rate": 4.052528707549475e-05,
"loss": 0.8543,
"step": 19400
},
{
"epoch": 0.1905506424976792,
"grad_norm": 0.4785550832748413,
"learning_rate": 4.0476423161495234e-05,
"loss": 0.863,
"step": 19500
},
{
"epoch": 0.19152782527971857,
"grad_norm": 0.6358627080917358,
"learning_rate": 4.042755924749573e-05,
"loss": 0.8842,
"step": 19600
},
{
"epoch": 0.19250500806175797,
"grad_norm": 0.7857956886291504,
"learning_rate": 4.037869533349621e-05,
"loss": 0.8698,
"step": 19700
},
{
"epoch": 0.19348219084379734,
"grad_norm": 0.5225537419319153,
"learning_rate": 4.0329831419496705e-05,
"loss": 0.8842,
"step": 19800
},
{
"epoch": 0.1944593736258367,
"grad_norm": 0.582313597202301,
"learning_rate": 4.0280967505497194e-05,
"loss": 0.8506,
"step": 19900
},
{
"epoch": 0.1954365564078761,
"grad_norm": 0.7206740379333496,
"learning_rate": 4.023210359149768e-05,
"loss": 0.8529,
"step": 20000
},
{
"epoch": 0.19641373918991548,
"grad_norm": 0.45054760575294495,
"learning_rate": 4.018323967749817e-05,
"loss": 0.8564,
"step": 20100
},
{
"epoch": 0.19739092197195485,
"grad_norm": 0.9214595556259155,
"learning_rate": 4.013437576349866e-05,
"loss": 0.8443,
"step": 20200
},
{
"epoch": 0.19836810475399425,
"grad_norm": 0.9843263626098633,
"learning_rate": 4.0085511849499146e-05,
"loss": 0.856,
"step": 20300
},
{
"epoch": 0.19934528753603362,
"grad_norm": 0.6508098840713501,
"learning_rate": 4.0036647935499634e-05,
"loss": 0.8532,
"step": 20400
},
{
"epoch": 0.200322470318073,
"grad_norm": 0.8091655969619751,
"learning_rate": 3.998778402150012e-05,
"loss": 0.8691,
"step": 20500
},
{
"epoch": 0.20129965310011239,
"grad_norm": 0.8139657378196716,
"learning_rate": 3.993892010750061e-05,
"loss": 0.8608,
"step": 20600
},
{
"epoch": 0.20227683588215176,
"grad_norm": 0.628423273563385,
"learning_rate": 3.9890056193501105e-05,
"loss": 0.8369,
"step": 20700
},
{
"epoch": 0.20325401866419113,
"grad_norm": 1.737331748008728,
"learning_rate": 3.9841192279501586e-05,
"loss": 0.8363,
"step": 20800
},
{
"epoch": 0.20423120144623053,
"grad_norm": 1.036280870437622,
"learning_rate": 3.979232836550208e-05,
"loss": 0.8387,
"step": 20900
},
{
"epoch": 0.2052083842282699,
"grad_norm": 0.35834863781929016,
"learning_rate": 3.974346445150256e-05,
"loss": 0.8565,
"step": 21000
},
{
"epoch": 0.20618556701030927,
"grad_norm": 0.7657331824302673,
"learning_rate": 3.969460053750306e-05,
"loss": 0.8654,
"step": 21100
},
{
"epoch": 0.20716274979234867,
"grad_norm": 1.077300786972046,
"learning_rate": 3.9645736623503545e-05,
"loss": 0.8218,
"step": 21200
},
{
"epoch": 0.20813993257438804,
"grad_norm": 0.5806353688240051,
"learning_rate": 3.959687270950403e-05,
"loss": 0.8375,
"step": 21300
},
{
"epoch": 0.2091171153564274,
"grad_norm": 0.3875705599784851,
"learning_rate": 3.954800879550452e-05,
"loss": 0.8342,
"step": 21400
},
{
"epoch": 0.2100942981384668,
"grad_norm": 0.7829961180686951,
"learning_rate": 3.949914488150501e-05,
"loss": 0.832,
"step": 21500
}
],
"logging_steps": 100,
"max_steps": 102335,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2.6967284815757312e+17,
"train_batch_size": 12,
"trial_name": null,
"trial_params": null
}