yalhessi's picture
End of training
86112b8 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 6.0,
"eval_steps": 2492,
"global_step": 74736,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.040141297366730895,
"grad_norm": 1.3413853645324707,
"learning_rate": 0.00019893081086874966,
"loss": 0.8055,
"step": 500
},
{
"epoch": 0.08028259473346179,
"grad_norm": 1.8721694946289062,
"learning_rate": 0.00019759365453668473,
"loss": 0.5632,
"step": 1000
},
{
"epoch": 0.12042389210019268,
"grad_norm": 1.127410650253296,
"learning_rate": 0.0001962538185326116,
"loss": 0.5087,
"step": 1500
},
{
"epoch": 0.16056518946692358,
"grad_norm": 1.7598766088485718,
"learning_rate": 0.00019491398252853852,
"loss": 0.4601,
"step": 2000
},
{
"epoch": 0.20006422607578678,
"eval_loss": 0.37580204010009766,
"eval_runtime": 9.2262,
"eval_samples_per_second": 10.839,
"eval_steps_per_second": 5.419,
"step": 2492
},
{
"epoch": 0.20070648683365447,
"grad_norm": 1.5294134616851807,
"learning_rate": 0.00019357414652446543,
"loss": 0.4423,
"step": 2500
},
{
"epoch": 0.24084778420038536,
"grad_norm": 2.4775636196136475,
"learning_rate": 0.00019223699019240047,
"loss": 0.4201,
"step": 3000
},
{
"epoch": 0.2809890815671163,
"grad_norm": 1.5586802959442139,
"learning_rate": 0.00019089715418832737,
"loss": 0.3942,
"step": 3500
},
{
"epoch": 0.32113037893384716,
"grad_norm": 2.8762762546539307,
"learning_rate": 0.00018955731818425425,
"loss": 0.3846,
"step": 4000
},
{
"epoch": 0.36127167630057805,
"grad_norm": 1.1510844230651855,
"learning_rate": 0.00018821748218018116,
"loss": 0.369,
"step": 4500
},
{
"epoch": 0.40012845215157355,
"eval_loss": 0.3327239751815796,
"eval_runtime": 9.2684,
"eval_samples_per_second": 10.789,
"eval_steps_per_second": 5.395,
"step": 4984
},
{
"epoch": 0.40141297366730894,
"grad_norm": 2.390737771987915,
"learning_rate": 0.00018687764617610804,
"loss": 0.3662,
"step": 5000
},
{
"epoch": 0.4415542710340398,
"grad_norm": 1.9558954238891602,
"learning_rate": 0.00018553781017203495,
"loss": 0.3516,
"step": 5500
},
{
"epoch": 0.4816955684007707,
"grad_norm": 1.7609800100326538,
"learning_rate": 0.00018419797416796186,
"loss": 0.3399,
"step": 6000
},
{
"epoch": 0.5218368657675017,
"grad_norm": 2.026273488998413,
"learning_rate": 0.00018285813816388874,
"loss": 0.3374,
"step": 6500
},
{
"epoch": 0.5619781631342325,
"grad_norm": 1.3401412963867188,
"learning_rate": 0.0001815209818318238,
"loss": 0.327,
"step": 7000
},
{
"epoch": 0.6001926782273603,
"eval_loss": 0.2893865406513214,
"eval_runtime": 9.2199,
"eval_samples_per_second": 10.846,
"eval_steps_per_second": 5.423,
"step": 7476
},
{
"epoch": 0.6021194605009634,
"grad_norm": 2.0142829418182373,
"learning_rate": 0.0001801811458277507,
"loss": 0.3291,
"step": 7500
},
{
"epoch": 0.6422607578676943,
"grad_norm": 0.9912647604942322,
"learning_rate": 0.0001788413098236776,
"loss": 0.3121,
"step": 8000
},
{
"epoch": 0.6824020552344252,
"grad_norm": 2.552272319793701,
"learning_rate": 0.00017750415349161264,
"loss": 0.3117,
"step": 8500
},
{
"epoch": 0.7225433526011561,
"grad_norm": 2.693617582321167,
"learning_rate": 0.00017616431748753954,
"loss": 0.3069,
"step": 9000
},
{
"epoch": 0.762684649967887,
"grad_norm": 3.3755452632904053,
"learning_rate": 0.00017482448148346645,
"loss": 0.2988,
"step": 9500
},
{
"epoch": 0.8002569043031471,
"eval_loss": 0.28478091955184937,
"eval_runtime": 9.2844,
"eval_samples_per_second": 10.771,
"eval_steps_per_second": 5.385,
"step": 9968
},
{
"epoch": 0.8028259473346179,
"grad_norm": 1.348958134651184,
"learning_rate": 0.00017348464547939333,
"loss": 0.297,
"step": 10000
},
{
"epoch": 0.8429672447013488,
"grad_norm": 1.8039051294326782,
"learning_rate": 0.00017214480947532024,
"loss": 0.3044,
"step": 10500
},
{
"epoch": 0.8831085420680796,
"grad_norm": 1.7400553226470947,
"learning_rate": 0.00017080497347124712,
"loss": 0.2938,
"step": 11000
},
{
"epoch": 0.9232498394348105,
"grad_norm": 1.4909226894378662,
"learning_rate": 0.00016946513746717403,
"loss": 0.2785,
"step": 11500
},
{
"epoch": 0.9633911368015414,
"grad_norm": 2.5195188522338867,
"learning_rate": 0.00016812530146310094,
"loss": 0.2873,
"step": 12000
},
{
"epoch": 1.0003211303789339,
"eval_loss": 0.2697654068470001,
"eval_runtime": 9.4165,
"eval_samples_per_second": 10.62,
"eval_steps_per_second": 5.31,
"step": 12460
},
{
"epoch": 1.0035324341682723,
"grad_norm": 2.530623435974121,
"learning_rate": 0.00016678546545902782,
"loss": 0.2723,
"step": 12500
},
{
"epoch": 1.0436737315350033,
"grad_norm": 1.2821416854858398,
"learning_rate": 0.00016544830912696288,
"loss": 0.2632,
"step": 13000
},
{
"epoch": 1.083815028901734,
"grad_norm": 1.7591134309768677,
"learning_rate": 0.00016410847312288976,
"loss": 0.2633,
"step": 13500
},
{
"epoch": 1.123956326268465,
"grad_norm": 1.4008091688156128,
"learning_rate": 0.00016276863711881667,
"loss": 0.245,
"step": 14000
},
{
"epoch": 1.1640976236351959,
"grad_norm": 1.567612886428833,
"learning_rate": 0.00016142880111474355,
"loss": 0.2487,
"step": 14500
},
{
"epoch": 1.2003853564547207,
"eval_loss": 0.26689666509628296,
"eval_runtime": 9.3095,
"eval_samples_per_second": 10.742,
"eval_steps_per_second": 5.371,
"step": 14952
},
{
"epoch": 1.2042389210019269,
"grad_norm": 1.9705049991607666,
"learning_rate": 0.00016008896511067046,
"loss": 0.247,
"step": 15000
},
{
"epoch": 1.2443802183686576,
"grad_norm": 1.31686270236969,
"learning_rate": 0.00015875180877860553,
"loss": 0.2459,
"step": 15500
},
{
"epoch": 1.2845215157353886,
"grad_norm": 1.3225802183151245,
"learning_rate": 0.0001574119727745324,
"loss": 0.253,
"step": 16000
},
{
"epoch": 1.3246628131021194,
"grad_norm": 1.7328706979751587,
"learning_rate": 0.00015607213677045932,
"loss": 0.2404,
"step": 16500
},
{
"epoch": 1.3648041104688504,
"grad_norm": 1.6794183254241943,
"learning_rate": 0.00015473498043839433,
"loss": 0.2371,
"step": 17000
},
{
"epoch": 1.4004495825305074,
"eval_loss": 0.24297775328159332,
"eval_runtime": 9.4719,
"eval_samples_per_second": 10.558,
"eval_steps_per_second": 5.279,
"step": 17444
},
{
"epoch": 1.4049454078355812,
"grad_norm": 1.2241907119750977,
"learning_rate": 0.00015339514443432124,
"loss": 0.2407,
"step": 17500
},
{
"epoch": 1.4450867052023122,
"grad_norm": 2.555837631225586,
"learning_rate": 0.00015205530843024815,
"loss": 0.2344,
"step": 18000
},
{
"epoch": 1.485228002569043,
"grad_norm": 2.1652560234069824,
"learning_rate": 0.00015071547242617505,
"loss": 0.2404,
"step": 18500
},
{
"epoch": 1.525369299935774,
"grad_norm": 3.3938980102539062,
"learning_rate": 0.00014937563642210196,
"loss": 0.2383,
"step": 19000
},
{
"epoch": 1.565510597302505,
"grad_norm": 2.078359365463257,
"learning_rate": 0.00014803580041802884,
"loss": 0.2324,
"step": 19500
},
{
"epoch": 1.6005138086062942,
"eval_loss": 0.23732736706733704,
"eval_runtime": 9.3828,
"eval_samples_per_second": 10.658,
"eval_steps_per_second": 5.329,
"step": 19936
},
{
"epoch": 1.6056518946692357,
"grad_norm": 2.344526767730713,
"learning_rate": 0.00014669596441395575,
"loss": 0.2337,
"step": 20000
},
{
"epoch": 1.6457931920359665,
"grad_norm": 3.3463504314422607,
"learning_rate": 0.00014535612840988263,
"loss": 0.2269,
"step": 20500
},
{
"epoch": 1.6859344894026975,
"grad_norm": 1.8101106882095337,
"learning_rate": 0.00014401629240580954,
"loss": 0.2281,
"step": 21000
},
{
"epoch": 1.7260757867694285,
"grad_norm": 2.2214066982269287,
"learning_rate": 0.00014267645640173645,
"loss": 0.2299,
"step": 21500
},
{
"epoch": 1.7662170841361593,
"grad_norm": 3.5231621265411377,
"learning_rate": 0.00014133930006967149,
"loss": 0.2278,
"step": 22000
},
{
"epoch": 1.800578034682081,
"eval_loss": 0.23720289766788483,
"eval_runtime": 9.3853,
"eval_samples_per_second": 10.655,
"eval_steps_per_second": 5.327,
"step": 22428
},
{
"epoch": 1.80635838150289,
"grad_norm": 1.503125548362732,
"learning_rate": 0.0001399994640655984,
"loss": 0.2286,
"step": 22500
},
{
"epoch": 1.846499678869621,
"grad_norm": 1.3709886074066162,
"learning_rate": 0.00013865962806152527,
"loss": 0.2204,
"step": 23000
},
{
"epoch": 1.886640976236352,
"grad_norm": 0.8027909398078918,
"learning_rate": 0.00013731979205745218,
"loss": 0.2122,
"step": 23500
},
{
"epoch": 1.9267822736030829,
"grad_norm": 1.0423333644866943,
"learning_rate": 0.0001359826357253872,
"loss": 0.2209,
"step": 24000
},
{
"epoch": 1.9669235709698136,
"grad_norm": 3.140822172164917,
"learning_rate": 0.00013464547939332226,
"loss": 0.2212,
"step": 24500
},
{
"epoch": 2.0006422607578678,
"eval_loss": 0.21895286440849304,
"eval_runtime": 9.4416,
"eval_samples_per_second": 10.591,
"eval_steps_per_second": 5.296,
"step": 24920
},
{
"epoch": 2.0070648683365446,
"grad_norm": 2.0828654766082764,
"learning_rate": 0.00013330564338924917,
"loss": 0.2157,
"step": 25000
},
{
"epoch": 2.0472061657032756,
"grad_norm": 1.9575061798095703,
"learning_rate": 0.00013196580738517605,
"loss": 0.1942,
"step": 25500
},
{
"epoch": 2.0873474630700066,
"grad_norm": 2.263435125350952,
"learning_rate": 0.00013062597138110296,
"loss": 0.1939,
"step": 26000
},
{
"epoch": 2.127488760436737,
"grad_norm": 1.9007372856140137,
"learning_rate": 0.00012928613537702984,
"loss": 0.1964,
"step": 26500
},
{
"epoch": 2.167630057803468,
"grad_norm": 1.8549082279205322,
"learning_rate": 0.00012794629937295675,
"loss": 0.1991,
"step": 27000
},
{
"epoch": 2.2007064868336546,
"eval_loss": 0.21308277547359467,
"eval_runtime": 9.4114,
"eval_samples_per_second": 10.625,
"eval_steps_per_second": 5.313,
"step": 27412
},
{
"epoch": 2.207771355170199,
"grad_norm": 1.3462762832641602,
"learning_rate": 0.00012660646336888366,
"loss": 0.1923,
"step": 27500
},
{
"epoch": 2.24791265253693,
"grad_norm": 1.3775655031204224,
"learning_rate": 0.00012526662736481056,
"loss": 0.1956,
"step": 28000
},
{
"epoch": 2.2880539499036607,
"grad_norm": 3.3470311164855957,
"learning_rate": 0.0001239294710327456,
"loss": 0.188,
"step": 28500
},
{
"epoch": 2.3281952472703917,
"grad_norm": 2.1436495780944824,
"learning_rate": 0.00012259231470068064,
"loss": 0.1963,
"step": 29000
},
{
"epoch": 2.3683365446371227,
"grad_norm": 1.8383592367172241,
"learning_rate": 0.00012125247869660754,
"loss": 0.1854,
"step": 29500
},
{
"epoch": 2.4007707129094413,
"eval_loss": 0.20915324985980988,
"eval_runtime": 9.2953,
"eval_samples_per_second": 10.758,
"eval_steps_per_second": 5.379,
"step": 29904
},
{
"epoch": 2.4084778420038537,
"grad_norm": 1.6092569828033447,
"learning_rate": 0.00011991264269253443,
"loss": 0.1949,
"step": 30000
},
{
"epoch": 2.4486191393705843,
"grad_norm": 1.8148771524429321,
"learning_rate": 0.00011857280668846134,
"loss": 0.1902,
"step": 30500
},
{
"epoch": 2.4887604367373153,
"grad_norm": 2.0251662731170654,
"learning_rate": 0.00011723297068438823,
"loss": 0.1921,
"step": 31000
},
{
"epoch": 2.5289017341040463,
"grad_norm": 4.4525861740112305,
"learning_rate": 0.00011589313468031513,
"loss": 0.1972,
"step": 31500
},
{
"epoch": 2.5690430314707773,
"grad_norm": 2.327648878097534,
"learning_rate": 0.00011455329867624202,
"loss": 0.1769,
"step": 32000
},
{
"epoch": 2.600834938985228,
"eval_loss": 0.20336776971817017,
"eval_runtime": 9.3362,
"eval_samples_per_second": 10.711,
"eval_steps_per_second": 5.356,
"step": 32396
},
{
"epoch": 2.6091843288375083,
"grad_norm": 1.7865198850631714,
"learning_rate": 0.00011321346267216893,
"loss": 0.1888,
"step": 32500
},
{
"epoch": 2.649325626204239,
"grad_norm": 2.9860942363739014,
"learning_rate": 0.00011187362666809582,
"loss": 0.1948,
"step": 33000
},
{
"epoch": 2.68946692357097,
"grad_norm": 1.5154398679733276,
"learning_rate": 0.00011053647033603088,
"loss": 0.1936,
"step": 33500
},
{
"epoch": 2.729608220937701,
"grad_norm": 1.3628119230270386,
"learning_rate": 0.00010919663433195777,
"loss": 0.1826,
"step": 34000
},
{
"epoch": 2.7697495183044314,
"grad_norm": 1.1184334754943848,
"learning_rate": 0.00010785679832788467,
"loss": 0.1818,
"step": 34500
},
{
"epoch": 2.800899165061015,
"eval_loss": 0.19036497175693512,
"eval_runtime": 9.2882,
"eval_samples_per_second": 10.766,
"eval_steps_per_second": 5.383,
"step": 34888
},
{
"epoch": 2.8098908156711624,
"grad_norm": 3.3239927291870117,
"learning_rate": 0.00010651696232381156,
"loss": 0.1738,
"step": 35000
},
{
"epoch": 2.8500321130378934,
"grad_norm": 1.8970750570297241,
"learning_rate": 0.00010517712631973846,
"loss": 0.1848,
"step": 35500
},
{
"epoch": 2.8901734104046244,
"grad_norm": 2.5936954021453857,
"learning_rate": 0.00010383996998767351,
"loss": 0.1834,
"step": 36000
},
{
"epoch": 2.9303147077713554,
"grad_norm": 4.450264930725098,
"learning_rate": 0.00010250013398360042,
"loss": 0.1768,
"step": 36500
},
{
"epoch": 2.970456005138086,
"grad_norm": 2.1718015670776367,
"learning_rate": 0.00010116029797952731,
"loss": 0.1798,
"step": 37000
},
{
"epoch": 3.0009633911368017,
"eval_loss": 0.19798487424850464,
"eval_runtime": 9.4331,
"eval_samples_per_second": 10.601,
"eval_steps_per_second": 5.3,
"step": 37380
},
{
"epoch": 3.010597302504817,
"grad_norm": 1.2895421981811523,
"learning_rate": 9.98204619754542e-05,
"loss": 0.1679,
"step": 37500
},
{
"epoch": 3.050738599871548,
"grad_norm": 2.5621845722198486,
"learning_rate": 9.848330564338926e-05,
"loss": 0.1556,
"step": 38000
},
{
"epoch": 3.090879897238279,
"grad_norm": 4.237947940826416,
"learning_rate": 9.714346963931615e-05,
"loss": 0.1627,
"step": 38500
},
{
"epoch": 3.1310211946050095,
"grad_norm": 0.6110877990722656,
"learning_rate": 9.580363363524305e-05,
"loss": 0.1577,
"step": 39000
},
{
"epoch": 3.1711624919717405,
"grad_norm": 1.6049071550369263,
"learning_rate": 9.446379763116994e-05,
"loss": 0.1656,
"step": 39500
},
{
"epoch": 3.2010276172125884,
"eval_loss": 0.18195098638534546,
"eval_runtime": 9.445,
"eval_samples_per_second": 10.588,
"eval_steps_per_second": 5.294,
"step": 39872
},
{
"epoch": 3.2113037893384715,
"grad_norm": 1.347450852394104,
"learning_rate": 9.3126641299105e-05,
"loss": 0.1523,
"step": 40000
},
{
"epoch": 3.2514450867052025,
"grad_norm": 2.5243570804595947,
"learning_rate": 9.17868052950319e-05,
"loss": 0.1549,
"step": 40500
},
{
"epoch": 3.291586384071933,
"grad_norm": 1.4282008409500122,
"learning_rate": 9.044964896296694e-05,
"loss": 0.1624,
"step": 41000
},
{
"epoch": 3.331727681438664,
"grad_norm": 1.0618197917938232,
"learning_rate": 8.910981295889384e-05,
"loss": 0.1506,
"step": 41500
},
{
"epoch": 3.371868978805395,
"grad_norm": 2.2008793354034424,
"learning_rate": 8.776997695482073e-05,
"loss": 0.1533,
"step": 42000
},
{
"epoch": 3.401091843288375,
"eval_loss": 0.1724245399236679,
"eval_runtime": 9.5166,
"eval_samples_per_second": 10.508,
"eval_steps_per_second": 5.254,
"step": 42364
},
{
"epoch": 3.412010276172126,
"grad_norm": 2.9689488410949707,
"learning_rate": 8.643014095074763e-05,
"loss": 0.1587,
"step": 42500
},
{
"epoch": 3.4521515735388566,
"grad_norm": 2.200533628463745,
"learning_rate": 8.509030494667452e-05,
"loss": 0.1521,
"step": 43000
},
{
"epoch": 3.4922928709055876,
"grad_norm": 2.433441162109375,
"learning_rate": 8.375046894260143e-05,
"loss": 0.1559,
"step": 43500
},
{
"epoch": 3.5324341682723186,
"grad_norm": 1.3238328695297241,
"learning_rate": 8.241063293852834e-05,
"loss": 0.1513,
"step": 44000
},
{
"epoch": 3.5725754656390496,
"grad_norm": 1.2658567428588867,
"learning_rate": 8.107079693445523e-05,
"loss": 0.1557,
"step": 44500
},
{
"epoch": 3.601156069364162,
"eval_loss": 0.1686776727437973,
"eval_runtime": 9.5131,
"eval_samples_per_second": 10.512,
"eval_steps_per_second": 5.256,
"step": 44856
},
{
"epoch": 3.61271676300578,
"grad_norm": 0.9541091918945312,
"learning_rate": 7.973096093038212e-05,
"loss": 0.1538,
"step": 45000
},
{
"epoch": 3.652858060372511,
"grad_norm": 2.3008804321289062,
"learning_rate": 7.839112492630902e-05,
"loss": 0.1595,
"step": 45500
},
{
"epoch": 3.692999357739242,
"grad_norm": 0.9546318054199219,
"learning_rate": 7.705128892223593e-05,
"loss": 0.1511,
"step": 46000
},
{
"epoch": 3.733140655105973,
"grad_norm": 2.0182559490203857,
"learning_rate": 7.571413259017097e-05,
"loss": 0.1524,
"step": 46500
},
{
"epoch": 3.773281952472704,
"grad_norm": 3.313910484313965,
"learning_rate": 7.437429658609787e-05,
"loss": 0.1536,
"step": 47000
},
{
"epoch": 3.8012202954399488,
"eval_loss": 0.17214125394821167,
"eval_runtime": 9.4988,
"eval_samples_per_second": 10.528,
"eval_steps_per_second": 5.264,
"step": 47348
},
{
"epoch": 3.8134232498394347,
"grad_norm": 1.0791243314743042,
"learning_rate": 7.303446058202477e-05,
"loss": 0.1559,
"step": 47500
},
{
"epoch": 3.8535645472061657,
"grad_norm": 1.7102487087249756,
"learning_rate": 7.169462457795166e-05,
"loss": 0.1558,
"step": 48000
},
{
"epoch": 3.8937058445728967,
"grad_norm": 1.6259522438049316,
"learning_rate": 7.03574682458867e-05,
"loss": 0.1519,
"step": 48500
},
{
"epoch": 3.9338471419396273,
"grad_norm": 1.94523286819458,
"learning_rate": 6.90176322418136e-05,
"loss": 0.1498,
"step": 49000
},
{
"epoch": 3.9739884393063583,
"grad_norm": 1.2504680156707764,
"learning_rate": 6.768047590974865e-05,
"loss": 0.1531,
"step": 49500
},
{
"epoch": 4.0012845215157355,
"eval_loss": 0.1633564978837967,
"eval_runtime": 9.4116,
"eval_samples_per_second": 10.625,
"eval_steps_per_second": 5.313,
"step": 49840
},
{
"epoch": 4.014129736673089,
"grad_norm": 2.535567283630371,
"learning_rate": 6.634063990567555e-05,
"loss": 0.1421,
"step": 50000
},
{
"epoch": 4.05427103403982,
"grad_norm": 2.8016157150268555,
"learning_rate": 6.500348357361059e-05,
"loss": 0.1304,
"step": 50500
},
{
"epoch": 4.094412331406551,
"grad_norm": 3.923179864883423,
"learning_rate": 6.366364756953749e-05,
"loss": 0.1222,
"step": 51000
},
{
"epoch": 4.134553628773282,
"grad_norm": 4.409369945526123,
"learning_rate": 6.232381156546439e-05,
"loss": 0.1344,
"step": 51500
},
{
"epoch": 4.174694926140013,
"grad_norm": 1.9566134214401245,
"learning_rate": 6.0983975561391295e-05,
"loss": 0.1256,
"step": 52000
},
{
"epoch": 4.201348747591522,
"eval_loss": 0.15292324125766754,
"eval_runtime": 9.4583,
"eval_samples_per_second": 10.573,
"eval_steps_per_second": 5.286,
"step": 52332
},
{
"epoch": 4.214836223506744,
"grad_norm": 1.6278679370880127,
"learning_rate": 5.964413955731819e-05,
"loss": 0.1274,
"step": 52500
},
{
"epoch": 4.254977520873474,
"grad_norm": 0.310170441865921,
"learning_rate": 5.830430355324509e-05,
"loss": 0.1277,
"step": 53000
},
{
"epoch": 4.295118818240206,
"grad_norm": 2.6923305988311768,
"learning_rate": 5.6964467549171985e-05,
"loss": 0.1284,
"step": 53500
},
{
"epoch": 4.335260115606936,
"grad_norm": 2.4311087131500244,
"learning_rate": 5.562463154509888e-05,
"loss": 0.1274,
"step": 54000
},
{
"epoch": 4.375401412973667,
"grad_norm": 0.18654285371303558,
"learning_rate": 5.4287475213033926e-05,
"loss": 0.1274,
"step": 54500
},
{
"epoch": 4.401412973667309,
"eval_loss": 0.14706894755363464,
"eval_runtime": 9.5362,
"eval_samples_per_second": 10.486,
"eval_steps_per_second": 5.243,
"step": 54824
},
{
"epoch": 4.415542710340398,
"grad_norm": 4.156493663787842,
"learning_rate": 5.2947639208960834e-05,
"loss": 0.1257,
"step": 55000
},
{
"epoch": 4.455684007707129,
"grad_norm": 1.4045820236206055,
"learning_rate": 5.160780320488773e-05,
"loss": 0.1302,
"step": 55500
},
{
"epoch": 4.49582530507386,
"grad_norm": 0.5079956650733948,
"learning_rate": 5.026796720081462e-05,
"loss": 0.1302,
"step": 56000
},
{
"epoch": 4.535966602440591,
"grad_norm": 2.671663522720337,
"learning_rate": 4.892813119674152e-05,
"loss": 0.1319,
"step": 56500
},
{
"epoch": 4.5761078998073215,
"grad_norm": 2.7671585083007812,
"learning_rate": 4.758829519266842e-05,
"loss": 0.1294,
"step": 57000
},
{
"epoch": 4.601477199743096,
"eval_loss": 0.1432746946811676,
"eval_runtime": 9.2082,
"eval_samples_per_second": 10.86,
"eval_steps_per_second": 5.43,
"step": 57316
},
{
"epoch": 4.616249197174053,
"grad_norm": 4.464232921600342,
"learning_rate": 4.624845918859532e-05,
"loss": 0.1236,
"step": 57500
},
{
"epoch": 4.6563904945407835,
"grad_norm": 0.22100146114826202,
"learning_rate": 4.490862318452222e-05,
"loss": 0.1214,
"step": 58000
},
{
"epoch": 4.696531791907514,
"grad_norm": 3.2810275554656982,
"learning_rate": 4.357146685245726e-05,
"loss": 0.1299,
"step": 58500
},
{
"epoch": 4.7366730892742455,
"grad_norm": 4.431360721588135,
"learning_rate": 4.2234310520392307e-05,
"loss": 0.129,
"step": 59000
},
{
"epoch": 4.776814386640976,
"grad_norm": 2.974076509475708,
"learning_rate": 4.08944745163192e-05,
"loss": 0.1209,
"step": 59500
},
{
"epoch": 4.801541425818883,
"eval_loss": 0.13750909268856049,
"eval_runtime": 9.2555,
"eval_samples_per_second": 10.804,
"eval_steps_per_second": 5.402,
"step": 59808
},
{
"epoch": 4.8169556840077075,
"grad_norm": 0.5916352272033691,
"learning_rate": 3.95546385122461e-05,
"loss": 0.1325,
"step": 60000
},
{
"epoch": 4.857096981374438,
"grad_norm": 3.2611584663391113,
"learning_rate": 3.8214802508173e-05,
"loss": 0.126,
"step": 60500
},
{
"epoch": 4.897238278741169,
"grad_norm": 2.499826192855835,
"learning_rate": 3.68749665040999e-05,
"loss": 0.127,
"step": 61000
},
{
"epoch": 4.9373795761079,
"grad_norm": 3.263596773147583,
"learning_rate": 3.55351305000268e-05,
"loss": 0.1263,
"step": 61500
},
{
"epoch": 4.977520873474631,
"grad_norm": 3.2504055500030518,
"learning_rate": 3.419529449595369e-05,
"loss": 0.1298,
"step": 62000
},
{
"epoch": 5.001605651894669,
"eval_loss": 0.13082526624202728,
"eval_runtime": 9.18,
"eval_samples_per_second": 10.893,
"eval_steps_per_second": 5.447,
"step": 62300
},
{
"epoch": 5.017662170841362,
"grad_norm": 1.3521040678024292,
"learning_rate": 3.28554584918806e-05,
"loss": 0.0973,
"step": 62500
},
{
"epoch": 5.057803468208093,
"grad_norm": 1.273701786994934,
"learning_rate": 3.151830215981564e-05,
"loss": 0.1059,
"step": 63000
},
{
"epoch": 5.097944765574823,
"grad_norm": 3.1900885105133057,
"learning_rate": 3.017846615574254e-05,
"loss": 0.1095,
"step": 63500
},
{
"epoch": 5.138086062941555,
"grad_norm": 1.1862250566482544,
"learning_rate": 2.883863015166944e-05,
"loss": 0.1053,
"step": 64000
},
{
"epoch": 5.178227360308285,
"grad_norm": 0.06329891085624695,
"learning_rate": 2.7498794147596337e-05,
"loss": 0.1064,
"step": 64500
},
{
"epoch": 5.201669877970456,
"eval_loss": 0.12689721584320068,
"eval_runtime": 9.2645,
"eval_samples_per_second": 10.794,
"eval_steps_per_second": 5.397,
"step": 64792
},
{
"epoch": 5.218368657675016,
"grad_norm": 1.3881759643554688,
"learning_rate": 2.615895814352323e-05,
"loss": 0.0983,
"step": 65000
},
{
"epoch": 5.258509955041747,
"grad_norm": 2.040412664413452,
"learning_rate": 2.4819122139450132e-05,
"loss": 0.11,
"step": 65500
},
{
"epoch": 5.298651252408478,
"grad_norm": 1.5292569398880005,
"learning_rate": 2.348196580738518e-05,
"loss": 0.1087,
"step": 66000
},
{
"epoch": 5.338792549775209,
"grad_norm": 1.2713596820831299,
"learning_rate": 2.2142129803312077e-05,
"loss": 0.1082,
"step": 66500
},
{
"epoch": 5.37893384714194,
"grad_norm": 1.5126820802688599,
"learning_rate": 2.080497347124712e-05,
"loss": 0.1063,
"step": 67000
},
{
"epoch": 5.401734104046243,
"eval_loss": 0.12622660398483276,
"eval_runtime": 9.2201,
"eval_samples_per_second": 10.846,
"eval_steps_per_second": 5.423,
"step": 67284
},
{
"epoch": 5.41907514450867,
"grad_norm": 1.634347915649414,
"learning_rate": 1.9465137467174018e-05,
"loss": 0.1076,
"step": 67500
},
{
"epoch": 5.459216441875402,
"grad_norm": 1.4614862203598022,
"learning_rate": 1.8125301463100915e-05,
"loss": 0.1001,
"step": 68000
},
{
"epoch": 5.499357739242132,
"grad_norm": 1.3177045583724976,
"learning_rate": 1.6785465459027816e-05,
"loss": 0.1041,
"step": 68500
},
{
"epoch": 5.539499036608863,
"grad_norm": 2.3282182216644287,
"learning_rate": 1.5445629454954714e-05,
"loss": 0.1021,
"step": 69000
},
{
"epoch": 5.579640333975594,
"grad_norm": 1.7429168224334717,
"learning_rate": 1.4105793450881613e-05,
"loss": 0.1036,
"step": 69500
},
{
"epoch": 5.60179833012203,
"eval_loss": 0.12125935405492783,
"eval_runtime": 9.2288,
"eval_samples_per_second": 10.836,
"eval_steps_per_second": 5.418,
"step": 69776
},
{
"epoch": 5.619781631342325,
"grad_norm": 0.19996996223926544,
"learning_rate": 1.2765957446808511e-05,
"loss": 0.098,
"step": 70000
},
{
"epoch": 5.659922928709056,
"grad_norm": 3.154064893722534,
"learning_rate": 1.142612144273541e-05,
"loss": 0.101,
"step": 70500
},
{
"epoch": 5.700064226075787,
"grad_norm": 0.9969759583473206,
"learning_rate": 1.0086285438662308e-05,
"loss": 0.1043,
"step": 71000
},
{
"epoch": 5.740205523442517,
"grad_norm": 0.29482555389404297,
"learning_rate": 8.749129106597353e-06,
"loss": 0.0978,
"step": 71500
},
{
"epoch": 5.780346820809249,
"grad_norm": 7.377176761627197,
"learning_rate": 7.409293102524252e-06,
"loss": 0.1084,
"step": 72000
},
{
"epoch": 5.8018625561978165,
"eval_loss": 0.11656199395656586,
"eval_runtime": 9.2458,
"eval_samples_per_second": 10.816,
"eval_steps_per_second": 5.408,
"step": 72268
},
{
"epoch": 5.820488118175979,
"grad_norm": 1.861075520515442,
"learning_rate": 6.06945709845115e-06,
"loss": 0.1045,
"step": 72500
},
{
"epoch": 5.860629415542711,
"grad_norm": 3.3500561714172363,
"learning_rate": 4.729621094378049e-06,
"loss": 0.103,
"step": 73000
},
{
"epoch": 5.900770712909441,
"grad_norm": 0.29247429966926575,
"learning_rate": 3.3924647623130934e-06,
"loss": 0.1022,
"step": 73500
},
{
"epoch": 5.940912010276172,
"grad_norm": 2.5670969486236572,
"learning_rate": 2.0526287582399915e-06,
"loss": 0.1048,
"step": 74000
},
{
"epoch": 5.981053307642903,
"grad_norm": 1.4825931787490845,
"learning_rate": 7.1279275416689e-07,
"loss": 0.1009,
"step": 74500
},
{
"epoch": 6.0,
"step": 74736,
"total_flos": 1.1033260642249114e+18,
"train_loss": 0.01737067021080218,
"train_runtime": 11655.5312,
"train_samples_per_second": 12.824,
"train_steps_per_second": 6.412
}
],
"logging_steps": 500,
"max_steps": 74736,
"num_input_tokens_seen": 0,
"num_train_epochs": 6,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.1033260642249114e+18,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}