Qwen2.5-1.5B-Open-R1-Distill / trainer_state.json
AlanWuxm's picture
Model save
82df75b verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9999069680900549,
"eval_steps": 100,
"global_step": 2687,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0018606381989022234,
"grad_norm": 0.025198739022016525,
"learning_rate": 3.7174721189591085e-07,
"loss": 0.7637,
"step": 5
},
{
"epoch": 0.003721276397804447,
"grad_norm": 0.017377199605107307,
"learning_rate": 7.434944237918217e-07,
"loss": 0.7786,
"step": 10
},
{
"epoch": 0.0055819145967066705,
"grad_norm": 0.019743537530303,
"learning_rate": 1.1152416356877324e-06,
"loss": 0.7174,
"step": 15
},
{
"epoch": 0.007442552795608894,
"grad_norm": 0.017805561423301697,
"learning_rate": 1.4869888475836434e-06,
"loss": 0.724,
"step": 20
},
{
"epoch": 0.009303190994511117,
"grad_norm": 0.017246991395950317,
"learning_rate": 1.858736059479554e-06,
"loss": 0.7668,
"step": 25
},
{
"epoch": 0.011163829193413341,
"grad_norm": 0.023152414709329605,
"learning_rate": 2.2304832713754648e-06,
"loss": 0.781,
"step": 30
},
{
"epoch": 0.013024467392315565,
"grad_norm": 0.019485710188746452,
"learning_rate": 2.6022304832713758e-06,
"loss": 0.7593,
"step": 35
},
{
"epoch": 0.014885105591217787,
"grad_norm": 0.01784881390631199,
"learning_rate": 2.973977695167287e-06,
"loss": 0.762,
"step": 40
},
{
"epoch": 0.01674574379012001,
"grad_norm": 0.01972021535038948,
"learning_rate": 3.3457249070631974e-06,
"loss": 0.7489,
"step": 45
},
{
"epoch": 0.018606381989022234,
"grad_norm": 0.019333072006702423,
"learning_rate": 3.717472118959108e-06,
"loss": 0.7461,
"step": 50
},
{
"epoch": 0.02046702018792446,
"grad_norm": 0.015675414353609085,
"learning_rate": 4.089219330855019e-06,
"loss": 0.7621,
"step": 55
},
{
"epoch": 0.022327658386826682,
"grad_norm": 0.021361010149121284,
"learning_rate": 4.4609665427509296e-06,
"loss": 0.7677,
"step": 60
},
{
"epoch": 0.024188296585728904,
"grad_norm": 0.021466689184308052,
"learning_rate": 4.83271375464684e-06,
"loss": 0.7857,
"step": 65
},
{
"epoch": 0.02604893478463113,
"grad_norm": 0.019599348306655884,
"learning_rate": 5.2044609665427516e-06,
"loss": 0.7308,
"step": 70
},
{
"epoch": 0.027909572983533353,
"grad_norm": 0.015855278819799423,
"learning_rate": 5.576208178438662e-06,
"loss": 0.7423,
"step": 75
},
{
"epoch": 0.029770211182435575,
"grad_norm": 0.019443219527602196,
"learning_rate": 5.947955390334574e-06,
"loss": 0.7782,
"step": 80
},
{
"epoch": 0.0316308493813378,
"grad_norm": 0.01575319655239582,
"learning_rate": 6.319702602230484e-06,
"loss": 0.7494,
"step": 85
},
{
"epoch": 0.03349148758024002,
"grad_norm": 0.0159482192248106,
"learning_rate": 6.691449814126395e-06,
"loss": 0.6999,
"step": 90
},
{
"epoch": 0.03535212577914225,
"grad_norm": 0.016392122954130173,
"learning_rate": 7.063197026022306e-06,
"loss": 0.6966,
"step": 95
},
{
"epoch": 0.03721276397804447,
"grad_norm": 0.017619600519537926,
"learning_rate": 7.434944237918216e-06,
"loss": 0.7659,
"step": 100
},
{
"epoch": 0.03721276397804447,
"eval_loss": 0.760661780834198,
"eval_runtime": 26.8914,
"eval_samples_per_second": 4.723,
"eval_steps_per_second": 4.723,
"step": 100
},
{
"epoch": 0.039073402176946694,
"grad_norm": 0.02020377479493618,
"learning_rate": 7.806691449814127e-06,
"loss": 0.7475,
"step": 105
},
{
"epoch": 0.04093404037584892,
"grad_norm": 0.018723690882325172,
"learning_rate": 8.178438661710038e-06,
"loss": 0.7233,
"step": 110
},
{
"epoch": 0.04279467857475114,
"grad_norm": 0.017194446176290512,
"learning_rate": 8.550185873605949e-06,
"loss": 0.7655,
"step": 115
},
{
"epoch": 0.044655316773653364,
"grad_norm": 0.013910962268710136,
"learning_rate": 8.921933085501859e-06,
"loss": 0.7493,
"step": 120
},
{
"epoch": 0.04651595497255559,
"grad_norm": 0.01600288413465023,
"learning_rate": 9.29368029739777e-06,
"loss": 0.7528,
"step": 125
},
{
"epoch": 0.04837659317145781,
"grad_norm": 0.01554879080504179,
"learning_rate": 9.66542750929368e-06,
"loss": 0.6947,
"step": 130
},
{
"epoch": 0.050237231370360035,
"grad_norm": 0.015109645202755928,
"learning_rate": 1.0037174721189591e-05,
"loss": 0.6647,
"step": 135
},
{
"epoch": 0.05209786956926226,
"grad_norm": 0.015620779246091843,
"learning_rate": 1.0408921933085503e-05,
"loss": 0.7298,
"step": 140
},
{
"epoch": 0.05395850776816448,
"grad_norm": 0.013970567844808102,
"learning_rate": 1.0780669144981412e-05,
"loss": 0.7485,
"step": 145
},
{
"epoch": 0.055819145967066705,
"grad_norm": 0.014776401221752167,
"learning_rate": 1.1152416356877324e-05,
"loss": 0.6767,
"step": 150
},
{
"epoch": 0.057679784165968924,
"grad_norm": 0.014368101954460144,
"learning_rate": 1.1524163568773235e-05,
"loss": 0.7037,
"step": 155
},
{
"epoch": 0.05954042236487115,
"grad_norm": 0.01449244562536478,
"learning_rate": 1.1895910780669147e-05,
"loss": 0.7305,
"step": 160
},
{
"epoch": 0.061401060563773376,
"grad_norm": 0.015350298024713993,
"learning_rate": 1.2267657992565056e-05,
"loss": 0.6804,
"step": 165
},
{
"epoch": 0.0632616987626756,
"grad_norm": 0.014395203441381454,
"learning_rate": 1.2639405204460968e-05,
"loss": 0.6832,
"step": 170
},
{
"epoch": 0.06512233696157782,
"grad_norm": 0.015541068278253078,
"learning_rate": 1.3011152416356879e-05,
"loss": 0.7168,
"step": 175
},
{
"epoch": 0.06698297516048005,
"grad_norm": 0.012579885311424732,
"learning_rate": 1.338289962825279e-05,
"loss": 0.693,
"step": 180
},
{
"epoch": 0.06884361335938227,
"grad_norm": 0.014168789610266685,
"learning_rate": 1.37546468401487e-05,
"loss": 0.7056,
"step": 185
},
{
"epoch": 0.0707042515582845,
"grad_norm": 0.012087655253708363,
"learning_rate": 1.4126394052044612e-05,
"loss": 0.7006,
"step": 190
},
{
"epoch": 0.07256488975718671,
"grad_norm": 0.014433121308684349,
"learning_rate": 1.4498141263940521e-05,
"loss": 0.7016,
"step": 195
},
{
"epoch": 0.07442552795608894,
"grad_norm": 0.015074139460921288,
"learning_rate": 1.4869888475836432e-05,
"loss": 0.6912,
"step": 200
},
{
"epoch": 0.07442552795608894,
"eval_loss": 0.7027233242988586,
"eval_runtime": 26.5611,
"eval_samples_per_second": 4.781,
"eval_steps_per_second": 4.781,
"step": 200
},
{
"epoch": 0.07628616615499116,
"grad_norm": 0.013617471791803837,
"learning_rate": 1.5241635687732344e-05,
"loss": 0.7139,
"step": 205
},
{
"epoch": 0.07814680435389339,
"grad_norm": 0.013274065218865871,
"learning_rate": 1.5613382899628255e-05,
"loss": 0.6955,
"step": 210
},
{
"epoch": 0.08000744255279561,
"grad_norm": 0.013781987130641937,
"learning_rate": 1.5985130111524165e-05,
"loss": 0.65,
"step": 215
},
{
"epoch": 0.08186808075169784,
"grad_norm": 0.01373015996068716,
"learning_rate": 1.6356877323420076e-05,
"loss": 0.6681,
"step": 220
},
{
"epoch": 0.08372871895060005,
"grad_norm": 0.01403126772493124,
"learning_rate": 1.6728624535315986e-05,
"loss": 0.6981,
"step": 225
},
{
"epoch": 0.08558935714950228,
"grad_norm": 0.01332685723900795,
"learning_rate": 1.7100371747211897e-05,
"loss": 0.6517,
"step": 230
},
{
"epoch": 0.0874499953484045,
"grad_norm": 0.013317782431840897,
"learning_rate": 1.7472118959107808e-05,
"loss": 0.6835,
"step": 235
},
{
"epoch": 0.08931063354730673,
"grad_norm": 0.015096917748451233,
"learning_rate": 1.7843866171003718e-05,
"loss": 0.7138,
"step": 240
},
{
"epoch": 0.09117127174620895,
"grad_norm": 0.013502717949450016,
"learning_rate": 1.8215613382899632e-05,
"loss": 0.6722,
"step": 245
},
{
"epoch": 0.09303190994511118,
"grad_norm": 0.01279054582118988,
"learning_rate": 1.858736059479554e-05,
"loss": 0.6323,
"step": 250
},
{
"epoch": 0.09489254814401339,
"grad_norm": 0.011777005158364773,
"learning_rate": 1.8959107806691453e-05,
"loss": 0.6332,
"step": 255
},
{
"epoch": 0.09675318634291562,
"grad_norm": 0.01320689544081688,
"learning_rate": 1.933085501858736e-05,
"loss": 0.7034,
"step": 260
},
{
"epoch": 0.09861382454181784,
"grad_norm": 0.012140162289142609,
"learning_rate": 1.970260223048327e-05,
"loss": 0.6154,
"step": 265
},
{
"epoch": 0.10047446274072007,
"grad_norm": 0.01772845722734928,
"learning_rate": 1.9999991559715313e-05,
"loss": 0.7119,
"step": 270
},
{
"epoch": 0.1023351009396223,
"grad_norm": 0.012948377057909966,
"learning_rate": 1.999969615124717e-05,
"loss": 0.6554,
"step": 275
},
{
"epoch": 0.10419573913852452,
"grad_norm": 0.013341420330107212,
"learning_rate": 1.9998978742792098e-05,
"loss": 0.636,
"step": 280
},
{
"epoch": 0.10605637733742673,
"grad_norm": 0.012272336520254612,
"learning_rate": 1.999783936462566e-05,
"loss": 0.6182,
"step": 285
},
{
"epoch": 0.10791701553632896,
"grad_norm": 0.012329615652561188,
"learning_rate": 1.999627806483107e-05,
"loss": 0.601,
"step": 290
},
{
"epoch": 0.10977765373523118,
"grad_norm": 0.01313008088618517,
"learning_rate": 1.999429490929718e-05,
"loss": 0.7002,
"step": 295
},
{
"epoch": 0.11163829193413341,
"grad_norm": 0.01170238945633173,
"learning_rate": 1.9991889981715696e-05,
"loss": 0.6784,
"step": 300
},
{
"epoch": 0.11163829193413341,
"eval_loss": 0.676296055316925,
"eval_runtime": 26.7211,
"eval_samples_per_second": 4.753,
"eval_steps_per_second": 4.753,
"step": 300
},
{
"epoch": 0.11349893013303564,
"grad_norm": 0.012631393037736416,
"learning_rate": 1.9989063383577644e-05,
"loss": 0.6332,
"step": 305
},
{
"epoch": 0.11535956833193785,
"grad_norm": 0.012392008677124977,
"learning_rate": 1.998581523416908e-05,
"loss": 0.713,
"step": 310
},
{
"epoch": 0.11722020653084007,
"grad_norm": 0.012342042289674282,
"learning_rate": 1.998214567056607e-05,
"loss": 0.6072,
"step": 315
},
{
"epoch": 0.1190808447297423,
"grad_norm": 0.011801215820014477,
"learning_rate": 1.9978054847628908e-05,
"loss": 0.6553,
"step": 320
},
{
"epoch": 0.12094148292864453,
"grad_norm": 0.013543626293540001,
"learning_rate": 1.997354293799555e-05,
"loss": 0.638,
"step": 325
},
{
"epoch": 0.12280212112754675,
"grad_norm": 0.012367943301796913,
"learning_rate": 1.9968610132074372e-05,
"loss": 0.6867,
"step": 330
},
{
"epoch": 0.12466275932644898,
"grad_norm": 0.01382706593722105,
"learning_rate": 1.99632566380361e-05,
"loss": 0.67,
"step": 335
},
{
"epoch": 0.1265233975253512,
"grad_norm": 0.01336714904755354,
"learning_rate": 1.9957482681805036e-05,
"loss": 0.6615,
"step": 340
},
{
"epoch": 0.12838403572425341,
"grad_norm": 0.012786686420440674,
"learning_rate": 1.9951288507049532e-05,
"loss": 0.6343,
"step": 345
},
{
"epoch": 0.13024467392315564,
"grad_norm": 0.014750728383660316,
"learning_rate": 1.9944674375171697e-05,
"loss": 0.6478,
"step": 350
},
{
"epoch": 0.13210531212205787,
"grad_norm": 0.01330367662012577,
"learning_rate": 1.9937640565296372e-05,
"loss": 0.6844,
"step": 355
},
{
"epoch": 0.1339659503209601,
"grad_norm": 0.011332959868013859,
"learning_rate": 1.9930187374259338e-05,
"loss": 0.6188,
"step": 360
},
{
"epoch": 0.13582658851986232,
"grad_norm": 0.013507647439837456,
"learning_rate": 1.992231511659481e-05,
"loss": 0.6844,
"step": 365
},
{
"epoch": 0.13768722671876454,
"grad_norm": 0.01494019664824009,
"learning_rate": 1.991402412452214e-05,
"loss": 0.6616,
"step": 370
},
{
"epoch": 0.13954786491766677,
"grad_norm": 0.012762832455337048,
"learning_rate": 1.9905314747931816e-05,
"loss": 0.6797,
"step": 375
},
{
"epoch": 0.141408503116569,
"grad_norm": 0.014913683757185936,
"learning_rate": 1.989618735437069e-05,
"loss": 0.6268,
"step": 380
},
{
"epoch": 0.1432691413154712,
"grad_norm": 0.012221734039485455,
"learning_rate": 1.9886642329026457e-05,
"loss": 0.6587,
"step": 385
},
{
"epoch": 0.14512977951437342,
"grad_norm": 0.01553855836391449,
"learning_rate": 1.9876680074711417e-05,
"loss": 0.6403,
"step": 390
},
{
"epoch": 0.14699041771327565,
"grad_norm": 0.013682518154382706,
"learning_rate": 1.986630101184546e-05,
"loss": 0.6287,
"step": 395
},
{
"epoch": 0.14885105591217787,
"grad_norm": 0.013481502421200275,
"learning_rate": 1.9855505578438343e-05,
"loss": 0.6757,
"step": 400
},
{
"epoch": 0.14885105591217787,
"eval_loss": 0.6647208333015442,
"eval_runtime": 27.129,
"eval_samples_per_second": 4.681,
"eval_steps_per_second": 4.681,
"step": 400
},
{
"epoch": 0.1507116941110801,
"grad_norm": 0.01228385604918003,
"learning_rate": 1.984429423007117e-05,
"loss": 0.6277,
"step": 405
},
{
"epoch": 0.15257233230998232,
"grad_norm": 0.014119746163487434,
"learning_rate": 1.9832667439877217e-05,
"loss": 0.615,
"step": 410
},
{
"epoch": 0.15443297050888455,
"grad_norm": 0.014395875856280327,
"learning_rate": 1.9820625698521918e-05,
"loss": 0.6417,
"step": 415
},
{
"epoch": 0.15629360870778677,
"grad_norm": 0.013175971806049347,
"learning_rate": 1.9808169514182182e-05,
"loss": 0.6509,
"step": 420
},
{
"epoch": 0.158154246906689,
"grad_norm": 0.015295376069843769,
"learning_rate": 1.9795299412524948e-05,
"loss": 0.6275,
"step": 425
},
{
"epoch": 0.16001488510559123,
"grad_norm": 0.014611025340855122,
"learning_rate": 1.9782015936684987e-05,
"loss": 0.6627,
"step": 430
},
{
"epoch": 0.16187552330449345,
"grad_norm": 0.01412207167595625,
"learning_rate": 1.9768319647242e-05,
"loss": 0.6362,
"step": 435
},
{
"epoch": 0.16373616150339568,
"grad_norm": 0.012070410884916782,
"learning_rate": 1.9754211122196945e-05,
"loss": 0.6429,
"step": 440
},
{
"epoch": 0.16559679970229788,
"grad_norm": 0.013232079334557056,
"learning_rate": 1.9739690956947652e-05,
"loss": 0.6941,
"step": 445
},
{
"epoch": 0.1674574379012001,
"grad_norm": 0.012606708332896233,
"learning_rate": 1.972475976426369e-05,
"loss": 0.6554,
"step": 450
},
{
"epoch": 0.16931807610010233,
"grad_norm": 0.012638423591852188,
"learning_rate": 1.9709418174260523e-05,
"loss": 0.645,
"step": 455
},
{
"epoch": 0.17117871429900455,
"grad_norm": 0.013119902461767197,
"learning_rate": 1.9693666834372896e-05,
"loss": 0.6128,
"step": 460
},
{
"epoch": 0.17303935249790678,
"grad_norm": 0.011363113299012184,
"learning_rate": 1.9677506409327532e-05,
"loss": 0.6294,
"step": 465
},
{
"epoch": 0.174899990696809,
"grad_norm": 0.014238959178328514,
"learning_rate": 1.9660937581115073e-05,
"loss": 0.6647,
"step": 470
},
{
"epoch": 0.17676062889571123,
"grad_norm": 0.013214629143476486,
"learning_rate": 1.9643961048961283e-05,
"loss": 0.6037,
"step": 475
},
{
"epoch": 0.17862126709461346,
"grad_norm": 0.012312485836446285,
"learning_rate": 1.9626577529297573e-05,
"loss": 0.6703,
"step": 480
},
{
"epoch": 0.18048190529351568,
"grad_norm": 0.012318914756178856,
"learning_rate": 1.9608787755730746e-05,
"loss": 0.6141,
"step": 485
},
{
"epoch": 0.1823425434924179,
"grad_norm": 0.01374764647334814,
"learning_rate": 1.9590592479012022e-05,
"loss": 0.673,
"step": 490
},
{
"epoch": 0.18420318169132013,
"grad_norm": 0.012634415179491043,
"learning_rate": 1.9571992467005395e-05,
"loss": 0.6135,
"step": 495
},
{
"epoch": 0.18606381989022236,
"grad_norm": 0.012813772074878216,
"learning_rate": 1.9552988504655194e-05,
"loss": 0.6648,
"step": 500
},
{
"epoch": 0.18606381989022236,
"eval_loss": 0.6582558751106262,
"eval_runtime": 26.9621,
"eval_samples_per_second": 4.71,
"eval_steps_per_second": 4.71,
"step": 500
},
{
"epoch": 0.18792445808912456,
"grad_norm": 0.012235240079462528,
"learning_rate": 1.9533581393952978e-05,
"loss": 0.6108,
"step": 505
},
{
"epoch": 0.18978509628802678,
"grad_norm": 0.012589952908456326,
"learning_rate": 1.951377195390367e-05,
"loss": 0.6218,
"step": 510
},
{
"epoch": 0.191645734486929,
"grad_norm": 0.012783786281943321,
"learning_rate": 1.9493561020491024e-05,
"loss": 0.6668,
"step": 515
},
{
"epoch": 0.19350637268583123,
"grad_norm": 0.013828632421791553,
"learning_rate": 1.9472949446642318e-05,
"loss": 0.6081,
"step": 520
},
{
"epoch": 0.19536701088473346,
"grad_norm": 0.011929171159863472,
"learning_rate": 1.945193810219237e-05,
"loss": 0.6329,
"step": 525
},
{
"epoch": 0.1972276490836357,
"grad_norm": 0.014584473334252834,
"learning_rate": 1.9430527873846826e-05,
"loss": 0.7017,
"step": 530
},
{
"epoch": 0.1990882872825379,
"grad_norm": 0.01474926806986332,
"learning_rate": 1.9408719665144756e-05,
"loss": 0.632,
"step": 535
},
{
"epoch": 0.20094892548144014,
"grad_norm": 0.015552829019725323,
"learning_rate": 1.9386514396420503e-05,
"loss": 0.6757,
"step": 540
},
{
"epoch": 0.20280956368034236,
"grad_norm": 0.013232480734586716,
"learning_rate": 1.9363913004764847e-05,
"loss": 0.6722,
"step": 545
},
{
"epoch": 0.2046702018792446,
"grad_norm": 0.012858827598392963,
"learning_rate": 1.9340916443985465e-05,
"loss": 0.6231,
"step": 550
},
{
"epoch": 0.20653084007814682,
"grad_norm": 0.012365566566586494,
"learning_rate": 1.9317525684566686e-05,
"loss": 0.5986,
"step": 555
},
{
"epoch": 0.20839147827704904,
"grad_norm": 0.01528852991759777,
"learning_rate": 1.9293741713628518e-05,
"loss": 0.6537,
"step": 560
},
{
"epoch": 0.21025211647595124,
"grad_norm": 0.014512522146105766,
"learning_rate": 1.9269565534885003e-05,
"loss": 0.6527,
"step": 565
},
{
"epoch": 0.21211275467485347,
"grad_norm": 0.013798325322568417,
"learning_rate": 1.9244998168601848e-05,
"loss": 0.6148,
"step": 570
},
{
"epoch": 0.2139733928737557,
"grad_norm": 0.013186248019337654,
"learning_rate": 1.9220040651553388e-05,
"loss": 0.6106,
"step": 575
},
{
"epoch": 0.21583403107265792,
"grad_norm": 0.013986771926283836,
"learning_rate": 1.9194694036978807e-05,
"loss": 0.6654,
"step": 580
},
{
"epoch": 0.21769466927156014,
"grad_norm": 0.016201818361878395,
"learning_rate": 1.9168959394537708e-05,
"loss": 0.6306,
"step": 585
},
{
"epoch": 0.21955530747046237,
"grad_norm": 0.013889294117689133,
"learning_rate": 1.9142837810264972e-05,
"loss": 0.6749,
"step": 590
},
{
"epoch": 0.2214159456693646,
"grad_norm": 0.013025142252445221,
"learning_rate": 1.911633038652491e-05,
"loss": 0.6075,
"step": 595
},
{
"epoch": 0.22327658386826682,
"grad_norm": 0.013716059736907482,
"learning_rate": 1.9089438241964764e-05,
"loss": 0.6516,
"step": 600
},
{
"epoch": 0.22327658386826682,
"eval_loss": 0.6547934412956238,
"eval_runtime": 26.7928,
"eval_samples_per_second": 4.74,
"eval_steps_per_second": 4.74,
"step": 600
},
{
"epoch": 0.22513722206716905,
"grad_norm": 0.016251811757683754,
"learning_rate": 1.906216251146748e-05,
"loss": 0.6265,
"step": 605
},
{
"epoch": 0.22699786026607127,
"grad_norm": 0.013359563425183296,
"learning_rate": 1.9034504346103825e-05,
"loss": 0.6052,
"step": 610
},
{
"epoch": 0.2288584984649735,
"grad_norm": 0.012794552370905876,
"learning_rate": 1.9006464913083807e-05,
"loss": 0.613,
"step": 615
},
{
"epoch": 0.2307191366638757,
"grad_norm": 0.012533072382211685,
"learning_rate": 1.897804539570742e-05,
"loss": 0.6735,
"step": 620
},
{
"epoch": 0.23257977486277792,
"grad_norm": 0.013286220841109753,
"learning_rate": 1.8949246993314694e-05,
"loss": 0.6692,
"step": 625
},
{
"epoch": 0.23444041306168015,
"grad_norm": 0.013466808013617992,
"learning_rate": 1.892007092123511e-05,
"loss": 0.6513,
"step": 630
},
{
"epoch": 0.23630105126058237,
"grad_norm": 0.012991335242986679,
"learning_rate": 1.8890518410736275e-05,
"loss": 0.6405,
"step": 635
},
{
"epoch": 0.2381616894594846,
"grad_norm": 0.013223089277744293,
"learning_rate": 1.8860590708971997e-05,
"loss": 0.6488,
"step": 640
},
{
"epoch": 0.24002232765838682,
"grad_norm": 0.012394067831337452,
"learning_rate": 1.8830289078929618e-05,
"loss": 0.6131,
"step": 645
},
{
"epoch": 0.24188296585728905,
"grad_norm": 0.013721502386033535,
"learning_rate": 1.8799614799376743e-05,
"loss": 0.681,
"step": 650
},
{
"epoch": 0.24374360405619128,
"grad_norm": 0.012726777233183384,
"learning_rate": 1.8768569164807272e-05,
"loss": 0.6837,
"step": 655
},
{
"epoch": 0.2456042422550935,
"grad_norm": 0.013720668852329254,
"learning_rate": 1.8737153485386737e-05,
"loss": 0.6007,
"step": 660
},
{
"epoch": 0.24746488045399573,
"grad_norm": 0.012646087445318699,
"learning_rate": 1.8705369086897063e-05,
"loss": 0.6545,
"step": 665
},
{
"epoch": 0.24932551865289795,
"grad_norm": 0.013658811338245869,
"learning_rate": 1.8673217310680578e-05,
"loss": 0.6379,
"step": 670
},
{
"epoch": 0.2511861568518002,
"grad_norm": 0.012248256243765354,
"learning_rate": 1.864069951358342e-05,
"loss": 0.6356,
"step": 675
},
{
"epoch": 0.2530467950507024,
"grad_norm": 0.0133894681930542,
"learning_rate": 1.860781706789829e-05,
"loss": 0.6737,
"step": 680
},
{
"epoch": 0.25490743324960463,
"grad_norm": 0.01430124044418335,
"learning_rate": 1.857457136130651e-05,
"loss": 0.6169,
"step": 685
},
{
"epoch": 0.25676807144850683,
"grad_norm": 0.013437042012810707,
"learning_rate": 1.854096379681949e-05,
"loss": 0.6021,
"step": 690
},
{
"epoch": 0.2586287096474091,
"grad_norm": 0.011497768573462963,
"learning_rate": 1.8506995792719498e-05,
"loss": 0.6119,
"step": 695
},
{
"epoch": 0.2604893478463113,
"grad_norm": 0.01319235097616911,
"learning_rate": 1.8472668782499817e-05,
"loss": 0.627,
"step": 700
},
{
"epoch": 0.2604893478463113,
"eval_loss": 0.6522720456123352,
"eval_runtime": 26.7201,
"eval_samples_per_second": 4.753,
"eval_steps_per_second": 4.753,
"step": 700
},
{
"epoch": 0.26234998604521353,
"grad_norm": 0.014044429175555706,
"learning_rate": 1.843798421480426e-05,
"loss": 0.6244,
"step": 705
},
{
"epoch": 0.26421062424411573,
"grad_norm": 0.011311609297990799,
"learning_rate": 1.8402943553365998e-05,
"loss": 0.5975,
"step": 710
},
{
"epoch": 0.26607126244301793,
"grad_norm": 0.012355692684650421,
"learning_rate": 1.8367548276945846e-05,
"loss": 0.6009,
"step": 715
},
{
"epoch": 0.2679319006419202,
"grad_norm": 0.012264437042176723,
"learning_rate": 1.83317998792698e-05,
"loss": 0.6103,
"step": 720
},
{
"epoch": 0.2697925388408224,
"grad_norm": 0.012037084437906742,
"learning_rate": 1.8295699868966038e-05,
"loss": 0.5602,
"step": 725
},
{
"epoch": 0.27165317703972464,
"grad_norm": 0.012773050926625729,
"learning_rate": 1.8259249769501237e-05,
"loss": 0.6215,
"step": 730
},
{
"epoch": 0.27351381523862683,
"grad_norm": 0.012415550649166107,
"learning_rate": 1.8222451119116288e-05,
"loss": 0.6364,
"step": 735
},
{
"epoch": 0.2753744534375291,
"grad_norm": 0.01302468404173851,
"learning_rate": 1.8185305470761366e-05,
"loss": 0.5994,
"step": 740
},
{
"epoch": 0.2772350916364313,
"grad_norm": 0.013447316363453865,
"learning_rate": 1.814781439203043e-05,
"loss": 0.6458,
"step": 745
},
{
"epoch": 0.27909572983533354,
"grad_norm": 0.012098093517124653,
"learning_rate": 1.8109979465095014e-05,
"loss": 0.6357,
"step": 750
},
{
"epoch": 0.28095636803423574,
"grad_norm": 0.012987499125301838,
"learning_rate": 1.8071802286637505e-05,
"loss": 0.6248,
"step": 755
},
{
"epoch": 0.282817006233138,
"grad_norm": 0.011747024022042751,
"learning_rate": 1.8033284467783742e-05,
"loss": 0.6202,
"step": 760
},
{
"epoch": 0.2846776444320402,
"grad_norm": 0.01332057174295187,
"learning_rate": 1.7994427634035016e-05,
"loss": 0.6347,
"step": 765
},
{
"epoch": 0.2865382826309424,
"grad_norm": 0.013383504003286362,
"learning_rate": 1.795523342519948e-05,
"loss": 0.6001,
"step": 770
},
{
"epoch": 0.28839892082984464,
"grad_norm": 0.013648821040987968,
"learning_rate": 1.7915703495322967e-05,
"loss": 0.6399,
"step": 775
},
{
"epoch": 0.29025955902874684,
"grad_norm": 0.012947522103786469,
"learning_rate": 1.7875839512619148e-05,
"loss": 0.6298,
"step": 780
},
{
"epoch": 0.2921201972276491,
"grad_norm": 0.01334394421428442,
"learning_rate": 1.7835643159399156e-05,
"loss": 0.6418,
"step": 785
},
{
"epoch": 0.2939808354265513,
"grad_norm": 0.014045110903680325,
"learning_rate": 1.7795116132000587e-05,
"loss": 0.6403,
"step": 790
},
{
"epoch": 0.29584147362545354,
"grad_norm": 0.015219368040561676,
"learning_rate": 1.7754260140715918e-05,
"loss": 0.6277,
"step": 795
},
{
"epoch": 0.29770211182435574,
"grad_norm": 0.01307649165391922,
"learning_rate": 1.771307690972031e-05,
"loss": 0.6271,
"step": 800
},
{
"epoch": 0.29770211182435574,
"eval_loss": 0.6502260565757751,
"eval_runtime": 26.7042,
"eval_samples_per_second": 4.756,
"eval_steps_per_second": 4.756,
"step": 800
},
{
"epoch": 0.299562750023258,
"grad_norm": 0.013835963793098927,
"learning_rate": 1.7671568176998865e-05,
"loss": 0.6286,
"step": 805
},
{
"epoch": 0.3014233882221602,
"grad_norm": 0.013574733398854733,
"learning_rate": 1.762973569427328e-05,
"loss": 0.6462,
"step": 810
},
{
"epoch": 0.30328402642106245,
"grad_norm": 0.01131366565823555,
"learning_rate": 1.758758122692791e-05,
"loss": 0.6167,
"step": 815
},
{
"epoch": 0.30514466461996465,
"grad_norm": 0.013482702895998955,
"learning_rate": 1.7545106553935277e-05,
"loss": 0.6413,
"step": 820
},
{
"epoch": 0.3070053028188669,
"grad_norm": 0.012861824594438076,
"learning_rate": 1.7502313467780988e-05,
"loss": 0.6027,
"step": 825
},
{
"epoch": 0.3088659410177691,
"grad_norm": 0.012556380592286587,
"learning_rate": 1.7459203774388097e-05,
"loss": 0.6603,
"step": 830
},
{
"epoch": 0.3107265792166713,
"grad_norm": 0.012278062291443348,
"learning_rate": 1.7415779293040887e-05,
"loss": 0.5803,
"step": 835
},
{
"epoch": 0.31258721741557355,
"grad_norm": 0.012213567271828651,
"learning_rate": 1.7372041856308098e-05,
"loss": 0.6624,
"step": 840
},
{
"epoch": 0.31444785561447575,
"grad_norm": 0.0131307952105999,
"learning_rate": 1.7327993309965583e-05,
"loss": 0.6447,
"step": 845
},
{
"epoch": 0.316308493813378,
"grad_norm": 0.0121999466791749,
"learning_rate": 1.7283635512918423e-05,
"loss": 0.6451,
"step": 850
},
{
"epoch": 0.3181691320122802,
"grad_norm": 0.012872702442109585,
"learning_rate": 1.7238970337122484e-05,
"loss": 0.5724,
"step": 855
},
{
"epoch": 0.32002977021118245,
"grad_norm": 0.013137550093233585,
"learning_rate": 1.7193999667505387e-05,
"loss": 0.6459,
"step": 860
},
{
"epoch": 0.32189040841008465,
"grad_norm": 0.013948196545243263,
"learning_rate": 1.7148725401887002e-05,
"loss": 0.651,
"step": 865
},
{
"epoch": 0.3237510466089869,
"grad_norm": 0.012517811730504036,
"learning_rate": 1.710314945089933e-05,
"loss": 0.6114,
"step": 870
},
{
"epoch": 0.3256116848078891,
"grad_norm": 0.014199101366102695,
"learning_rate": 1.7057273737905887e-05,
"loss": 0.6405,
"step": 875
},
{
"epoch": 0.32747232300679135,
"grad_norm": 0.012591714970767498,
"learning_rate": 1.7011100198920528e-05,
"loss": 0.6767,
"step": 880
},
{
"epoch": 0.32933296120569355,
"grad_norm": 0.012114683166146278,
"learning_rate": 1.6964630782525743e-05,
"loss": 0.6037,
"step": 885
},
{
"epoch": 0.33119359940459575,
"grad_norm": 0.013139299117028713,
"learning_rate": 1.6917867449790432e-05,
"loss": 0.643,
"step": 890
},
{
"epoch": 0.333054237603498,
"grad_norm": 0.012391779571771622,
"learning_rate": 1.6870812174187136e-05,
"loss": 0.647,
"step": 895
},
{
"epoch": 0.3349148758024002,
"grad_norm": 0.013360656797885895,
"learning_rate": 1.6823466941508762e-05,
"loss": 0.725,
"step": 900
},
{
"epoch": 0.3349148758024002,
"eval_loss": 0.6488396525382996,
"eval_runtime": 27.3774,
"eval_samples_per_second": 4.639,
"eval_steps_per_second": 4.639,
"step": 900
},
{
"epoch": 0.33677551400130246,
"grad_norm": 0.013424807228147984,
"learning_rate": 1.677583374978478e-05,
"loss": 0.6342,
"step": 905
},
{
"epoch": 0.33863615220020465,
"grad_norm": 0.01573541946709156,
"learning_rate": 1.6727914609196895e-05,
"loss": 0.6562,
"step": 910
},
{
"epoch": 0.3404967903991069,
"grad_norm": 0.011363287456333637,
"learning_rate": 1.6679711541994227e-05,
"loss": 0.6492,
"step": 915
},
{
"epoch": 0.3423574285980091,
"grad_norm": 0.012154373340308666,
"learning_rate": 1.6631226582407954e-05,
"loss": 0.602,
"step": 920
},
{
"epoch": 0.34421806679691136,
"grad_norm": 0.01744014024734497,
"learning_rate": 1.658246177656548e-05,
"loss": 0.6318,
"step": 925
},
{
"epoch": 0.34607870499581356,
"grad_norm": 0.012465902604162693,
"learning_rate": 1.6533419182404078e-05,
"loss": 0.6522,
"step": 930
},
{
"epoch": 0.3479393431947158,
"grad_norm": 0.015055039897561073,
"learning_rate": 1.6484100869584044e-05,
"loss": 0.6376,
"step": 935
},
{
"epoch": 0.349799981393618,
"grad_norm": 0.012351201847195625,
"learning_rate": 1.6434508919401357e-05,
"loss": 0.6206,
"step": 940
},
{
"epoch": 0.3516606195925202,
"grad_norm": 0.012793191708624363,
"learning_rate": 1.6384645424699835e-05,
"loss": 0.6182,
"step": 945
},
{
"epoch": 0.35352125779142246,
"grad_norm": 0.012946651317179203,
"learning_rate": 1.6334512489782833e-05,
"loss": 0.5839,
"step": 950
},
{
"epoch": 0.35538189599032466,
"grad_norm": 0.012998082675039768,
"learning_rate": 1.628411223032442e-05,
"loss": 0.6517,
"step": 955
},
{
"epoch": 0.3572425341892269,
"grad_norm": 0.012614963576197624,
"learning_rate": 1.6233446773280113e-05,
"loss": 0.6235,
"step": 960
},
{
"epoch": 0.3591031723881291,
"grad_norm": 0.012318151071667671,
"learning_rate": 1.6182518256797095e-05,
"loss": 0.664,
"step": 965
},
{
"epoch": 0.36096381058703136,
"grad_norm": 0.012551162391901016,
"learning_rate": 1.6131328830123997e-05,
"loss": 0.6317,
"step": 970
},
{
"epoch": 0.36282444878593356,
"grad_norm": 0.013372802175581455,
"learning_rate": 1.60798806535202e-05,
"loss": 0.6418,
"step": 975
},
{
"epoch": 0.3646850869848358,
"grad_norm": 0.011675640940666199,
"learning_rate": 1.6028175898164665e-05,
"loss": 0.6118,
"step": 980
},
{
"epoch": 0.366545725183738,
"grad_norm": 0.013295911252498627,
"learning_rate": 1.5976216746064294e-05,
"loss": 0.6217,
"step": 985
},
{
"epoch": 0.36840636338264027,
"grad_norm": 0.012895721010863781,
"learning_rate": 1.5924005389961866e-05,
"loss": 0.6436,
"step": 990
},
{
"epoch": 0.37026700158154247,
"grad_norm": 0.014572090469300747,
"learning_rate": 1.5871544033243488e-05,
"loss": 0.6342,
"step": 995
},
{
"epoch": 0.3721276397804447,
"grad_norm": 0.012393898330628872,
"learning_rate": 1.581883488984562e-05,
"loss": 0.6218,
"step": 1000
},
{
"epoch": 0.3721276397804447,
"eval_loss": 0.6476932168006897,
"eval_runtime": 27.1488,
"eval_samples_per_second": 4.678,
"eval_steps_per_second": 4.678,
"step": 1000
},
{
"epoch": 0.3739882779793469,
"grad_norm": 0.014692210592329502,
"learning_rate": 1.5765880184161625e-05,
"loss": 0.6216,
"step": 1005
},
{
"epoch": 0.3758489161782491,
"grad_norm": 0.012152746319770813,
"learning_rate": 1.5712682150947926e-05,
"loss": 0.6243,
"step": 1010
},
{
"epoch": 0.37770955437715137,
"grad_norm": 0.012929155491292477,
"learning_rate": 1.5659243035229657e-05,
"loss": 0.6493,
"step": 1015
},
{
"epoch": 0.37957019257605357,
"grad_norm": 0.0136475944891572,
"learning_rate": 1.5605565092205973e-05,
"loss": 0.6506,
"step": 1020
},
{
"epoch": 0.3814308307749558,
"grad_norm": 0.014008302241563797,
"learning_rate": 1.5551650587154815e-05,
"loss": 0.6429,
"step": 1025
},
{
"epoch": 0.383291468973858,
"grad_norm": 0.014000017195940018,
"learning_rate": 1.5497501795337366e-05,
"loss": 0.6277,
"step": 1030
},
{
"epoch": 0.38515210717276027,
"grad_norm": 0.012146887369453907,
"learning_rate": 1.5443121001901994e-05,
"loss": 0.635,
"step": 1035
},
{
"epoch": 0.38701274537166247,
"grad_norm": 0.013878699392080307,
"learning_rate": 1.5388510501787855e-05,
"loss": 0.6416,
"step": 1040
},
{
"epoch": 0.3888733835705647,
"grad_norm": 0.011823480948805809,
"learning_rate": 1.5333672599628005e-05,
"loss": 0.637,
"step": 1045
},
{
"epoch": 0.3907340217694669,
"grad_norm": 0.012524113990366459,
"learning_rate": 1.527860960965216e-05,
"loss": 0.6763,
"step": 1050
},
{
"epoch": 0.3925946599683692,
"grad_norm": 0.013192784041166306,
"learning_rate": 1.5223323855589027e-05,
"loss": 0.6501,
"step": 1055
},
{
"epoch": 0.3944552981672714,
"grad_norm": 0.012439992278814316,
"learning_rate": 1.5167817670568253e-05,
"loss": 0.5886,
"step": 1060
},
{
"epoch": 0.39631593636617357,
"grad_norm": 0.013605669140815735,
"learning_rate": 1.5112093397021945e-05,
"loss": 0.5925,
"step": 1065
},
{
"epoch": 0.3981765745650758,
"grad_norm": 0.01384530495852232,
"learning_rate": 1.5056153386585828e-05,
"loss": 0.6607,
"step": 1070
},
{
"epoch": 0.400037212763978,
"grad_norm": 0.014060786925256252,
"learning_rate": 1.5000000000000002e-05,
"loss": 0.6458,
"step": 1075
},
{
"epoch": 0.4018978509628803,
"grad_norm": 0.012853951193392277,
"learning_rate": 1.494363560700931e-05,
"loss": 0.6028,
"step": 1080
},
{
"epoch": 0.4037584891617825,
"grad_norm": 0.012846381403505802,
"learning_rate": 1.4887062586263334e-05,
"loss": 0.6543,
"step": 1085
},
{
"epoch": 0.40561912736068473,
"grad_norm": 0.013001542538404465,
"learning_rate": 1.4830283325216026e-05,
"loss": 0.5654,
"step": 1090
},
{
"epoch": 0.4074797655595869,
"grad_norm": 0.012333175167441368,
"learning_rate": 1.477330022002493e-05,
"loss": 0.6465,
"step": 1095
},
{
"epoch": 0.4093404037584892,
"grad_norm": 0.01342159602791071,
"learning_rate": 1.4716115675450078e-05,
"loss": 0.6168,
"step": 1100
},
{
"epoch": 0.4093404037584892,
"eval_loss": 0.6468775272369385,
"eval_runtime": 26.8263,
"eval_samples_per_second": 4.734,
"eval_steps_per_second": 4.734,
"step": 1100
},
{
"epoch": 0.4112010419573914,
"grad_norm": 0.0139292748644948,
"learning_rate": 1.4658732104752507e-05,
"loss": 0.634,
"step": 1105
},
{
"epoch": 0.41306168015629363,
"grad_norm": 0.013169731944799423,
"learning_rate": 1.4601151929592403e-05,
"loss": 0.6227,
"step": 1110
},
{
"epoch": 0.41492231835519583,
"grad_norm": 0.013705245219171047,
"learning_rate": 1.4543377579926915e-05,
"loss": 0.6441,
"step": 1115
},
{
"epoch": 0.4167829565540981,
"grad_norm": 0.013035726733505726,
"learning_rate": 1.4485411493907617e-05,
"loss": 0.6498,
"step": 1120
},
{
"epoch": 0.4186435947530003,
"grad_norm": 0.01190096139907837,
"learning_rate": 1.442725611777758e-05,
"loss": 0.6285,
"step": 1125
},
{
"epoch": 0.4205042329519025,
"grad_norm": 0.013753347098827362,
"learning_rate": 1.4368913905768178e-05,
"loss": 0.6541,
"step": 1130
},
{
"epoch": 0.42236487115080473,
"grad_norm": 0.012330746278166771,
"learning_rate": 1.4310387319995492e-05,
"loss": 0.6721,
"step": 1135
},
{
"epoch": 0.42422550934970693,
"grad_norm": 0.01278294064104557,
"learning_rate": 1.4251678830356408e-05,
"loss": 0.6589,
"step": 1140
},
{
"epoch": 0.4260861475486092,
"grad_norm": 0.012772184796631336,
"learning_rate": 1.41927909144244e-05,
"loss": 0.6411,
"step": 1145
},
{
"epoch": 0.4279467857475114,
"grad_norm": 0.012047790922224522,
"learning_rate": 1.413372605734495e-05,
"loss": 0.5759,
"step": 1150
},
{
"epoch": 0.42980742394641364,
"grad_norm": 0.014543715864419937,
"learning_rate": 1.4074486751730687e-05,
"loss": 0.6578,
"step": 1155
},
{
"epoch": 0.43166806214531583,
"grad_norm": 0.013436605222523212,
"learning_rate": 1.4015075497556193e-05,
"loss": 0.5876,
"step": 1160
},
{
"epoch": 0.4335287003442181,
"grad_norm": 0.011584432795643806,
"learning_rate": 1.3955494802052498e-05,
"loss": 0.656,
"step": 1165
},
{
"epoch": 0.4353893385431203,
"grad_norm": 0.012196795083582401,
"learning_rate": 1.3895747179601275e-05,
"loss": 0.6562,
"step": 1170
},
{
"epoch": 0.43724997674202254,
"grad_norm": 0.011762428097426891,
"learning_rate": 1.3835835151628728e-05,
"loss": 0.5918,
"step": 1175
},
{
"epoch": 0.43911061494092474,
"grad_norm": 0.014100808650255203,
"learning_rate": 1.3775761246499177e-05,
"loss": 0.6216,
"step": 1180
},
{
"epoch": 0.44097125313982694,
"grad_norm": 0.011724433861672878,
"learning_rate": 1.3715527999408376e-05,
"loss": 0.6434,
"step": 1185
},
{
"epoch": 0.4428318913387292,
"grad_norm": 0.013045977801084518,
"learning_rate": 1.365513795227651e-05,
"loss": 0.5915,
"step": 1190
},
{
"epoch": 0.4446925295376314,
"grad_norm": 0.013339078053832054,
"learning_rate": 1.359459365364092e-05,
"loss": 0.6148,
"step": 1195
},
{
"epoch": 0.44655316773653364,
"grad_norm": 0.011851554736495018,
"learning_rate": 1.3533897658548571e-05,
"loss": 0.6294,
"step": 1200
},
{
"epoch": 0.44655316773653364,
"eval_loss": 0.6459712982177734,
"eval_runtime": 27.1858,
"eval_samples_per_second": 4.672,
"eval_steps_per_second": 4.672,
"step": 1200
},
{
"epoch": 0.44841380593543584,
"grad_norm": 0.014516811817884445,
"learning_rate": 1.3473052528448203e-05,
"loss": 0.6052,
"step": 1205
},
{
"epoch": 0.4502744441343381,
"grad_norm": 0.013740907423198223,
"learning_rate": 1.341206083108225e-05,
"loss": 0.6035,
"step": 1210
},
{
"epoch": 0.4521350823332403,
"grad_norm": 0.013699905015528202,
"learning_rate": 1.3350925140378465e-05,
"loss": 0.64,
"step": 1215
},
{
"epoch": 0.45399572053214254,
"grad_norm": 0.013219136744737625,
"learning_rate": 1.328964803634131e-05,
"loss": 0.6102,
"step": 1220
},
{
"epoch": 0.45585635873104474,
"grad_norm": 0.012603058479726315,
"learning_rate": 1.3228232104943073e-05,
"loss": 0.5452,
"step": 1225
},
{
"epoch": 0.457716996929947,
"grad_norm": 0.012072132900357246,
"learning_rate": 1.3166679938014728e-05,
"loss": 0.5864,
"step": 1230
},
{
"epoch": 0.4595776351288492,
"grad_norm": 0.01093310210853815,
"learning_rate": 1.3104994133136563e-05,
"loss": 0.6122,
"step": 1235
},
{
"epoch": 0.4614382733277514,
"grad_norm": 0.01245942059904337,
"learning_rate": 1.3043177293528571e-05,
"loss": 0.5889,
"step": 1240
},
{
"epoch": 0.46329891152665365,
"grad_norm": 0.013127041980624199,
"learning_rate": 1.2981232027940562e-05,
"loss": 0.6225,
"step": 1245
},
{
"epoch": 0.46515954972555584,
"grad_norm": 0.018618909642100334,
"learning_rate": 1.2919160950542095e-05,
"loss": 0.6189,
"step": 1250
},
{
"epoch": 0.4670201879244581,
"grad_norm": 0.013042682781815529,
"learning_rate": 1.2856966680812148e-05,
"loss": 0.674,
"step": 1255
},
{
"epoch": 0.4688808261233603,
"grad_norm": 0.013208975084125996,
"learning_rate": 1.2794651843428575e-05,
"loss": 0.6084,
"step": 1260
},
{
"epoch": 0.47074146432226255,
"grad_norm": 0.012473770417273045,
"learning_rate": 1.2732219068157335e-05,
"loss": 0.5748,
"step": 1265
},
{
"epoch": 0.47260210252116475,
"grad_norm": 0.013629582710564137,
"learning_rate": 1.2669670989741519e-05,
"loss": 0.6358,
"step": 1270
},
{
"epoch": 0.474462740720067,
"grad_norm": 0.014257396571338177,
"learning_rate": 1.2607010247790158e-05,
"loss": 0.6794,
"step": 1275
},
{
"epoch": 0.4763233789189692,
"grad_norm": 0.0164741612970829,
"learning_rate": 1.2544239486666831e-05,
"loss": 0.6647,
"step": 1280
},
{
"epoch": 0.47818401711787145,
"grad_norm": 0.012896180152893066,
"learning_rate": 1.2481361355378066e-05,
"loss": 0.6413,
"step": 1285
},
{
"epoch": 0.48004465531677365,
"grad_norm": 0.012296337634325027,
"learning_rate": 1.2418378507461544e-05,
"loss": 0.62,
"step": 1290
},
{
"epoch": 0.4819052935156759,
"grad_norm": 0.01213445421308279,
"learning_rate": 1.2355293600874132e-05,
"loss": 0.6611,
"step": 1295
},
{
"epoch": 0.4837659317145781,
"grad_norm": 0.01198558695614338,
"learning_rate": 1.229210929787969e-05,
"loss": 0.6438,
"step": 1300
},
{
"epoch": 0.4837659317145781,
"eval_loss": 0.6453238725662231,
"eval_runtime": 26.6037,
"eval_samples_per_second": 4.774,
"eval_steps_per_second": 4.774,
"step": 1300
},
{
"epoch": 0.4856265699134803,
"grad_norm": 0.012105841189622879,
"learning_rate": 1.2228828264936755e-05,
"loss": 0.675,
"step": 1305
},
{
"epoch": 0.48748720811238255,
"grad_norm": 0.012657279148697853,
"learning_rate": 1.2165453172585964e-05,
"loss": 0.6066,
"step": 1310
},
{
"epoch": 0.48934784631128475,
"grad_norm": 0.01320530753582716,
"learning_rate": 1.2101986695337407e-05,
"loss": 0.6578,
"step": 1315
},
{
"epoch": 0.491208484510187,
"grad_norm": 0.012736879289150238,
"learning_rate": 1.2038431511557715e-05,
"loss": 0.6596,
"step": 1320
},
{
"epoch": 0.4930691227090892,
"grad_norm": 0.013182558119297028,
"learning_rate": 1.197479030335706e-05,
"loss": 0.595,
"step": 1325
},
{
"epoch": 0.49492976090799146,
"grad_norm": 0.013970241881906986,
"learning_rate": 1.1911065756475953e-05,
"loss": 0.6525,
"step": 1330
},
{
"epoch": 0.49679039910689365,
"grad_norm": 0.012158108875155449,
"learning_rate": 1.1847260560171895e-05,
"loss": 0.576,
"step": 1335
},
{
"epoch": 0.4986510373057959,
"grad_norm": 0.012398924678564072,
"learning_rate": 1.1783377407105907e-05,
"loss": 0.6039,
"step": 1340
},
{
"epoch": 0.5005116755046981,
"grad_norm": 0.013791786506772041,
"learning_rate": 1.1719418993228883e-05,
"loss": 0.6585,
"step": 1345
},
{
"epoch": 0.5023723137036004,
"grad_norm": 0.011280239559710026,
"learning_rate": 1.1655388017667812e-05,
"loss": 0.5919,
"step": 1350
},
{
"epoch": 0.5042329519025026,
"grad_norm": 0.015436794608831406,
"learning_rate": 1.159128718261189e-05,
"loss": 0.6632,
"step": 1355
},
{
"epoch": 0.5060935901014048,
"grad_norm": 0.015739573165774345,
"learning_rate": 1.1527119193198466e-05,
"loss": 0.6384,
"step": 1360
},
{
"epoch": 0.507954228300307,
"grad_norm": 0.01318281702697277,
"learning_rate": 1.146288675739889e-05,
"loss": 0.6312,
"step": 1365
},
{
"epoch": 0.5098148664992093,
"grad_norm": 0.01480270829051733,
"learning_rate": 1.1398592585904234e-05,
"loss": 0.6453,
"step": 1370
},
{
"epoch": 0.5116755046981114,
"grad_norm": 0.011967113241553307,
"learning_rate": 1.133423939201089e-05,
"loss": 0.6335,
"step": 1375
},
{
"epoch": 0.5135361428970137,
"grad_norm": 0.013129732571542263,
"learning_rate": 1.1269829891506081e-05,
"loss": 0.5852,
"step": 1380
},
{
"epoch": 0.5153967810959159,
"grad_norm": 0.0132956113666296,
"learning_rate": 1.1205366802553231e-05,
"loss": 0.6477,
"step": 1385
},
{
"epoch": 0.5172574192948182,
"grad_norm": 0.01314165536314249,
"learning_rate": 1.1140852845577273e-05,
"loss": 0.6441,
"step": 1390
},
{
"epoch": 0.5191180574937203,
"grad_norm": 0.01178740430623293,
"learning_rate": 1.1076290743149827e-05,
"loss": 0.6035,
"step": 1395
},
{
"epoch": 0.5209786956926226,
"grad_norm": 0.013940893113613129,
"learning_rate": 1.1011683219874324e-05,
"loss": 0.6492,
"step": 1400
},
{
"epoch": 0.5209786956926226,
"eval_loss": 0.644782543182373,
"eval_runtime": 26.5697,
"eval_samples_per_second": 4.78,
"eval_steps_per_second": 4.78,
"step": 1400
},
{
"epoch": 0.5228393338915248,
"grad_norm": 0.013468287885189056,
"learning_rate": 1.0947033002271001e-05,
"loss": 0.6135,
"step": 1405
},
{
"epoch": 0.5246999720904271,
"grad_norm": 0.015101495198905468,
"learning_rate": 1.0882342818661859e-05,
"loss": 0.6449,
"step": 1410
},
{
"epoch": 0.5265606102893292,
"grad_norm": 0.013072527013719082,
"learning_rate": 1.0817615399055513e-05,
"loss": 0.6252,
"step": 1415
},
{
"epoch": 0.5284212484882315,
"grad_norm": 0.01259040180593729,
"learning_rate": 1.075285347503198e-05,
"loss": 0.6274,
"step": 1420
},
{
"epoch": 0.5302818866871337,
"grad_norm": 0.014084907248616219,
"learning_rate": 1.0688059779627417e-05,
"loss": 0.6298,
"step": 1425
},
{
"epoch": 0.5321425248860359,
"grad_norm": 0.015118095092475414,
"learning_rate": 1.0623237047218771e-05,
"loss": 0.6638,
"step": 1430
},
{
"epoch": 0.5340031630849381,
"grad_norm": 0.013168774545192719,
"learning_rate": 1.0558388013408378e-05,
"loss": 0.6134,
"step": 1435
},
{
"epoch": 0.5358638012838404,
"grad_norm": 0.011875185184180737,
"learning_rate": 1.0493515414908542e-05,
"loss": 0.6396,
"step": 1440
},
{
"epoch": 0.5377244394827426,
"grad_norm": 0.013676362112164497,
"learning_rate": 1.0428621989426016e-05,
"loss": 0.6286,
"step": 1445
},
{
"epoch": 0.5395850776816448,
"grad_norm": 0.012775209732353687,
"learning_rate": 1.0363710475546483e-05,
"loss": 0.6156,
"step": 1450
},
{
"epoch": 0.541445715880547,
"grad_norm": 0.014763396233320236,
"learning_rate": 1.0298783612618977e-05,
"loss": 0.6713,
"step": 1455
},
{
"epoch": 0.5433063540794493,
"grad_norm": 0.014012233354151249,
"learning_rate": 1.0233844140640287e-05,
"loss": 0.5887,
"step": 1460
},
{
"epoch": 0.5451669922783515,
"grad_norm": 0.013994456268846989,
"learning_rate": 1.0168894800139311e-05,
"loss": 0.6509,
"step": 1465
},
{
"epoch": 0.5470276304772537,
"grad_norm": 0.014792009256780148,
"learning_rate": 1.0103938332061422e-05,
"loss": 0.6434,
"step": 1470
},
{
"epoch": 0.5488882686761559,
"grad_norm": 0.01255644578486681,
"learning_rate": 1.0038977477652779e-05,
"loss": 0.6407,
"step": 1475
},
{
"epoch": 0.5507489068750582,
"grad_norm": 0.013431582599878311,
"learning_rate": 9.974014978344646e-06,
"loss": 0.6528,
"step": 1480
},
{
"epoch": 0.5526095450739603,
"grad_norm": 0.0123568931594491,
"learning_rate": 9.909053575637717e-06,
"loss": 0.602,
"step": 1485
},
{
"epoch": 0.5544701832728626,
"grad_norm": 0.014569776132702827,
"learning_rate": 9.844096010986392e-06,
"loss": 0.6268,
"step": 1490
},
{
"epoch": 0.5563308214717648,
"grad_norm": 0.012518184259533882,
"learning_rate": 9.779145025683114e-06,
"loss": 0.5936,
"step": 1495
},
{
"epoch": 0.5581914596706671,
"grad_norm": 0.011615007184445858,
"learning_rate": 9.714203360742666e-06,
"loss": 0.6275,
"step": 1500
},
{
"epoch": 0.5581914596706671,
"eval_loss": 0.6442868709564209,
"eval_runtime": 26.5597,
"eval_samples_per_second": 4.782,
"eval_steps_per_second": 4.782,
"step": 1500
},
{
"epoch": 0.5600520978695692,
"grad_norm": 0.013426556251943111,
"learning_rate": 9.649273756786486e-06,
"loss": 0.6291,
"step": 1505
},
{
"epoch": 0.5619127360684715,
"grad_norm": 0.01306986529380083,
"learning_rate": 9.584358953927043e-06,
"loss": 0.6211,
"step": 1510
},
{
"epoch": 0.5637733742673737,
"grad_norm": 0.012594708241522312,
"learning_rate": 9.519461691652169e-06,
"loss": 0.5803,
"step": 1515
},
{
"epoch": 0.565634012466276,
"grad_norm": 0.014830299653112888,
"learning_rate": 9.454584708709462e-06,
"loss": 0.5976,
"step": 1520
},
{
"epoch": 0.5674946506651781,
"grad_norm": 0.013333864510059357,
"learning_rate": 9.389730742990714e-06,
"loss": 0.6154,
"step": 1525
},
{
"epoch": 0.5693552888640804,
"grad_norm": 0.013150627724826336,
"learning_rate": 9.324902531416348e-06,
"loss": 0.581,
"step": 1530
},
{
"epoch": 0.5712159270629826,
"grad_norm": 0.012874056585133076,
"learning_rate": 9.260102809819939e-06,
"loss": 0.6224,
"step": 1535
},
{
"epoch": 0.5730765652618848,
"grad_norm": 0.012678616680204868,
"learning_rate": 9.195334312832742e-06,
"loss": 0.6705,
"step": 1540
},
{
"epoch": 0.574937203460787,
"grad_norm": 0.011814710684120655,
"learning_rate": 9.1305997737683e-06,
"loss": 0.6103,
"step": 1545
},
{
"epoch": 0.5767978416596893,
"grad_norm": 0.013010908849537373,
"learning_rate": 9.065901924507085e-06,
"loss": 0.655,
"step": 1550
},
{
"epoch": 0.5786584798585915,
"grad_norm": 0.013622297905385494,
"learning_rate": 9.001243495381207e-06,
"loss": 0.5961,
"step": 1555
},
{
"epoch": 0.5805191180574937,
"grad_norm": 0.013876644894480705,
"learning_rate": 8.936627215059206e-06,
"loss": 0.6789,
"step": 1560
},
{
"epoch": 0.5823797562563959,
"grad_norm": 0.01281541958451271,
"learning_rate": 8.872055810430881e-06,
"loss": 0.6567,
"step": 1565
},
{
"epoch": 0.5842403944552982,
"grad_norm": 0.012182512320578098,
"learning_rate": 8.80753200649222e-06,
"loss": 0.6338,
"step": 1570
},
{
"epoch": 0.5861010326542004,
"grad_norm": 0.012718496844172478,
"learning_rate": 8.743058526230409e-06,
"loss": 0.6151,
"step": 1575
},
{
"epoch": 0.5879616708531026,
"grad_norm": 0.01301120687276125,
"learning_rate": 8.678638090508897e-06,
"loss": 0.6147,
"step": 1580
},
{
"epoch": 0.5898223090520048,
"grad_norm": 0.011715607717633247,
"learning_rate": 8.614273417952593e-06,
"loss": 0.5776,
"step": 1585
},
{
"epoch": 0.5916829472509071,
"grad_norm": 0.013846187852323055,
"learning_rate": 8.549967224833131e-06,
"loss": 0.6604,
"step": 1590
},
{
"epoch": 0.5935435854498092,
"grad_norm": 0.013312350027263165,
"learning_rate": 8.485722224954237e-06,
"loss": 0.6395,
"step": 1595
},
{
"epoch": 0.5954042236487115,
"grad_norm": 0.014313463121652603,
"learning_rate": 8.421541129537194e-06,
"loss": 0.6848,
"step": 1600
},
{
"epoch": 0.5954042236487115,
"eval_loss": 0.6439012885093689,
"eval_runtime": 26.5935,
"eval_samples_per_second": 4.776,
"eval_steps_per_second": 4.776,
"step": 1600
},
{
"epoch": 0.5972648618476137,
"grad_norm": 0.012196023017168045,
"learning_rate": 8.357426647106451e-06,
"loss": 0.6079,
"step": 1605
},
{
"epoch": 0.599125500046516,
"grad_norm": 0.0135956397280097,
"learning_rate": 8.293381483375293e-06,
"loss": 0.6463,
"step": 1610
},
{
"epoch": 0.6009861382454181,
"grad_norm": 0.011982797645032406,
"learning_rate": 8.229408341131665e-06,
"loss": 0.6113,
"step": 1615
},
{
"epoch": 0.6028467764443204,
"grad_norm": 0.014127896167337894,
"learning_rate": 8.165509920124125e-06,
"loss": 0.6602,
"step": 1620
},
{
"epoch": 0.6047074146432226,
"grad_norm": 0.012588880024850368,
"learning_rate": 8.10168891694789e-06,
"loss": 0.6376,
"step": 1625
},
{
"epoch": 0.6065680528421249,
"grad_norm": 0.012559432536363602,
"learning_rate": 8.037948024931039e-06,
"loss": 0.6336,
"step": 1630
},
{
"epoch": 0.608428691041027,
"grad_norm": 0.012455436401069164,
"learning_rate": 7.974289934020879e-06,
"loss": 0.6403,
"step": 1635
},
{
"epoch": 0.6102893292399293,
"grad_norm": 0.01241991762071848,
"learning_rate": 7.91071733067038e-06,
"loss": 0.6518,
"step": 1640
},
{
"epoch": 0.6121499674388315,
"grad_norm": 0.012328005395829678,
"learning_rate": 7.84723289772484e-06,
"loss": 0.6162,
"step": 1645
},
{
"epoch": 0.6140106056377338,
"grad_norm": 0.012782512232661247,
"learning_rate": 7.783839314308656e-06,
"loss": 0.6624,
"step": 1650
},
{
"epoch": 0.6158712438366359,
"grad_norm": 0.012050081044435501,
"learning_rate": 7.720539255712252e-06,
"loss": 0.6565,
"step": 1655
},
{
"epoch": 0.6177318820355382,
"grad_norm": 0.013794245198369026,
"learning_rate": 7.657335393279179e-06,
"loss": 0.6475,
"step": 1660
},
{
"epoch": 0.6195925202344404,
"grad_norm": 0.012430761009454727,
"learning_rate": 7.594230394293404e-06,
"loss": 0.5821,
"step": 1665
},
{
"epoch": 0.6214531584333426,
"grad_norm": 0.013150406070053577,
"learning_rate": 7.531226921866715e-06,
"loss": 0.6023,
"step": 1670
},
{
"epoch": 0.6233137966322448,
"grad_norm": 0.013466687873005867,
"learning_rate": 7.468327634826354e-06,
"loss": 0.637,
"step": 1675
},
{
"epoch": 0.6251744348311471,
"grad_norm": 0.014000273309648037,
"learning_rate": 7.405535187602809e-06,
"loss": 0.6113,
"step": 1680
},
{
"epoch": 0.6270350730300494,
"grad_norm": 0.013736380264163017,
"learning_rate": 7.3428522301177894e-06,
"loss": 0.5914,
"step": 1685
},
{
"epoch": 0.6288957112289515,
"grad_norm": 0.013854081742465496,
"learning_rate": 7.2802814076723896e-06,
"loss": 0.6744,
"step": 1690
},
{
"epoch": 0.6307563494278537,
"grad_norm": 0.012866591103374958,
"learning_rate": 7.217825360835475e-06,
"loss": 0.6209,
"step": 1695
},
{
"epoch": 0.632616987626756,
"grad_norm": 0.012788123451173306,
"learning_rate": 7.155486725332224e-06,
"loss": 0.5764,
"step": 1700
},
{
"epoch": 0.632616987626756,
"eval_loss": 0.6434745192527771,
"eval_runtime": 26.5737,
"eval_samples_per_second": 4.779,
"eval_steps_per_second": 4.779,
"step": 1700
},
{
"epoch": 0.6344776258256583,
"grad_norm": 0.013442011550068855,
"learning_rate": 7.093268131932905e-06,
"loss": 0.6522,
"step": 1705
},
{
"epoch": 0.6363382640245604,
"grad_norm": 0.013659958727657795,
"learning_rate": 7.03117220634187e-06,
"loss": 0.5949,
"step": 1710
},
{
"epoch": 0.6381989022234626,
"grad_norm": 0.013367819599807262,
"learning_rate": 6.9692015690867135e-06,
"loss": 0.5959,
"step": 1715
},
{
"epoch": 0.6400595404223649,
"grad_norm": 0.013543561100959778,
"learning_rate": 6.9073588354077125e-06,
"loss": 0.6539,
"step": 1720
},
{
"epoch": 0.641920178621267,
"grad_norm": 0.014200146310031414,
"learning_rate": 6.845646615147445e-06,
"loss": 0.6438,
"step": 1725
},
{
"epoch": 0.6437808168201693,
"grad_norm": 0.012410931289196014,
"learning_rate": 6.784067512640666e-06,
"loss": 0.6035,
"step": 1730
},
{
"epoch": 0.6456414550190716,
"grad_norm": 0.01299245934933424,
"learning_rate": 6.7226241266043735e-06,
"loss": 0.6507,
"step": 1735
},
{
"epoch": 0.6475020932179738,
"grad_norm": 0.013046164996922016,
"learning_rate": 6.661319050028167e-06,
"loss": 0.6277,
"step": 1740
},
{
"epoch": 0.649362731416876,
"grad_norm": 0.013011117465794086,
"learning_rate": 6.600154870064812e-06,
"loss": 0.6415,
"step": 1745
},
{
"epoch": 0.6512233696157782,
"grad_norm": 0.012195507064461708,
"learning_rate": 6.53913416792105e-06,
"loss": 0.5718,
"step": 1750
},
{
"epoch": 0.6530840078146805,
"grad_norm": 0.012789854779839516,
"learning_rate": 6.478259518748675e-06,
"loss": 0.5963,
"step": 1755
},
{
"epoch": 0.6549446460135827,
"grad_norm": 0.01217294204980135,
"learning_rate": 6.41753349153587e-06,
"loss": 0.6361,
"step": 1760
},
{
"epoch": 0.6568052842124849,
"grad_norm": 0.014139696955680847,
"learning_rate": 6.356958648998762e-06,
"loss": 0.6321,
"step": 1765
},
{
"epoch": 0.6586659224113871,
"grad_norm": 0.014064906165003777,
"learning_rate": 6.296537547473302e-06,
"loss": 0.6519,
"step": 1770
},
{
"epoch": 0.6605265606102894,
"grad_norm": 0.012787656858563423,
"learning_rate": 6.236272736807378e-06,
"loss": 0.6033,
"step": 1775
},
{
"epoch": 0.6623871988091915,
"grad_norm": 0.012475831434130669,
"learning_rate": 6.176166760253196e-06,
"loss": 0.5947,
"step": 1780
},
{
"epoch": 0.6642478370080938,
"grad_norm": 0.013439202681183815,
"learning_rate": 6.116222154359952e-06,
"loss": 0.636,
"step": 1785
},
{
"epoch": 0.666108475206996,
"grad_norm": 0.01295219361782074,
"learning_rate": 6.056441448866817e-06,
"loss": 0.6203,
"step": 1790
},
{
"epoch": 0.6679691134058983,
"grad_norm": 0.012275603599846363,
"learning_rate": 5.996827166596129e-06,
"loss": 0.6743,
"step": 1795
},
{
"epoch": 0.6698297516048004,
"grad_norm": 0.01329441275447607,
"learning_rate": 5.937381823346964e-06,
"loss": 0.5975,
"step": 1800
},
{
"epoch": 0.6698297516048004,
"eval_loss": 0.6431623101234436,
"eval_runtime": 26.5644,
"eval_samples_per_second": 4.781,
"eval_steps_per_second": 4.781,
"step": 1800
},
{
"epoch": 0.6716903898037027,
"grad_norm": 0.013007073663175106,
"learning_rate": 5.878107927788962e-06,
"loss": 0.6165,
"step": 1805
},
{
"epoch": 0.6735510280026049,
"grad_norm": 0.013040522113442421,
"learning_rate": 5.819007981356441e-06,
"loss": 0.6107,
"step": 1810
},
{
"epoch": 0.6754116662015072,
"grad_norm": 0.014542641118168831,
"learning_rate": 5.760084478142842e-06,
"loss": 0.6284,
"step": 1815
},
{
"epoch": 0.6772723044004093,
"grad_norm": 0.01339266262948513,
"learning_rate": 5.701339904795486e-06,
"loss": 0.6228,
"step": 1820
},
{
"epoch": 0.6791329425993116,
"grad_norm": 0.013091943226754665,
"learning_rate": 5.642776740410618e-06,
"loss": 0.5995,
"step": 1825
},
{
"epoch": 0.6809935807982138,
"grad_norm": 0.013321259059011936,
"learning_rate": 5.584397456428785e-06,
"loss": 0.627,
"step": 1830
},
{
"epoch": 0.682854218997116,
"grad_norm": 0.012838364578783512,
"learning_rate": 5.5262045165305615e-06,
"loss": 0.658,
"step": 1835
},
{
"epoch": 0.6847148571960182,
"grad_norm": 0.01306592021137476,
"learning_rate": 5.468200376532552e-06,
"loss": 0.6756,
"step": 1840
},
{
"epoch": 0.6865754953949205,
"grad_norm": 0.014314945787191391,
"learning_rate": 5.410387484283767e-06,
"loss": 0.6598,
"step": 1845
},
{
"epoch": 0.6884361335938227,
"grad_norm": 0.013973386958241463,
"learning_rate": 5.352768279562315e-06,
"loss": 0.6535,
"step": 1850
},
{
"epoch": 0.6902967717927249,
"grad_norm": 0.012864621356129646,
"learning_rate": 5.295345193972445e-06,
"loss": 0.6422,
"step": 1855
},
{
"epoch": 0.6921574099916271,
"grad_norm": 0.014392906799912453,
"learning_rate": 5.238120650841925e-06,
"loss": 0.664,
"step": 1860
},
{
"epoch": 0.6940180481905294,
"grad_norm": 0.01327193807810545,
"learning_rate": 5.18109706511978e-06,
"loss": 0.6179,
"step": 1865
},
{
"epoch": 0.6958786863894316,
"grad_norm": 0.01355487760156393,
"learning_rate": 5.124276843274372e-06,
"loss": 0.605,
"step": 1870
},
{
"epoch": 0.6977393245883338,
"grad_norm": 0.012646518647670746,
"learning_rate": 5.067662383191845e-06,
"loss": 0.6608,
"step": 1875
},
{
"epoch": 0.699599962787236,
"grad_norm": 0.013363865204155445,
"learning_rate": 5.011256074074945e-06,
"loss": 0.6975,
"step": 1880
},
{
"epoch": 0.7014606009861383,
"grad_norm": 0.013032359071075916,
"learning_rate": 4.955060296342163e-06,
"loss": 0.6041,
"step": 1885
},
{
"epoch": 0.7033212391850404,
"grad_norm": 0.012049228884279728,
"learning_rate": 4.899077421527304e-06,
"loss": 0.6078,
"step": 1890
},
{
"epoch": 0.7051818773839427,
"grad_norm": 0.013416007161140442,
"learning_rate": 4.843309812179405e-06,
"loss": 0.6514,
"step": 1895
},
{
"epoch": 0.7070425155828449,
"grad_norm": 0.013701760210096836,
"learning_rate": 4.787759821763017e-06,
"loss": 0.6606,
"step": 1900
},
{
"epoch": 0.7070425155828449,
"eval_loss": 0.6429811716079712,
"eval_runtime": 26.563,
"eval_samples_per_second": 4.781,
"eval_steps_per_second": 4.781,
"step": 1900
},
{
"epoch": 0.7089031537817472,
"grad_norm": 0.013226517476141453,
"learning_rate": 4.732429794558887e-06,
"loss": 0.6391,
"step": 1905
},
{
"epoch": 0.7107637919806493,
"grad_norm": 0.012964668683707714,
"learning_rate": 4.677322065565039e-06,
"loss": 0.6768,
"step": 1910
},
{
"epoch": 0.7126244301795516,
"grad_norm": 0.01224041823297739,
"learning_rate": 4.622438960398234e-06,
"loss": 0.645,
"step": 1915
},
{
"epoch": 0.7144850683784538,
"grad_norm": 0.013205330818891525,
"learning_rate": 4.567782795195816e-06,
"loss": 0.6051,
"step": 1920
},
{
"epoch": 0.7163457065773561,
"grad_norm": 0.01323428563773632,
"learning_rate": 4.5133558765179576e-06,
"loss": 0.6113,
"step": 1925
},
{
"epoch": 0.7182063447762582,
"grad_norm": 0.012786868028342724,
"learning_rate": 4.459160501250358e-06,
"loss": 0.6677,
"step": 1930
},
{
"epoch": 0.7200669829751605,
"grad_norm": 0.010466697625815868,
"learning_rate": 4.405198956507272e-06,
"loss": 0.586,
"step": 1935
},
{
"epoch": 0.7219276211740627,
"grad_norm": 0.012410039082169533,
"learning_rate": 4.35147351953501e-06,
"loss": 0.6111,
"step": 1940
},
{
"epoch": 0.723788259372965,
"grad_norm": 0.013155271299183369,
"learning_rate": 4.297986457615836e-06,
"loss": 0.6404,
"step": 1945
},
{
"epoch": 0.7256488975718671,
"grad_norm": 0.01472384575754404,
"learning_rate": 4.244740027972275e-06,
"loss": 0.6553,
"step": 1950
},
{
"epoch": 0.7275095357707694,
"grad_norm": 0.012972739525139332,
"learning_rate": 4.191736477671864e-06,
"loss": 0.6613,
"step": 1955
},
{
"epoch": 0.7293701739696716,
"grad_norm": 0.013699792325496674,
"learning_rate": 4.138978043532332e-06,
"loss": 0.6178,
"step": 1960
},
{
"epoch": 0.7312308121685738,
"grad_norm": 0.012072841636836529,
"learning_rate": 4.086466952027171e-06,
"loss": 0.5865,
"step": 1965
},
{
"epoch": 0.733091450367476,
"grad_norm": 0.013510013930499554,
"learning_rate": 4.034205419191709e-06,
"loss": 0.6387,
"step": 1970
},
{
"epoch": 0.7349520885663783,
"grad_norm": 0.01405192632228136,
"learning_rate": 3.982195650529583e-06,
"loss": 0.5677,
"step": 1975
},
{
"epoch": 0.7368127267652805,
"grad_norm": 0.012013067491352558,
"learning_rate": 3.930439840919652e-06,
"loss": 0.5868,
"step": 1980
},
{
"epoch": 0.7386733649641827,
"grad_norm": 0.01283415500074625,
"learning_rate": 3.878940174523371e-06,
"loss": 0.597,
"step": 1985
},
{
"epoch": 0.7405340031630849,
"grad_norm": 0.013413486070930958,
"learning_rate": 3.827698824692643e-06,
"loss": 0.6074,
"step": 1990
},
{
"epoch": 0.7423946413619872,
"grad_norm": 0.013137887232005596,
"learning_rate": 3.776717953878064e-06,
"loss": 0.6599,
"step": 1995
},
{
"epoch": 0.7442552795608894,
"grad_norm": 0.013435564935207367,
"learning_rate": 3.725999713537689e-06,
"loss": 0.6191,
"step": 2000
},
{
"epoch": 0.7442552795608894,
"eval_loss": 0.6428595185279846,
"eval_runtime": 26.6061,
"eval_samples_per_second": 4.773,
"eval_steps_per_second": 4.773,
"step": 2000
},
{
"epoch": 0.7461159177597916,
"grad_norm": 0.013417122885584831,
"learning_rate": 3.6755462440462288e-06,
"loss": 0.6012,
"step": 2005
},
{
"epoch": 0.7479765559586938,
"grad_norm": 0.01378430612385273,
"learning_rate": 3.625359674604725e-06,
"loss": 0.607,
"step": 2010
},
{
"epoch": 0.7498371941575961,
"grad_norm": 0.012614194303750992,
"learning_rate": 3.5754421231506953e-06,
"loss": 0.6364,
"step": 2015
},
{
"epoch": 0.7516978323564982,
"grad_norm": 0.013113941997289658,
"learning_rate": 3.5257956962687545e-06,
"loss": 0.6148,
"step": 2020
},
{
"epoch": 0.7535584705554005,
"grad_norm": 0.012877865694463253,
"learning_rate": 3.476422489101713e-06,
"loss": 0.6369,
"step": 2025
},
{
"epoch": 0.7554191087543027,
"grad_norm": 0.013343026861548424,
"learning_rate": 3.427324585262156e-06,
"loss": 0.6257,
"step": 2030
},
{
"epoch": 0.757279746953205,
"grad_norm": 0.01313395518809557,
"learning_rate": 3.3785040567445282e-06,
"loss": 0.6301,
"step": 2035
},
{
"epoch": 0.7591403851521071,
"grad_norm": 0.013584131374955177,
"learning_rate": 3.329962963837661e-06,
"loss": 0.6244,
"step": 2040
},
{
"epoch": 0.7610010233510094,
"grad_norm": 0.011685586534440517,
"learning_rate": 3.281703355037854e-06,
"loss": 0.5818,
"step": 2045
},
{
"epoch": 0.7628616615499116,
"grad_norm": 0.013454140163958073,
"learning_rate": 3.233727266962425e-06,
"loss": 0.6151,
"step": 2050
},
{
"epoch": 0.7647222997488139,
"grad_norm": 0.012276149354875088,
"learning_rate": 3.186036724263748e-06,
"loss": 0.621,
"step": 2055
},
{
"epoch": 0.766582937947716,
"grad_norm": 0.013796020299196243,
"learning_rate": 3.138633739543805e-06,
"loss": 0.6466,
"step": 2060
},
{
"epoch": 0.7684435761466183,
"grad_norm": 0.013075289316475391,
"learning_rate": 3.0915203132692805e-06,
"loss": 0.6116,
"step": 2065
},
{
"epoch": 0.7703042143455205,
"grad_norm": 0.014096383936703205,
"learning_rate": 3.0446984336871144e-06,
"loss": 0.5877,
"step": 2070
},
{
"epoch": 0.7721648525444227,
"grad_norm": 0.013074836693704128,
"learning_rate": 2.998170076740601e-06,
"loss": 0.5829,
"step": 2075
},
{
"epoch": 0.7740254907433249,
"grad_norm": 0.01671191304922104,
"learning_rate": 2.951937205986004e-06,
"loss": 0.6439,
"step": 2080
},
{
"epoch": 0.7758861289422272,
"grad_norm": 0.013207321055233479,
"learning_rate": 2.9060017725096943e-06,
"loss": 0.621,
"step": 2085
},
{
"epoch": 0.7777467671411294,
"grad_norm": 0.014768741093575954,
"learning_rate": 2.8603657148458053e-06,
"loss": 0.6272,
"step": 2090
},
{
"epoch": 0.7796074053400316,
"grad_norm": 0.015293040312826633,
"learning_rate": 2.8150309588944304e-06,
"loss": 0.6388,
"step": 2095
},
{
"epoch": 0.7814680435389338,
"grad_norm": 0.01271377969533205,
"learning_rate": 2.769999417840341e-06,
"loss": 0.6249,
"step": 2100
},
{
"epoch": 0.7814680435389338,
"eval_loss": 0.6426796913146973,
"eval_runtime": 26.6067,
"eval_samples_per_second": 4.773,
"eval_steps_per_second": 4.773,
"step": 2100
},
{
"epoch": 0.7833286817378361,
"grad_norm": 0.013834511861205101,
"learning_rate": 2.7252729920722564e-06,
"loss": 0.6254,
"step": 2105
},
{
"epoch": 0.7851893199367384,
"grad_norm": 0.020366957411170006,
"learning_rate": 2.680853569102633e-06,
"loss": 0.632,
"step": 2110
},
{
"epoch": 0.7870499581356405,
"grad_norm": 0.012656195089221,
"learning_rate": 2.6367430234880286e-06,
"loss": 0.6274,
"step": 2115
},
{
"epoch": 0.7889105963345427,
"grad_norm": 0.01308165118098259,
"learning_rate": 2.5929432167499658e-06,
"loss": 0.6457,
"step": 2120
},
{
"epoch": 0.790771234533445,
"grad_norm": 0.012838170863687992,
"learning_rate": 2.5494559972963928e-06,
"loss": 0.6436,
"step": 2125
},
{
"epoch": 0.7926318727323471,
"grad_norm": 0.012966095469892025,
"learning_rate": 2.5062832003436833e-06,
"loss": 0.6449,
"step": 2130
},
{
"epoch": 0.7944925109312494,
"grad_norm": 0.013781096786260605,
"learning_rate": 2.463426647839173e-06,
"loss": 0.584,
"step": 2135
},
{
"epoch": 0.7963531491301516,
"grad_norm": 0.012445746921002865,
"learning_rate": 2.420888148384265e-06,
"loss": 0.6397,
"step": 2140
},
{
"epoch": 0.7982137873290539,
"grad_norm": 0.012599524110555649,
"learning_rate": 2.378669497158138e-06,
"loss": 0.5974,
"step": 2145
},
{
"epoch": 0.800074425527956,
"grad_norm": 0.011499716900289059,
"learning_rate": 2.3367724758419495e-06,
"loss": 0.5552,
"step": 2150
},
{
"epoch": 0.8019350637268583,
"grad_norm": 0.012771397829055786,
"learning_rate": 2.2951988525436695e-06,
"loss": 0.666,
"step": 2155
},
{
"epoch": 0.8037957019257606,
"grad_norm": 0.013215843588113785,
"learning_rate": 2.2539503817234553e-06,
"loss": 0.5925,
"step": 2160
},
{
"epoch": 0.8056563401246628,
"grad_norm": 0.014011417515575886,
"learning_rate": 2.2130288041196135e-06,
"loss": 0.6216,
"step": 2165
},
{
"epoch": 0.807516978323565,
"grad_norm": 0.014169846661388874,
"learning_rate": 2.1724358466751394e-06,
"loss": 0.625,
"step": 2170
},
{
"epoch": 0.8093776165224672,
"grad_norm": 0.013908619992434978,
"learning_rate": 2.132173222464834e-06,
"loss": 0.6641,
"step": 2175
},
{
"epoch": 0.8112382547213695,
"grad_norm": 0.01255475077778101,
"learning_rate": 2.092242630623016e-06,
"loss": 0.6135,
"step": 2180
},
{
"epoch": 0.8130988929202716,
"grad_norm": 0.01320129819214344,
"learning_rate": 2.0526457562718074e-06,
"loss": 0.5893,
"step": 2185
},
{
"epoch": 0.8149595311191739,
"grad_norm": 0.012163982726633549,
"learning_rate": 2.013384270450036e-06,
"loss": 0.6552,
"step": 2190
},
{
"epoch": 0.8168201693180761,
"grad_norm": 0.012990654446184635,
"learning_rate": 1.974459830042691e-06,
"loss": 0.6309,
"step": 2195
},
{
"epoch": 0.8186808075169784,
"grad_norm": 0.015373739413917065,
"learning_rate": 1.9358740777110154e-06,
"loss": 0.6761,
"step": 2200
},
{
"epoch": 0.8186808075169784,
"eval_loss": 0.6425639390945435,
"eval_runtime": 26.5817,
"eval_samples_per_second": 4.778,
"eval_steps_per_second": 4.778,
"step": 2200
},
{
"epoch": 0.8205414457158805,
"grad_norm": 0.012748058885335922,
"learning_rate": 1.8976286418231916e-06,
"loss": 0.6313,
"step": 2205
},
{
"epoch": 0.8224020839147828,
"grad_norm": 0.014108446426689625,
"learning_rate": 1.8597251363856061e-06,
"loss": 0.5997,
"step": 2210
},
{
"epoch": 0.824262722113685,
"grad_norm": 0.0133186811581254,
"learning_rate": 1.8221651609747337e-06,
"loss": 0.6732,
"step": 2215
},
{
"epoch": 0.8261233603125873,
"grad_norm": 0.012968887574970722,
"learning_rate": 1.7849503006696566e-06,
"loss": 0.6129,
"step": 2220
},
{
"epoch": 0.8279839985114894,
"grad_norm": 0.012473355047404766,
"learning_rate": 1.7480821259851488e-06,
"loss": 0.6131,
"step": 2225
},
{
"epoch": 0.8298446367103917,
"grad_norm": 0.012497167102992535,
"learning_rate": 1.7115621928054105e-06,
"loss": 0.6199,
"step": 2230
},
{
"epoch": 0.8317052749092939,
"grad_norm": 0.014741111546754837,
"learning_rate": 1.6753920423184022e-06,
"loss": 0.6159,
"step": 2235
},
{
"epoch": 0.8335659131081962,
"grad_norm": 0.013057067058980465,
"learning_rate": 1.6395732009508058e-06,
"loss": 0.5898,
"step": 2240
},
{
"epoch": 0.8354265513070983,
"grad_norm": 0.014299440197646618,
"learning_rate": 1.60410718030361e-06,
"loss": 0.6486,
"step": 2245
},
{
"epoch": 0.8372871895060006,
"grad_norm": 0.013178296387195587,
"learning_rate": 1.568995477088323e-06,
"loss": 0.6254,
"step": 2250
},
{
"epoch": 0.8391478277049028,
"grad_norm": 0.013072814792394638,
"learning_rate": 1.5342395730637904e-06,
"loss": 0.6385,
"step": 2255
},
{
"epoch": 0.841008465903805,
"grad_norm": 0.014187455177307129,
"learning_rate": 1.4998409349736841e-06,
"loss": 0.6458,
"step": 2260
},
{
"epoch": 0.8428691041027072,
"grad_norm": 0.011844666674733162,
"learning_rate": 1.4658010144846001e-06,
"loss": 0.6556,
"step": 2265
},
{
"epoch": 0.8447297423016095,
"grad_norm": 0.013700997456908226,
"learning_rate": 1.432121248124786e-06,
"loss": 0.6381,
"step": 2270
},
{
"epoch": 0.8465903805005117,
"grad_norm": 0.014204096049070358,
"learning_rate": 1.3988030572235212e-06,
"loss": 0.624,
"step": 2275
},
{
"epoch": 0.8484510186994139,
"grad_norm": 0.013039126060903072,
"learning_rate": 1.3658478478511416e-06,
"loss": 0.593,
"step": 2280
},
{
"epoch": 0.8503116568983161,
"grad_norm": 0.014153816737234592,
"learning_rate": 1.333257010759702e-06,
"loss": 0.5909,
"step": 2285
},
{
"epoch": 0.8521722950972184,
"grad_norm": 0.012623702175915241,
"learning_rate": 1.3010319213242762e-06,
"loss": 0.6006,
"step": 2290
},
{
"epoch": 0.8540329332961206,
"grad_norm": 0.013684497214853764,
"learning_rate": 1.2691739394849089e-06,
"loss": 0.6112,
"step": 2295
},
{
"epoch": 0.8558935714950228,
"grad_norm": 0.012923210859298706,
"learning_rate": 1.2376844096892526e-06,
"loss": 0.6239,
"step": 2300
},
{
"epoch": 0.8558935714950228,
"eval_loss": 0.6425209641456604,
"eval_runtime": 26.5582,
"eval_samples_per_second": 4.782,
"eval_steps_per_second": 4.782,
"step": 2300
},
{
"epoch": 0.857754209693925,
"grad_norm": 0.01325127761811018,
"learning_rate": 1.2065646608357972e-06,
"loss": 0.6537,
"step": 2305
},
{
"epoch": 0.8596148478928273,
"grad_norm": 0.01215402316302061,
"learning_rate": 1.1758160062178093e-06,
"loss": 0.634,
"step": 2310
},
{
"epoch": 0.8614754860917294,
"grad_norm": 0.015918321907520294,
"learning_rate": 1.1454397434679022e-06,
"loss": 0.6296,
"step": 2315
},
{
"epoch": 0.8633361242906317,
"grad_norm": 0.011971482075750828,
"learning_rate": 1.1154371545032738e-06,
"loss": 0.5983,
"step": 2320
},
{
"epoch": 0.8651967624895339,
"grad_norm": 0.01542737614363432,
"learning_rate": 1.0858095054716111e-06,
"loss": 0.6468,
"step": 2325
},
{
"epoch": 0.8670574006884362,
"grad_norm": 0.012823596596717834,
"learning_rate": 1.0565580466976566e-06,
"loss": 0.6222,
"step": 2330
},
{
"epoch": 0.8689180388873383,
"grad_norm": 0.012822597287595272,
"learning_rate": 1.027684012630441e-06,
"loss": 0.6385,
"step": 2335
},
{
"epoch": 0.8707786770862406,
"grad_norm": 0.014365943148732185,
"learning_rate": 9.991886217911851e-07,
"loss": 0.6541,
"step": 2340
},
{
"epoch": 0.8726393152851428,
"grad_norm": 0.013635417446494102,
"learning_rate": 9.710730767218913e-07,
"loss": 0.6323,
"step": 2345
},
{
"epoch": 0.8744999534840451,
"grad_norm": 0.012839280068874359,
"learning_rate": 9.433385639345705e-07,
"loss": 0.5944,
"step": 2350
},
{
"epoch": 0.8763605916829472,
"grad_norm": 0.0141257019713521,
"learning_rate": 9.159862538611908e-07,
"loss": 0.6632,
"step": 2355
},
{
"epoch": 0.8782212298818495,
"grad_norm": 0.013571621850132942,
"learning_rate": 8.890173008042768e-07,
"loss": 0.6086,
"step": 2360
},
{
"epoch": 0.8800818680807517,
"grad_norm": 0.01433682069182396,
"learning_rate": 8.624328428881945e-07,
"loss": 0.6482,
"step": 2365
},
{
"epoch": 0.8819425062796539,
"grad_norm": 0.014090972021222115,
"learning_rate": 8.36234002011117e-07,
"loss": 0.6335,
"step": 2370
},
{
"epoch": 0.8838031444785561,
"grad_norm": 0.013004067353904247,
"learning_rate": 8.10421883797694e-07,
"loss": 0.6145,
"step": 2375
},
{
"epoch": 0.8856637826774584,
"grad_norm": 0.01391025260090828,
"learning_rate": 7.849975775523777e-07,
"loss": 0.6415,
"step": 2380
},
{
"epoch": 0.8875244208763606,
"grad_norm": 0.012305272743105888,
"learning_rate": 7.599621562134596e-07,
"loss": 0.6462,
"step": 2385
},
{
"epoch": 0.8893850590752628,
"grad_norm": 0.013947544619441032,
"learning_rate": 7.35316676307789e-07,
"loss": 0.6266,
"step": 2390
},
{
"epoch": 0.891245697274165,
"grad_norm": 0.016055511310696602,
"learning_rate": 7.110621779061889e-07,
"loss": 0.6501,
"step": 2395
},
{
"epoch": 0.8931063354730673,
"grad_norm": 0.01238927897065878,
"learning_rate": 6.871996845795581e-07,
"loss": 0.5895,
"step": 2400
},
{
"epoch": 0.8931063354730673,
"eval_loss": 0.6424703598022461,
"eval_runtime": 26.6049,
"eval_samples_per_second": 4.774,
"eval_steps_per_second": 4.774,
"step": 2400
},
{
"epoch": 0.8949669736719695,
"grad_norm": 0.012343904934823513,
"learning_rate": 6.637302033556891e-07,
"loss": 0.6105,
"step": 2405
},
{
"epoch": 0.8968276118708717,
"grad_norm": 0.01289752684533596,
"learning_rate": 6.40654724676748e-07,
"loss": 0.6265,
"step": 2410
},
{
"epoch": 0.8986882500697739,
"grad_norm": 0.013334677554666996,
"learning_rate": 6.179742223574936e-07,
"loss": 0.6261,
"step": 2415
},
{
"epoch": 0.9005488882686762,
"grad_norm": 0.013512643985450268,
"learning_rate": 5.956896535441803e-07,
"loss": 0.5797,
"step": 2420
},
{
"epoch": 0.9024095264675783,
"grad_norm": 0.01403987966477871,
"learning_rate": 5.738019586741573e-07,
"loss": 0.616,
"step": 2425
},
{
"epoch": 0.9042701646664806,
"grad_norm": 0.013224196620285511,
"learning_rate": 5.523120614361821e-07,
"loss": 0.608,
"step": 2430
},
{
"epoch": 0.9061308028653828,
"grad_norm": 0.01317799836397171,
"learning_rate": 5.312208687314502e-07,
"loss": 0.6206,
"step": 2435
},
{
"epoch": 0.9079914410642851,
"grad_norm": 0.01357815321534872,
"learning_rate": 5.105292706353093e-07,
"loss": 0.6759,
"step": 2440
},
{
"epoch": 0.9098520792631872,
"grad_norm": 0.012360199354588985,
"learning_rate": 4.902381403597046e-07,
"loss": 0.6182,
"step": 2445
},
{
"epoch": 0.9117127174620895,
"grad_norm": 0.012978832237422466,
"learning_rate": 4.703483342163262e-07,
"loss": 0.5892,
"step": 2450
},
{
"epoch": 0.9135733556609917,
"grad_norm": 0.013323403894901276,
"learning_rate": 4.5086069158047143e-07,
"loss": 0.636,
"step": 2455
},
{
"epoch": 0.915433993859894,
"grad_norm": 0.013127562589943409,
"learning_rate": 4.3177603485562327e-07,
"loss": 0.5847,
"step": 2460
},
{
"epoch": 0.9172946320587961,
"grad_norm": 0.012560434639453888,
"learning_rate": 4.1309516943874196e-07,
"loss": 0.6073,
"step": 2465
},
{
"epoch": 0.9191552702576984,
"grad_norm": 0.01229447964578867,
"learning_rate": 3.9481888368627764e-07,
"loss": 0.5846,
"step": 2470
},
{
"epoch": 0.9210159084566006,
"grad_norm": 0.012554515153169632,
"learning_rate": 3.7694794888090025e-07,
"loss": 0.6186,
"step": 2475
},
{
"epoch": 0.9228765466555028,
"grad_norm": 0.011958773247897625,
"learning_rate": 3.594831191989523e-07,
"loss": 0.6217,
"step": 2480
},
{
"epoch": 0.924737184854405,
"grad_norm": 0.01367896143347025,
"learning_rate": 3.424251316786165e-07,
"loss": 0.6572,
"step": 2485
},
{
"epoch": 0.9265978230533073,
"grad_norm": 0.013063084334135056,
"learning_rate": 3.2577470618881726e-07,
"loss": 0.5973,
"step": 2490
},
{
"epoch": 0.9284584612522095,
"grad_norm": 0.012596881948411465,
"learning_rate": 3.095325453988385e-07,
"loss": 0.6153,
"step": 2495
},
{
"epoch": 0.9303190994511117,
"grad_norm": 0.013333328068256378,
"learning_rate": 2.9369933474867496e-07,
"loss": 0.6337,
"step": 2500
},
{
"epoch": 0.9303190994511117,
"eval_loss": 0.6424322128295898,
"eval_runtime": 26.5954,
"eval_samples_per_second": 4.775,
"eval_steps_per_second": 4.775,
"step": 2500
},
{
"epoch": 0.9321797376500139,
"grad_norm": 0.012263627722859383,
"learning_rate": 2.7827574242009434e-07,
"loss": 0.6258,
"step": 2505
},
{
"epoch": 0.9340403758489162,
"grad_norm": 0.01295017171651125,
"learning_rate": 2.632624193084499e-07,
"loss": 0.6071,
"step": 2510
},
{
"epoch": 0.9359010140478184,
"grad_norm": 0.013952870853245258,
"learning_rate": 2.48659998995211e-07,
"loss": 0.6594,
"step": 2515
},
{
"epoch": 0.9377616522467206,
"grad_norm": 0.01351676881313324,
"learning_rate": 2.344690977212205e-07,
"loss": 0.6072,
"step": 2520
},
{
"epoch": 0.9396222904456228,
"grad_norm": 0.012661050073802471,
"learning_rate": 2.2069031436068643e-07,
"loss": 0.6152,
"step": 2525
},
{
"epoch": 0.9414829286445251,
"grad_norm": 0.013936568051576614,
"learning_rate": 2.0732423039591998e-07,
"loss": 0.6149,
"step": 2530
},
{
"epoch": 0.9433435668434274,
"grad_norm": 0.012601389549672604,
"learning_rate": 1.9437140989278624e-07,
"loss": 0.6205,
"step": 2535
},
{
"epoch": 0.9452042050423295,
"grad_norm": 0.011978083290159702,
"learning_rate": 1.8183239947690112e-07,
"loss": 0.6252,
"step": 2540
},
{
"epoch": 0.9470648432412317,
"grad_norm": 0.012075323611497879,
"learning_rate": 1.6970772831056637e-07,
"loss": 0.6438,
"step": 2545
},
{
"epoch": 0.948925481440134,
"grad_norm": 0.012837120331823826,
"learning_rate": 1.5799790807043857e-07,
"loss": 0.6304,
"step": 2550
},
{
"epoch": 0.9507861196390361,
"grad_norm": 0.012269029393792152,
"learning_rate": 1.467034329259287e-07,
"loss": 0.5938,
"step": 2555
},
{
"epoch": 0.9526467578379384,
"grad_norm": 0.01306453812867403,
"learning_rate": 1.358247795183587e-07,
"loss": 0.5981,
"step": 2560
},
{
"epoch": 0.9545073960368406,
"grad_norm": 0.012171389535069466,
"learning_rate": 1.2536240694083658e-07,
"loss": 0.5599,
"step": 2565
},
{
"epoch": 0.9563680342357429,
"grad_norm": 0.013338044285774231,
"learning_rate": 1.1531675671888621e-07,
"loss": 0.6366,
"step": 2570
},
{
"epoch": 0.958228672434645,
"grad_norm": 0.012387475930154324,
"learning_rate": 1.0568825279181572e-07,
"loss": 0.6376,
"step": 2575
},
{
"epoch": 0.9600893106335473,
"grad_norm": 0.013054094277322292,
"learning_rate": 9.647730149482614e-08,
"loss": 0.6386,
"step": 2580
},
{
"epoch": 0.9619499488324496,
"grad_norm": 0.01319506112486124,
"learning_rate": 8.768429154185853e-08,
"loss": 0.6067,
"step": 2585
},
{
"epoch": 0.9638105870313518,
"grad_norm": 0.013576803728938103,
"learning_rate": 7.930959400919924e-08,
"loss": 0.6258,
"step": 2590
},
{
"epoch": 0.965671225230254,
"grad_norm": 0.013341645710170269,
"learning_rate": 7.135356231981028e-08,
"loss": 0.6344,
"step": 2595
},
{
"epoch": 0.9675318634291562,
"grad_norm": 0.013685373589396477,
"learning_rate": 6.381653222842011e-08,
"loss": 0.6144,
"step": 2600
},
{
"epoch": 0.9675318634291562,
"eval_loss": 0.642424464225769,
"eval_runtime": 26.5716,
"eval_samples_per_second": 4.78,
"eval_steps_per_second": 4.78,
"step": 2600
},
{
"epoch": 0.9693925016280585,
"grad_norm": 0.018206071108579636,
"learning_rate": 5.6698821807354975e-08,
"loss": 0.5716,
"step": 2605
},
{
"epoch": 0.9712531398269606,
"grad_norm": 0.014082850888371468,
"learning_rate": 5.000073143310969e-08,
"loss": 0.641,
"step": 2610
},
{
"epoch": 0.9731137780258629,
"grad_norm": 0.01310745719820261,
"learning_rate": 4.3722543773681016e-08,
"loss": 0.6206,
"step": 2615
},
{
"epoch": 0.9749744162247651,
"grad_norm": 0.013652559369802475,
"learning_rate": 3.7864523776628414e-08,
"loss": 0.6405,
"step": 2620
},
{
"epoch": 0.9768350544236674,
"grad_norm": 0.01430275198072195,
"learning_rate": 3.242691865790071e-08,
"loss": 0.6191,
"step": 2625
},
{
"epoch": 0.9786956926225695,
"grad_norm": 0.013149100355803967,
"learning_rate": 2.7409957891397775e-08,
"loss": 0.6676,
"step": 2630
},
{
"epoch": 0.9805563308214718,
"grad_norm": 0.01429106667637825,
"learning_rate": 2.2813853199292745e-08,
"loss": 0.656,
"step": 2635
},
{
"epoch": 0.982416969020374,
"grad_norm": 0.012532561086118221,
"learning_rate": 1.8638798543090253e-08,
"loss": 0.6065,
"step": 2640
},
{
"epoch": 0.9842776072192763,
"grad_norm": 0.01296111661940813,
"learning_rate": 1.4884970115444097e-08,
"loss": 0.604,
"step": 2645
},
{
"epoch": 0.9861382454181784,
"grad_norm": 0.012804310768842697,
"learning_rate": 1.1552526332723191e-08,
"loss": 0.6236,
"step": 2650
},
{
"epoch": 0.9879988836170807,
"grad_norm": 0.012863220646977425,
"learning_rate": 8.641607828324682e-09,
"loss": 0.6245,
"step": 2655
},
{
"epoch": 0.9898595218159829,
"grad_norm": 0.013312343508005142,
"learning_rate": 6.152337446736489e-09,
"loss": 0.6561,
"step": 2660
},
{
"epoch": 0.991720160014885,
"grad_norm": 0.011470803059637547,
"learning_rate": 4.0848202383581e-09,
"loss": 0.5847,
"step": 2665
},
{
"epoch": 0.9935807982137873,
"grad_norm": 0.013984150253236294,
"learning_rate": 2.4391434550652403e-09,
"loss": 0.5881,
"step": 2670
},
{
"epoch": 0.9954414364126896,
"grad_norm": 0.012489933520555496,
"learning_rate": 1.2153765465250378e-09,
"loss": 0.6653,
"step": 2675
},
{
"epoch": 0.9973020746115918,
"grad_norm": 0.012180610559880733,
"learning_rate": 4.1357115726947674e-10,
"loss": 0.629,
"step": 2680
},
{
"epoch": 0.999162712810494,
"grad_norm": 0.014413848519325256,
"learning_rate": 3.376112451158875e-11,
"loss": 0.6525,
"step": 2685
},
{
"epoch": 0.9999069680900549,
"step": 2687,
"total_flos": 8.247315480402985e+17,
"train_loss": 0.638783668202896,
"train_runtime": 14784.0018,
"train_samples_per_second": 1.454,
"train_steps_per_second": 0.182
}
],
"logging_steps": 5,
"max_steps": 2687,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 8.247315480402985e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}