Instruments-8bit-3B-4Epoch-RanOrd / trainer_state.json
WangXFng's picture
Model save
c9accb5 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.9996055735491556,
"eval_steps": 500,
"global_step": 32956,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.030340496218815658,
"grad_norm": 1.0293707847595215,
"learning_rate": 9.924743581962737e-05,
"loss": 1.2653,
"step": 250
},
{
"epoch": 0.060680992437631316,
"grad_norm": 0.8869585990905762,
"learning_rate": 9.848880257328398e-05,
"loss": 0.757,
"step": 500
},
{
"epoch": 0.09102148865644698,
"grad_norm": 0.9423841834068298,
"learning_rate": 9.773016932694059e-05,
"loss": 0.6578,
"step": 750
},
{
"epoch": 0.12136198487526263,
"grad_norm": 0.7434495091438293,
"learning_rate": 9.697153608059721e-05,
"loss": 0.6204,
"step": 1000
},
{
"epoch": 0.1517024810940783,
"grad_norm": 0.6856955885887146,
"learning_rate": 9.621290283425382e-05,
"loss": 0.6006,
"step": 1250
},
{
"epoch": 0.18204297731289396,
"grad_norm": 0.6636335849761963,
"learning_rate": 9.545426958791042e-05,
"loss": 0.5822,
"step": 1500
},
{
"epoch": 0.2123834735317096,
"grad_norm": 0.6647191643714905,
"learning_rate": 9.469563634156705e-05,
"loss": 0.5681,
"step": 1750
},
{
"epoch": 0.24272396975052526,
"grad_norm": 0.825007438659668,
"learning_rate": 9.393700309522365e-05,
"loss": 0.5531,
"step": 2000
},
{
"epoch": 0.2730644659693409,
"grad_norm": 0.651090681552887,
"learning_rate": 9.317836984888026e-05,
"loss": 0.5387,
"step": 2250
},
{
"epoch": 0.3034049621881566,
"grad_norm": 0.7366721034049988,
"learning_rate": 9.241973660253687e-05,
"loss": 0.517,
"step": 2500
},
{
"epoch": 0.33374545840697223,
"grad_norm": 0.8244208693504333,
"learning_rate": 9.166110335619349e-05,
"loss": 0.4931,
"step": 2750
},
{
"epoch": 0.3640859546257879,
"grad_norm": 0.7282492518424988,
"learning_rate": 9.09024701098501e-05,
"loss": 0.4741,
"step": 3000
},
{
"epoch": 0.39442645084460354,
"grad_norm": 0.877441942691803,
"learning_rate": 9.014383686350671e-05,
"loss": 0.4543,
"step": 3250
},
{
"epoch": 0.4247669470634192,
"grad_norm": 0.8125002980232239,
"learning_rate": 8.938520361716333e-05,
"loss": 0.4301,
"step": 3500
},
{
"epoch": 0.4551074432822349,
"grad_norm": 0.7860125303268433,
"learning_rate": 8.862657037081994e-05,
"loss": 0.4165,
"step": 3750
},
{
"epoch": 0.48544793950105053,
"grad_norm": 0.8634310364723206,
"learning_rate": 8.786793712447655e-05,
"loss": 0.3964,
"step": 4000
},
{
"epoch": 0.5157884357198662,
"grad_norm": 0.8578837513923645,
"learning_rate": 8.710930387813317e-05,
"loss": 0.3847,
"step": 4250
},
{
"epoch": 0.5461289319386818,
"grad_norm": 0.6492015719413757,
"learning_rate": 8.635067063178978e-05,
"loss": 0.3763,
"step": 4500
},
{
"epoch": 0.5764694281574975,
"grad_norm": 0.7463727593421936,
"learning_rate": 8.559203738544639e-05,
"loss": 0.36,
"step": 4750
},
{
"epoch": 0.6068099243763132,
"grad_norm": 0.7797712683677673,
"learning_rate": 8.483340413910301e-05,
"loss": 0.3597,
"step": 5000
},
{
"epoch": 0.6371504205951288,
"grad_norm": 0.8424202799797058,
"learning_rate": 8.407477089275962e-05,
"loss": 0.344,
"step": 5250
},
{
"epoch": 0.6674909168139445,
"grad_norm": 0.7570486664772034,
"learning_rate": 8.331613764641622e-05,
"loss": 0.3402,
"step": 5500
},
{
"epoch": 0.6978314130327602,
"grad_norm": 0.741788923740387,
"learning_rate": 8.255750440007283e-05,
"loss": 0.3378,
"step": 5750
},
{
"epoch": 0.7281719092515758,
"grad_norm": 0.7841416597366333,
"learning_rate": 8.179887115372944e-05,
"loss": 0.3305,
"step": 6000
},
{
"epoch": 0.7585124054703914,
"grad_norm": 0.7679227590560913,
"learning_rate": 8.104023790738605e-05,
"loss": 0.3263,
"step": 6250
},
{
"epoch": 0.7888529016892071,
"grad_norm": 0.7030394673347473,
"learning_rate": 8.028160466104267e-05,
"loss": 0.3232,
"step": 6500
},
{
"epoch": 0.8191933979080228,
"grad_norm": 0.7032948136329651,
"learning_rate": 7.952297141469928e-05,
"loss": 0.3142,
"step": 6750
},
{
"epoch": 0.8495338941268384,
"grad_norm": 0.7051456570625305,
"learning_rate": 7.876433816835589e-05,
"loss": 0.3101,
"step": 7000
},
{
"epoch": 0.8798743903456541,
"grad_norm": 0.680454432964325,
"learning_rate": 7.80057049220125e-05,
"loss": 0.3128,
"step": 7250
},
{
"epoch": 0.9102148865644698,
"grad_norm": 0.6378083229064941,
"learning_rate": 7.724707167566912e-05,
"loss": 0.3058,
"step": 7500
},
{
"epoch": 0.9405553827832854,
"grad_norm": 0.6744751930236816,
"learning_rate": 7.648843842932573e-05,
"loss": 0.3021,
"step": 7750
},
{
"epoch": 0.9708958790021011,
"grad_norm": 0.6540088057518005,
"learning_rate": 7.572980518298233e-05,
"loss": 0.3028,
"step": 8000
},
{
"epoch": 1.0012363752209168,
"grad_norm": 0.6479789614677429,
"learning_rate": 7.497117193663896e-05,
"loss": 0.2993,
"step": 8250
},
{
"epoch": 1.0315768714397324,
"grad_norm": 0.696811854839325,
"learning_rate": 7.421253869029556e-05,
"loss": 0.29,
"step": 8500
},
{
"epoch": 1.061917367658548,
"grad_norm": 0.6420058608055115,
"learning_rate": 7.345390544395217e-05,
"loss": 0.2862,
"step": 8750
},
{
"epoch": 1.0922578638773637,
"grad_norm": 0.7226221561431885,
"learning_rate": 7.26952721976088e-05,
"loss": 0.2876,
"step": 9000
},
{
"epoch": 1.1225983600961793,
"grad_norm": 0.5515549182891846,
"learning_rate": 7.19366389512654e-05,
"loss": 0.284,
"step": 9250
},
{
"epoch": 1.1529388563149952,
"grad_norm": 0.5973398685455322,
"learning_rate": 7.117800570492201e-05,
"loss": 0.2828,
"step": 9500
},
{
"epoch": 1.1832793525338108,
"grad_norm": 0.6326724886894226,
"learning_rate": 7.041937245857862e-05,
"loss": 0.281,
"step": 9750
},
{
"epoch": 1.2136198487526264,
"grad_norm": 0.5345270037651062,
"learning_rate": 6.966073921223524e-05,
"loss": 0.2817,
"step": 10000
},
{
"epoch": 1.243960344971442,
"grad_norm": 0.6941035389900208,
"learning_rate": 6.890210596589185e-05,
"loss": 0.2783,
"step": 10250
},
{
"epoch": 1.2743008411902577,
"grad_norm": 0.499006062746048,
"learning_rate": 6.814347271954846e-05,
"loss": 0.2765,
"step": 10500
},
{
"epoch": 1.3046413374090733,
"grad_norm": 0.5554171204566956,
"learning_rate": 6.738483947320508e-05,
"loss": 0.2769,
"step": 10750
},
{
"epoch": 1.334981833627889,
"grad_norm": 0.6185881495475769,
"learning_rate": 6.662620622686169e-05,
"loss": 0.2737,
"step": 11000
},
{
"epoch": 1.3653223298467045,
"grad_norm": 0.5652614235877991,
"learning_rate": 6.58675729805183e-05,
"loss": 0.2736,
"step": 11250
},
{
"epoch": 1.3956628260655202,
"grad_norm": 0.602508008480072,
"learning_rate": 6.510893973417492e-05,
"loss": 0.2726,
"step": 11500
},
{
"epoch": 1.426003322284336,
"grad_norm": 0.6180042028427124,
"learning_rate": 6.435030648783153e-05,
"loss": 0.2713,
"step": 11750
},
{
"epoch": 1.4563438185031516,
"grad_norm": 0.510152280330658,
"learning_rate": 6.359167324148813e-05,
"loss": 0.2697,
"step": 12000
},
{
"epoch": 1.4866843147219673,
"grad_norm": 0.5209817886352539,
"learning_rate": 6.283303999514476e-05,
"loss": 0.2724,
"step": 12250
},
{
"epoch": 1.517024810940783,
"grad_norm": 0.5717406868934631,
"learning_rate": 6.207440674880136e-05,
"loss": 0.2675,
"step": 12500
},
{
"epoch": 1.5473653071595987,
"grad_norm": 0.5593615770339966,
"learning_rate": 6.131577350245797e-05,
"loss": 0.2683,
"step": 12750
},
{
"epoch": 1.5777058033784144,
"grad_norm": 0.6112098693847656,
"learning_rate": 6.055714025611459e-05,
"loss": 0.2637,
"step": 13000
},
{
"epoch": 1.60804629959723,
"grad_norm": 0.6081228256225586,
"learning_rate": 5.97985070097712e-05,
"loss": 0.2662,
"step": 13250
},
{
"epoch": 1.6383867958160456,
"grad_norm": 0.6320655345916748,
"learning_rate": 5.903987376342781e-05,
"loss": 0.2622,
"step": 13500
},
{
"epoch": 1.6687272920348613,
"grad_norm": 0.5471298098564148,
"learning_rate": 5.8281240517084425e-05,
"loss": 0.2596,
"step": 13750
},
{
"epoch": 1.6990677882536769,
"grad_norm": 0.5194515585899353,
"learning_rate": 5.7522607270741034e-05,
"loss": 0.2627,
"step": 14000
},
{
"epoch": 1.7294082844724925,
"grad_norm": 0.64277184009552,
"learning_rate": 5.676397402439765e-05,
"loss": 0.26,
"step": 14250
},
{
"epoch": 1.7597487806913081,
"grad_norm": 0.5324087738990784,
"learning_rate": 5.6005340778054264e-05,
"loss": 0.2608,
"step": 14500
},
{
"epoch": 1.7900892769101238,
"grad_norm": 0.5574278235435486,
"learning_rate": 5.524670753171087e-05,
"loss": 0.2617,
"step": 14750
},
{
"epoch": 1.8204297731289394,
"grad_norm": 0.5711286664009094,
"learning_rate": 5.448807428536749e-05,
"loss": 0.2571,
"step": 15000
},
{
"epoch": 1.8507702693477552,
"grad_norm": 0.5730472207069397,
"learning_rate": 5.3729441039024095e-05,
"loss": 0.2563,
"step": 15250
},
{
"epoch": 1.8811107655665709,
"grad_norm": 0.6286032199859619,
"learning_rate": 5.297080779268071e-05,
"loss": 0.2546,
"step": 15500
},
{
"epoch": 1.9114512617853865,
"grad_norm": 0.5809808373451233,
"learning_rate": 5.2212174546337325e-05,
"loss": 0.2548,
"step": 15750
},
{
"epoch": 1.9417917580042021,
"grad_norm": 0.5805879831314087,
"learning_rate": 5.1453541299993933e-05,
"loss": 0.2553,
"step": 16000
},
{
"epoch": 1.972132254223018,
"grad_norm": 0.6372638940811157,
"learning_rate": 5.069490805365055e-05,
"loss": 0.2524,
"step": 16250
},
{
"epoch": 2.0024727504418336,
"grad_norm": 0.643139660358429,
"learning_rate": 4.993627480730716e-05,
"loss": 0.251,
"step": 16500
},
{
"epoch": 2.032813246660649,
"grad_norm": 0.5501179099082947,
"learning_rate": 4.917764156096377e-05,
"loss": 0.244,
"step": 16750
},
{
"epoch": 2.063153742879465,
"grad_norm": 0.6013950109481812,
"learning_rate": 4.841900831462039e-05,
"loss": 0.2448,
"step": 17000
},
{
"epoch": 2.0934942390982805,
"grad_norm": 0.4996771216392517,
"learning_rate": 4.7660375068276995e-05,
"loss": 0.2442,
"step": 17250
},
{
"epoch": 2.123834735317096,
"grad_norm": 0.6059885025024414,
"learning_rate": 4.690174182193361e-05,
"loss": 0.242,
"step": 17500
},
{
"epoch": 2.1541752315359117,
"grad_norm": 0.48200371861457825,
"learning_rate": 4.614310857559022e-05,
"loss": 0.2418,
"step": 17750
},
{
"epoch": 2.1845157277547274,
"grad_norm": 0.6055967211723328,
"learning_rate": 4.5384475329246827e-05,
"loss": 0.2428,
"step": 18000
},
{
"epoch": 2.214856223973543,
"grad_norm": 0.5236734747886658,
"learning_rate": 4.462584208290344e-05,
"loss": 0.2409,
"step": 18250
},
{
"epoch": 2.2451967201923586,
"grad_norm": 0.6422255635261536,
"learning_rate": 4.386720883656005e-05,
"loss": 0.2415,
"step": 18500
},
{
"epoch": 2.2755372164111742,
"grad_norm": 0.545559823513031,
"learning_rate": 4.3108575590216665e-05,
"loss": 0.242,
"step": 18750
},
{
"epoch": 2.3058777126299903,
"grad_norm": 0.547564685344696,
"learning_rate": 4.234994234387328e-05,
"loss": 0.2406,
"step": 19000
},
{
"epoch": 2.3362182088488055,
"grad_norm": 0.5706421732902527,
"learning_rate": 4.159130909752989e-05,
"loss": 0.2431,
"step": 19250
},
{
"epoch": 2.3665587050676216,
"grad_norm": 0.6321772933006287,
"learning_rate": 4.08326758511865e-05,
"loss": 0.2378,
"step": 19500
},
{
"epoch": 2.396899201286437,
"grad_norm": 0.6109116077423096,
"learning_rate": 4.007404260484312e-05,
"loss": 0.2374,
"step": 19750
},
{
"epoch": 2.427239697505253,
"grad_norm": 0.4645892083644867,
"learning_rate": 3.9315409358499727e-05,
"loss": 0.2365,
"step": 20000
},
{
"epoch": 2.4575801937240684,
"grad_norm": 0.5845937728881836,
"learning_rate": 3.855677611215634e-05,
"loss": 0.2396,
"step": 20250
},
{
"epoch": 2.487920689942884,
"grad_norm": 0.6609899401664734,
"learning_rate": 3.779814286581295e-05,
"loss": 0.2358,
"step": 20500
},
{
"epoch": 2.5182611861616997,
"grad_norm": 0.6136410236358643,
"learning_rate": 3.7039509619469565e-05,
"loss": 0.2367,
"step": 20750
},
{
"epoch": 2.5486016823805153,
"grad_norm": 0.6023163795471191,
"learning_rate": 3.628087637312618e-05,
"loss": 0.2342,
"step": 21000
},
{
"epoch": 2.578942178599331,
"grad_norm": 0.5570552349090576,
"learning_rate": 3.552224312678279e-05,
"loss": 0.2368,
"step": 21250
},
{
"epoch": 2.6092826748181466,
"grad_norm": 0.5860863327980042,
"learning_rate": 3.47636098804394e-05,
"loss": 0.2345,
"step": 21500
},
{
"epoch": 2.639623171036962,
"grad_norm": 0.6390525698661804,
"learning_rate": 3.400497663409601e-05,
"loss": 0.2365,
"step": 21750
},
{
"epoch": 2.669963667255778,
"grad_norm": 0.6538860201835632,
"learning_rate": 3.3246343387752626e-05,
"loss": 0.2335,
"step": 22000
},
{
"epoch": 2.700304163474594,
"grad_norm": 0.5609804391860962,
"learning_rate": 3.248771014140924e-05,
"loss": 0.2366,
"step": 22250
},
{
"epoch": 2.730644659693409,
"grad_norm": 0.5518357753753662,
"learning_rate": 3.172907689506585e-05,
"loss": 0.2335,
"step": 22500
},
{
"epoch": 2.760985155912225,
"grad_norm": 0.6421113014221191,
"learning_rate": 3.0970443648722465e-05,
"loss": 0.2348,
"step": 22750
},
{
"epoch": 2.7913256521310403,
"grad_norm": 0.6312738656997681,
"learning_rate": 3.0211810402379076e-05,
"loss": 0.2324,
"step": 23000
},
{
"epoch": 2.8216661483498564,
"grad_norm": 0.6342710256576538,
"learning_rate": 2.9453177156035688e-05,
"loss": 0.2291,
"step": 23250
},
{
"epoch": 2.852006644568672,
"grad_norm": 0.5744002461433411,
"learning_rate": 2.86945439096923e-05,
"loss": 0.2318,
"step": 23500
},
{
"epoch": 2.8823471407874877,
"grad_norm": 0.6222126483917236,
"learning_rate": 2.7935910663348915e-05,
"loss": 0.2319,
"step": 23750
},
{
"epoch": 2.9126876370063033,
"grad_norm": 0.5072076916694641,
"learning_rate": 2.7177277417005526e-05,
"loss": 0.2288,
"step": 24000
},
{
"epoch": 2.943028133225119,
"grad_norm": 0.6769903898239136,
"learning_rate": 2.6418644170662138e-05,
"loss": 0.2296,
"step": 24250
},
{
"epoch": 2.9733686294439345,
"grad_norm": 0.6178023815155029,
"learning_rate": 2.566001092431875e-05,
"loss": 0.228,
"step": 24500
},
{
"epoch": 3.00370912566275,
"grad_norm": 0.623375654220581,
"learning_rate": 2.490137767797536e-05,
"loss": 0.2271,
"step": 24750
},
{
"epoch": 3.034049621881566,
"grad_norm": 0.5049629211425781,
"learning_rate": 2.4142744431631973e-05,
"loss": 0.2177,
"step": 25000
},
{
"epoch": 3.0643901181003814,
"grad_norm": 0.5621640682220459,
"learning_rate": 2.3384111185288585e-05,
"loss": 0.2203,
"step": 25250
},
{
"epoch": 3.094730614319197,
"grad_norm": 0.6677132844924927,
"learning_rate": 2.2625477938945196e-05,
"loss": 0.2188,
"step": 25500
},
{
"epoch": 3.1250711105380127,
"grad_norm": 0.6030067801475525,
"learning_rate": 2.1866844692601808e-05,
"loss": 0.2197,
"step": 25750
},
{
"epoch": 3.1554116067568283,
"grad_norm": 0.6289698481559753,
"learning_rate": 2.1108211446258423e-05,
"loss": 0.2206,
"step": 26000
},
{
"epoch": 3.1857521029756444,
"grad_norm": 0.650068461894989,
"learning_rate": 2.0349578199915035e-05,
"loss": 0.2193,
"step": 26250
},
{
"epoch": 3.21609259919446,
"grad_norm": 0.6510699987411499,
"learning_rate": 1.9590944953571646e-05,
"loss": 0.218,
"step": 26500
},
{
"epoch": 3.2464330954132756,
"grad_norm": 0.6897627115249634,
"learning_rate": 1.8832311707228258e-05,
"loss": 0.2172,
"step": 26750
},
{
"epoch": 3.2767735916320913,
"grad_norm": 0.6440379023551941,
"learning_rate": 1.8073678460884873e-05,
"loss": 0.2157,
"step": 27000
},
{
"epoch": 3.307114087850907,
"grad_norm": 0.6011075973510742,
"learning_rate": 1.7315045214541485e-05,
"loss": 0.2159,
"step": 27250
},
{
"epoch": 3.3374545840697225,
"grad_norm": 0.6770527362823486,
"learning_rate": 1.6556411968198096e-05,
"loss": 0.2178,
"step": 27500
},
{
"epoch": 3.367795080288538,
"grad_norm": 0.5674268007278442,
"learning_rate": 1.5797778721854708e-05,
"loss": 0.2155,
"step": 27750
},
{
"epoch": 3.3981355765073538,
"grad_norm": 0.6782290935516357,
"learning_rate": 1.5039145475511321e-05,
"loss": 0.2165,
"step": 28000
},
{
"epoch": 3.4284760727261694,
"grad_norm": 0.6378525495529175,
"learning_rate": 1.4280512229167931e-05,
"loss": 0.2165,
"step": 28250
},
{
"epoch": 3.458816568944985,
"grad_norm": 0.6417750716209412,
"learning_rate": 1.3521878982824543e-05,
"loss": 0.2154,
"step": 28500
},
{
"epoch": 3.4891570651638006,
"grad_norm": 0.733974039554596,
"learning_rate": 1.2763245736481154e-05,
"loss": 0.2145,
"step": 28750
},
{
"epoch": 3.5194975613826163,
"grad_norm": 0.6935612559318542,
"learning_rate": 1.200461249013777e-05,
"loss": 0.2126,
"step": 29000
},
{
"epoch": 3.5498380576014323,
"grad_norm": 0.6449461579322815,
"learning_rate": 1.124597924379438e-05,
"loss": 0.2133,
"step": 29250
},
{
"epoch": 3.5801785538202475,
"grad_norm": 0.630962610244751,
"learning_rate": 1.0487345997450993e-05,
"loss": 0.2137,
"step": 29500
},
{
"epoch": 3.6105190500390636,
"grad_norm": 0.6320120692253113,
"learning_rate": 9.728712751107604e-06,
"loss": 0.2151,
"step": 29750
},
{
"epoch": 3.6408595462578788,
"grad_norm": 0.6066524386405945,
"learning_rate": 8.970079504764218e-06,
"loss": 0.2132,
"step": 30000
},
{
"epoch": 3.671200042476695,
"grad_norm": 0.707771897315979,
"learning_rate": 8.21144625842083e-06,
"loss": 0.2122,
"step": 30250
},
{
"epoch": 3.7015405386955105,
"grad_norm": 0.6526840925216675,
"learning_rate": 7.452813012077442e-06,
"loss": 0.2107,
"step": 30500
},
{
"epoch": 3.731881034914326,
"grad_norm": 0.7959334254264832,
"learning_rate": 6.694179765734054e-06,
"loss": 0.2119,
"step": 30750
},
{
"epoch": 3.7622215311331417,
"grad_norm": 0.6414825320243835,
"learning_rate": 5.935546519390666e-06,
"loss": 0.2092,
"step": 31000
},
{
"epoch": 3.7925620273519574,
"grad_norm": 0.6206453442573547,
"learning_rate": 5.1769132730472785e-06,
"loss": 0.2105,
"step": 31250
},
{
"epoch": 3.822902523570773,
"grad_norm": 0.6464530825614929,
"learning_rate": 4.418280026703891e-06,
"loss": 0.2109,
"step": 31500
},
{
"epoch": 3.8532430197895886,
"grad_norm": 0.6245775818824768,
"learning_rate": 3.6596467803605027e-06,
"loss": 0.2108,
"step": 31750
},
{
"epoch": 3.8835835160084042,
"grad_norm": 0.6799646615982056,
"learning_rate": 2.9010135340171147e-06,
"loss": 0.2116,
"step": 32000
},
{
"epoch": 3.91392401222722,
"grad_norm": 0.6648467183113098,
"learning_rate": 2.1423802876737272e-06,
"loss": 0.2108,
"step": 32250
},
{
"epoch": 3.9442645084460355,
"grad_norm": 0.6555809378623962,
"learning_rate": 1.3837470413303393e-06,
"loss": 0.2117,
"step": 32500
},
{
"epoch": 3.974605004664851,
"grad_norm": 0.5603693127632141,
"learning_rate": 6.251137949869516e-07,
"loss": 0.209,
"step": 32750
},
{
"epoch": 3.9996055735491556,
"step": 32956,
"total_flos": 5.234804195348718e+18,
"train_loss": 0.2934305334803057,
"train_runtime": 64350.2603,
"train_samples_per_second": 32.78,
"train_steps_per_second": 0.512
}
],
"logging_steps": 250,
"max_steps": 32956,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 5.234804195348718e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}