cbb-1b / checkpoint-1098 /trainer_state.json
awilliamson's picture
Upload folder using huggingface_hub
9f0cda1 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9979522184300342,
"eval_steps": 500,
"global_step": 1098,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0027303754266211604,
"grad_norm": 1.3254656791687012,
"learning_rate": 1.818181818181818e-06,
"loss": 1.2897,
"step": 1
},
{
"epoch": 0.005460750853242321,
"grad_norm": 1.3267881870269775,
"learning_rate": 3.636363636363636e-06,
"loss": 1.2823,
"step": 2
},
{
"epoch": 0.008191126279863481,
"grad_norm": 1.2982481718063354,
"learning_rate": 5.4545454545454545e-06,
"loss": 1.2595,
"step": 3
},
{
"epoch": 0.010921501706484642,
"grad_norm": 1.2894413471221924,
"learning_rate": 7.272727272727272e-06,
"loss": 1.2653,
"step": 4
},
{
"epoch": 0.013651877133105802,
"grad_norm": 1.2869772911071777,
"learning_rate": 9.090909090909091e-06,
"loss": 1.2545,
"step": 5
},
{
"epoch": 0.016382252559726963,
"grad_norm": 1.2373387813568115,
"learning_rate": 1.0909090909090909e-05,
"loss": 1.2194,
"step": 6
},
{
"epoch": 0.01911262798634812,
"grad_norm": 1.20195734500885,
"learning_rate": 1.2727272727272727e-05,
"loss": 1.2139,
"step": 7
},
{
"epoch": 0.021843003412969283,
"grad_norm": 1.1426103115081787,
"learning_rate": 1.4545454545454545e-05,
"loss": 1.2315,
"step": 8
},
{
"epoch": 0.024573378839590442,
"grad_norm": 1.0495123863220215,
"learning_rate": 1.6363636363636366e-05,
"loss": 1.1944,
"step": 9
},
{
"epoch": 0.027303754266211604,
"grad_norm": 0.8776500821113586,
"learning_rate": 1.8181818181818182e-05,
"loss": 1.183,
"step": 10
},
{
"epoch": 0.030034129692832763,
"grad_norm": 0.8687052130699158,
"learning_rate": 2e-05,
"loss": 1.1592,
"step": 11
},
{
"epoch": 0.032764505119453925,
"grad_norm": 0.7476271390914917,
"learning_rate": 2.1818181818181818e-05,
"loss": 1.1483,
"step": 12
},
{
"epoch": 0.03549488054607509,
"grad_norm": 0.6418495774269104,
"learning_rate": 2.3636363636363637e-05,
"loss": 1.0762,
"step": 13
},
{
"epoch": 0.03822525597269624,
"grad_norm": 0.600390613079071,
"learning_rate": 2.5454545454545454e-05,
"loss": 1.0492,
"step": 14
},
{
"epoch": 0.040955631399317405,
"grad_norm": 0.5653348565101624,
"learning_rate": 2.7272727272727273e-05,
"loss": 1.0548,
"step": 15
},
{
"epoch": 0.04368600682593857,
"grad_norm": 0.5357097387313843,
"learning_rate": 2.909090909090909e-05,
"loss": 1.0273,
"step": 16
},
{
"epoch": 0.04641638225255973,
"grad_norm": 0.4480445683002472,
"learning_rate": 3.090909090909091e-05,
"loss": 1.0065,
"step": 17
},
{
"epoch": 0.049146757679180884,
"grad_norm": 0.40983352065086365,
"learning_rate": 3.272727272727273e-05,
"loss": 0.9908,
"step": 18
},
{
"epoch": 0.05187713310580205,
"grad_norm": 0.42159780859947205,
"learning_rate": 3.454545454545455e-05,
"loss": 0.9888,
"step": 19
},
{
"epoch": 0.05460750853242321,
"grad_norm": 0.41620174050331116,
"learning_rate": 3.6363636363636364e-05,
"loss": 0.9575,
"step": 20
},
{
"epoch": 0.05733788395904437,
"grad_norm": 0.3804452419281006,
"learning_rate": 3.818181818181819e-05,
"loss": 0.9413,
"step": 21
},
{
"epoch": 0.060068259385665526,
"grad_norm": 0.37021100521087646,
"learning_rate": 4e-05,
"loss": 0.9223,
"step": 22
},
{
"epoch": 0.06279863481228669,
"grad_norm": 0.34090206027030945,
"learning_rate": 4.181818181818182e-05,
"loss": 0.8878,
"step": 23
},
{
"epoch": 0.06552901023890785,
"grad_norm": 0.32232972979545593,
"learning_rate": 4.3636363636363636e-05,
"loss": 0.8986,
"step": 24
},
{
"epoch": 0.06825938566552901,
"grad_norm": 0.2941684424877167,
"learning_rate": 4.545454545454546e-05,
"loss": 0.8857,
"step": 25
},
{
"epoch": 0.07098976109215017,
"grad_norm": 0.27072674036026,
"learning_rate": 4.7272727272727275e-05,
"loss": 0.8736,
"step": 26
},
{
"epoch": 0.07372013651877134,
"grad_norm": 0.2696637511253357,
"learning_rate": 4.909090909090909e-05,
"loss": 0.8698,
"step": 27
},
{
"epoch": 0.07645051194539249,
"grad_norm": 0.2565267086029053,
"learning_rate": 5.090909090909091e-05,
"loss": 0.8324,
"step": 28
},
{
"epoch": 0.07918088737201365,
"grad_norm": 0.2474038451910019,
"learning_rate": 5.272727272727272e-05,
"loss": 0.841,
"step": 29
},
{
"epoch": 0.08191126279863481,
"grad_norm": 0.22430865466594696,
"learning_rate": 5.4545454545454546e-05,
"loss": 0.8219,
"step": 30
},
{
"epoch": 0.08464163822525597,
"grad_norm": 0.21238166093826294,
"learning_rate": 5.636363636363636e-05,
"loss": 0.8328,
"step": 31
},
{
"epoch": 0.08737201365187713,
"grad_norm": 0.2210783213376999,
"learning_rate": 5.818181818181818e-05,
"loss": 0.8187,
"step": 32
},
{
"epoch": 0.0901023890784983,
"grad_norm": 0.24119816720485687,
"learning_rate": 6e-05,
"loss": 0.8078,
"step": 33
},
{
"epoch": 0.09283276450511946,
"grad_norm": 0.21313577890396118,
"learning_rate": 6.181818181818182e-05,
"loss": 0.8051,
"step": 34
},
{
"epoch": 0.09556313993174062,
"grad_norm": 0.17824789881706238,
"learning_rate": 6.363636363636364e-05,
"loss": 0.7841,
"step": 35
},
{
"epoch": 0.09829351535836177,
"grad_norm": 0.18413369357585907,
"learning_rate": 6.545454545454546e-05,
"loss": 0.7851,
"step": 36
},
{
"epoch": 0.10102389078498293,
"grad_norm": 0.2003067582845688,
"learning_rate": 6.727272727272727e-05,
"loss": 0.8064,
"step": 37
},
{
"epoch": 0.1037542662116041,
"grad_norm": 0.1989540457725525,
"learning_rate": 6.90909090909091e-05,
"loss": 0.7841,
"step": 38
},
{
"epoch": 0.10648464163822526,
"grad_norm": 0.14544272422790527,
"learning_rate": 7.090909090909092e-05,
"loss": 0.7745,
"step": 39
},
{
"epoch": 0.10921501706484642,
"grad_norm": 0.1559988260269165,
"learning_rate": 7.272727272727273e-05,
"loss": 0.7845,
"step": 40
},
{
"epoch": 0.11194539249146758,
"grad_norm": 0.1705523580312729,
"learning_rate": 7.454545454545455e-05,
"loss": 0.7876,
"step": 41
},
{
"epoch": 0.11467576791808874,
"grad_norm": 0.1464846283197403,
"learning_rate": 7.636363636363637e-05,
"loss": 0.7846,
"step": 42
},
{
"epoch": 0.1174061433447099,
"grad_norm": 0.1304199993610382,
"learning_rate": 7.818181818181818e-05,
"loss": 0.7734,
"step": 43
},
{
"epoch": 0.12013651877133105,
"grad_norm": 0.1516261249780655,
"learning_rate": 8e-05,
"loss": 0.7836,
"step": 44
},
{
"epoch": 0.12286689419795221,
"grad_norm": 0.1361905336380005,
"learning_rate": 8.181818181818183e-05,
"loss": 0.7751,
"step": 45
},
{
"epoch": 0.12559726962457338,
"grad_norm": 0.14435631036758423,
"learning_rate": 8.363636363636364e-05,
"loss": 0.7655,
"step": 46
},
{
"epoch": 0.12832764505119454,
"grad_norm": 0.13407501578330994,
"learning_rate": 8.545454545454545e-05,
"loss": 0.7729,
"step": 47
},
{
"epoch": 0.1310580204778157,
"grad_norm": 0.13555769622325897,
"learning_rate": 8.727272727272727e-05,
"loss": 0.7693,
"step": 48
},
{
"epoch": 0.13378839590443686,
"grad_norm": 0.1439952850341797,
"learning_rate": 8.90909090909091e-05,
"loss": 0.7787,
"step": 49
},
{
"epoch": 0.13651877133105803,
"grad_norm": 0.14037510752677917,
"learning_rate": 9.090909090909092e-05,
"loss": 0.7757,
"step": 50
},
{
"epoch": 0.1392491467576792,
"grad_norm": 0.11772281676530838,
"learning_rate": 9.272727272727273e-05,
"loss": 0.7408,
"step": 51
},
{
"epoch": 0.14197952218430035,
"grad_norm": 0.1545950025320053,
"learning_rate": 9.454545454545455e-05,
"loss": 0.7381,
"step": 52
},
{
"epoch": 0.1447098976109215,
"grad_norm": 0.12565699219703674,
"learning_rate": 9.636363636363637e-05,
"loss": 0.7769,
"step": 53
},
{
"epoch": 0.14744027303754267,
"grad_norm": 0.15412947535514832,
"learning_rate": 9.818181818181818e-05,
"loss": 0.7585,
"step": 54
},
{
"epoch": 0.15017064846416384,
"grad_norm": 0.11638892441987991,
"learning_rate": 0.0001,
"loss": 0.7661,
"step": 55
},
{
"epoch": 0.15290102389078497,
"grad_norm": 0.16432470083236694,
"learning_rate": 0.00010181818181818181,
"loss": 0.7546,
"step": 56
},
{
"epoch": 0.15563139931740613,
"grad_norm": 0.11529026180505753,
"learning_rate": 0.00010363636363636364,
"loss": 0.7535,
"step": 57
},
{
"epoch": 0.1583617747440273,
"grad_norm": 0.23582805693149567,
"learning_rate": 0.00010545454545454545,
"loss": 0.7683,
"step": 58
},
{
"epoch": 0.16109215017064846,
"grad_norm": 0.12393908202648163,
"learning_rate": 0.00010727272727272728,
"loss": 0.7526,
"step": 59
},
{
"epoch": 0.16382252559726962,
"grad_norm": 0.20981422066688538,
"learning_rate": 0.00010909090909090909,
"loss": 0.7397,
"step": 60
},
{
"epoch": 0.16655290102389078,
"grad_norm": 0.1519405096769333,
"learning_rate": 0.00011090909090909092,
"loss": 0.7472,
"step": 61
},
{
"epoch": 0.16928327645051194,
"grad_norm": 0.15360122919082642,
"learning_rate": 0.00011272727272727272,
"loss": 0.7452,
"step": 62
},
{
"epoch": 0.1720136518771331,
"grad_norm": 0.14730164408683777,
"learning_rate": 0.00011454545454545456,
"loss": 0.7068,
"step": 63
},
{
"epoch": 0.17474402730375427,
"grad_norm": 0.166826993227005,
"learning_rate": 0.00011636363636363636,
"loss": 0.7476,
"step": 64
},
{
"epoch": 0.17747440273037543,
"grad_norm": 0.17365337908267975,
"learning_rate": 0.0001181818181818182,
"loss": 0.7083,
"step": 65
},
{
"epoch": 0.1802047781569966,
"grad_norm": 0.2101927250623703,
"learning_rate": 0.00012,
"loss": 0.755,
"step": 66
},
{
"epoch": 0.18293515358361775,
"grad_norm": 0.1802549660205841,
"learning_rate": 0.00012181818181818183,
"loss": 0.7388,
"step": 67
},
{
"epoch": 0.18566552901023892,
"grad_norm": 0.17372193932533264,
"learning_rate": 0.00012363636363636364,
"loss": 0.7067,
"step": 68
},
{
"epoch": 0.18839590443686008,
"grad_norm": 0.15312258899211884,
"learning_rate": 0.00012545454545454546,
"loss": 0.7424,
"step": 69
},
{
"epoch": 0.19112627986348124,
"grad_norm": 0.24253840744495392,
"learning_rate": 0.00012727272727272728,
"loss": 0.7418,
"step": 70
},
{
"epoch": 0.19385665529010238,
"grad_norm": 0.2043231725692749,
"learning_rate": 0.0001290909090909091,
"loss": 0.7362,
"step": 71
},
{
"epoch": 0.19658703071672354,
"grad_norm": 0.19859246909618378,
"learning_rate": 0.00013090909090909093,
"loss": 0.7071,
"step": 72
},
{
"epoch": 0.1993174061433447,
"grad_norm": 0.20175667107105255,
"learning_rate": 0.00013272727272727275,
"loss": 0.7202,
"step": 73
},
{
"epoch": 0.20204778156996586,
"grad_norm": 0.1653033196926117,
"learning_rate": 0.00013454545454545455,
"loss": 0.7229,
"step": 74
},
{
"epoch": 0.20477815699658702,
"grad_norm": 0.17003491520881653,
"learning_rate": 0.00013636363636363637,
"loss": 0.7353,
"step": 75
},
{
"epoch": 0.2075085324232082,
"grad_norm": 0.18356764316558838,
"learning_rate": 0.0001381818181818182,
"loss": 0.7138,
"step": 76
},
{
"epoch": 0.21023890784982935,
"grad_norm": 0.2215511053800583,
"learning_rate": 0.00014,
"loss": 0.7238,
"step": 77
},
{
"epoch": 0.2129692832764505,
"grad_norm": 0.34184327721595764,
"learning_rate": 0.00014181818181818184,
"loss": 0.7297,
"step": 78
},
{
"epoch": 0.21569965870307167,
"grad_norm": 0.25543472170829773,
"learning_rate": 0.00014363636363636363,
"loss": 0.742,
"step": 79
},
{
"epoch": 0.21843003412969283,
"grad_norm": 0.2220849245786667,
"learning_rate": 0.00014545454545454546,
"loss": 0.6897,
"step": 80
},
{
"epoch": 0.221160409556314,
"grad_norm": 0.25552013516426086,
"learning_rate": 0.00014727272727272728,
"loss": 0.744,
"step": 81
},
{
"epoch": 0.22389078498293516,
"grad_norm": 0.2621108889579773,
"learning_rate": 0.0001490909090909091,
"loss": 0.7193,
"step": 82
},
{
"epoch": 0.22662116040955632,
"grad_norm": 0.1840047836303711,
"learning_rate": 0.0001509090909090909,
"loss": 0.6991,
"step": 83
},
{
"epoch": 0.22935153583617748,
"grad_norm": 0.21538959443569183,
"learning_rate": 0.00015272727272727275,
"loss": 0.7058,
"step": 84
},
{
"epoch": 0.23208191126279865,
"grad_norm": 0.34613293409347534,
"learning_rate": 0.00015454545454545454,
"loss": 0.7011,
"step": 85
},
{
"epoch": 0.2348122866894198,
"grad_norm": 0.26010966300964355,
"learning_rate": 0.00015636363636363637,
"loss": 0.7009,
"step": 86
},
{
"epoch": 0.23754266211604094,
"grad_norm": 0.18031255900859833,
"learning_rate": 0.0001581818181818182,
"loss": 0.7036,
"step": 87
},
{
"epoch": 0.2402730375426621,
"grad_norm": 0.24567286670207977,
"learning_rate": 0.00016,
"loss": 0.6921,
"step": 88
},
{
"epoch": 0.24300341296928327,
"grad_norm": 0.19522973895072937,
"learning_rate": 0.00016181818181818184,
"loss": 0.7109,
"step": 89
},
{
"epoch": 0.24573378839590443,
"grad_norm": 0.2405068725347519,
"learning_rate": 0.00016363636363636366,
"loss": 0.7134,
"step": 90
},
{
"epoch": 0.2484641638225256,
"grad_norm": 0.15669392049312592,
"learning_rate": 0.00016545454545454545,
"loss": 0.6966,
"step": 91
},
{
"epoch": 0.25119453924914675,
"grad_norm": 0.23415732383728027,
"learning_rate": 0.00016727272727272728,
"loss": 0.6771,
"step": 92
},
{
"epoch": 0.25392491467576794,
"grad_norm": 0.1842266321182251,
"learning_rate": 0.0001690909090909091,
"loss": 0.6895,
"step": 93
},
{
"epoch": 0.2566552901023891,
"grad_norm": 0.21642841398715973,
"learning_rate": 0.0001709090909090909,
"loss": 0.6913,
"step": 94
},
{
"epoch": 0.2593856655290102,
"grad_norm": 0.26327016949653625,
"learning_rate": 0.00017272727272727275,
"loss": 0.6856,
"step": 95
},
{
"epoch": 0.2621160409556314,
"grad_norm": 0.20735357701778412,
"learning_rate": 0.00017454545454545454,
"loss": 0.6769,
"step": 96
},
{
"epoch": 0.26484641638225254,
"grad_norm": 0.3127861022949219,
"learning_rate": 0.00017636363636363637,
"loss": 0.6518,
"step": 97
},
{
"epoch": 0.2675767918088737,
"grad_norm": 0.34650346636772156,
"learning_rate": 0.0001781818181818182,
"loss": 0.6937,
"step": 98
},
{
"epoch": 0.27030716723549486,
"grad_norm": 0.252059668302536,
"learning_rate": 0.00018,
"loss": 0.6589,
"step": 99
},
{
"epoch": 0.27303754266211605,
"grad_norm": 0.1896669715642929,
"learning_rate": 0.00018181818181818183,
"loss": 0.6717,
"step": 100
},
{
"epoch": 0.2757679180887372,
"grad_norm": 0.2724236249923706,
"learning_rate": 0.00018363636363636366,
"loss": 0.6671,
"step": 101
},
{
"epoch": 0.2784982935153584,
"grad_norm": 0.1814826875925064,
"learning_rate": 0.00018545454545454545,
"loss": 0.6564,
"step": 102
},
{
"epoch": 0.2812286689419795,
"grad_norm": 0.24140000343322754,
"learning_rate": 0.00018727272727272728,
"loss": 0.6708,
"step": 103
},
{
"epoch": 0.2839590443686007,
"grad_norm": 0.19333204627037048,
"learning_rate": 0.0001890909090909091,
"loss": 0.6628,
"step": 104
},
{
"epoch": 0.28668941979522183,
"grad_norm": 0.21080803871154785,
"learning_rate": 0.00019090909090909092,
"loss": 0.6457,
"step": 105
},
{
"epoch": 0.289419795221843,
"grad_norm": 0.20848962664604187,
"learning_rate": 0.00019272727272727274,
"loss": 0.6589,
"step": 106
},
{
"epoch": 0.29215017064846416,
"grad_norm": 0.2381501942873001,
"learning_rate": 0.00019454545454545457,
"loss": 0.6846,
"step": 107
},
{
"epoch": 0.29488054607508535,
"grad_norm": 0.1734190434217453,
"learning_rate": 0.00019636363636363636,
"loss": 0.6691,
"step": 108
},
{
"epoch": 0.2976109215017065,
"grad_norm": 0.2187424749135971,
"learning_rate": 0.00019818181818181821,
"loss": 0.6806,
"step": 109
},
{
"epoch": 0.3003412969283277,
"grad_norm": 0.21607345342636108,
"learning_rate": 0.0002,
"loss": 0.6588,
"step": 110
},
{
"epoch": 0.3030716723549488,
"grad_norm": 0.1873304545879364,
"learning_rate": 0.00019999949446003433,
"loss": 0.6513,
"step": 111
},
{
"epoch": 0.30580204778156994,
"grad_norm": 0.21443282067775726,
"learning_rate": 0.00019999797784524866,
"loss": 0.6704,
"step": 112
},
{
"epoch": 0.30853242320819113,
"grad_norm": 0.18743731081485748,
"learning_rate": 0.00019999545017097728,
"loss": 0.6346,
"step": 113
},
{
"epoch": 0.31126279863481227,
"grad_norm": 0.18916335701942444,
"learning_rate": 0.0001999919114627769,
"loss": 0.68,
"step": 114
},
{
"epoch": 0.31399317406143346,
"grad_norm": 0.1925644427537918,
"learning_rate": 0.00019998736175642673,
"loss": 0.6408,
"step": 115
},
{
"epoch": 0.3167235494880546,
"grad_norm": 0.17722898721694946,
"learning_rate": 0.0001999818010979279,
"loss": 0.6787,
"step": 116
},
{
"epoch": 0.3194539249146758,
"grad_norm": 0.19374825060367584,
"learning_rate": 0.0001999752295435032,
"loss": 0.669,
"step": 117
},
{
"epoch": 0.3221843003412969,
"grad_norm": 0.20013949275016785,
"learning_rate": 0.00019996764715959618,
"loss": 0.653,
"step": 118
},
{
"epoch": 0.3249146757679181,
"grad_norm": 0.18780681490898132,
"learning_rate": 0.00019995905402287094,
"loss": 0.6557,
"step": 119
},
{
"epoch": 0.32764505119453924,
"grad_norm": 0.1718084216117859,
"learning_rate": 0.00019994945022021082,
"loss": 0.6272,
"step": 120
},
{
"epoch": 0.33037542662116043,
"grad_norm": 0.1613592952489853,
"learning_rate": 0.00019993883584871808,
"loss": 0.6515,
"step": 121
},
{
"epoch": 0.33310580204778156,
"grad_norm": 0.1737043410539627,
"learning_rate": 0.00019992721101571236,
"loss": 0.6134,
"step": 122
},
{
"epoch": 0.33583617747440275,
"grad_norm": 0.16362418234348297,
"learning_rate": 0.0001999145758387301,
"loss": 0.6448,
"step": 123
},
{
"epoch": 0.3385665529010239,
"grad_norm": 0.19181552529335022,
"learning_rate": 0.00019990093044552304,
"loss": 0.6497,
"step": 124
},
{
"epoch": 0.3412969283276451,
"grad_norm": 0.15803317725658417,
"learning_rate": 0.00019988627497405696,
"loss": 0.6116,
"step": 125
},
{
"epoch": 0.3440273037542662,
"grad_norm": 0.2208717167377472,
"learning_rate": 0.00019987060957251047,
"loss": 0.6459,
"step": 126
},
{
"epoch": 0.34675767918088735,
"grad_norm": 0.20142869651317596,
"learning_rate": 0.00019985393439927323,
"loss": 0.6589,
"step": 127
},
{
"epoch": 0.34948805460750854,
"grad_norm": 0.17945925891399384,
"learning_rate": 0.00019983624962294458,
"loss": 0.6252,
"step": 128
},
{
"epoch": 0.35221843003412967,
"grad_norm": 0.22226247191429138,
"learning_rate": 0.00019981755542233177,
"loss": 0.6379,
"step": 129
},
{
"epoch": 0.35494880546075086,
"grad_norm": 0.1588139832019806,
"learning_rate": 0.00019979785198644806,
"loss": 0.6408,
"step": 130
},
{
"epoch": 0.357679180887372,
"grad_norm": 0.22639498114585876,
"learning_rate": 0.00019977713951451102,
"loss": 0.6434,
"step": 131
},
{
"epoch": 0.3604095563139932,
"grad_norm": 0.16015386581420898,
"learning_rate": 0.00019975541821594026,
"loss": 0.6151,
"step": 132
},
{
"epoch": 0.3631399317406143,
"grad_norm": 0.21671050786972046,
"learning_rate": 0.00019973268831035545,
"loss": 0.6357,
"step": 133
},
{
"epoch": 0.3658703071672355,
"grad_norm": 0.1871589720249176,
"learning_rate": 0.00019970895002757413,
"loss": 0.6436,
"step": 134
},
{
"epoch": 0.36860068259385664,
"grad_norm": 0.19527480006217957,
"learning_rate": 0.00019968420360760926,
"loss": 0.6308,
"step": 135
},
{
"epoch": 0.37133105802047783,
"grad_norm": 0.20158074796199799,
"learning_rate": 0.000199658449300667,
"loss": 0.6227,
"step": 136
},
{
"epoch": 0.37406143344709897,
"grad_norm": 0.15605966746807098,
"learning_rate": 0.00019963168736714392,
"loss": 0.615,
"step": 137
},
{
"epoch": 0.37679180887372016,
"grad_norm": 0.22042252123355865,
"learning_rate": 0.00019960391807762463,
"loss": 0.6263,
"step": 138
},
{
"epoch": 0.3795221843003413,
"grad_norm": 0.16206978261470795,
"learning_rate": 0.00019957514171287875,
"loss": 0.6182,
"step": 139
},
{
"epoch": 0.3822525597269625,
"grad_norm": 0.2251751869916916,
"learning_rate": 0.00019954535856385837,
"loss": 0.6376,
"step": 140
},
{
"epoch": 0.3849829351535836,
"grad_norm": 0.16586551070213318,
"learning_rate": 0.00019951456893169497,
"loss": 0.6285,
"step": 141
},
{
"epoch": 0.38771331058020475,
"grad_norm": 0.27427414059638977,
"learning_rate": 0.0001994827731276963,
"loss": 0.6397,
"step": 142
},
{
"epoch": 0.39044368600682594,
"grad_norm": 0.21177491545677185,
"learning_rate": 0.00019944997147334337,
"loss": 0.6034,
"step": 143
},
{
"epoch": 0.3931740614334471,
"grad_norm": 0.25477880239486694,
"learning_rate": 0.0001994161643002871,
"loss": 0.6199,
"step": 144
},
{
"epoch": 0.39590443686006827,
"grad_norm": 0.23290970921516418,
"learning_rate": 0.00019938135195034508,
"loss": 0.6201,
"step": 145
},
{
"epoch": 0.3986348122866894,
"grad_norm": 0.19198672473430634,
"learning_rate": 0.00019934553477549794,
"loss": 0.6213,
"step": 146
},
{
"epoch": 0.4013651877133106,
"grad_norm": 0.1911400705575943,
"learning_rate": 0.000199308713137886,
"loss": 0.6146,
"step": 147
},
{
"epoch": 0.4040955631399317,
"grad_norm": 0.18605491518974304,
"learning_rate": 0.0001992708874098054,
"loss": 0.6123,
"step": 148
},
{
"epoch": 0.4068259385665529,
"grad_norm": 0.18028293550014496,
"learning_rate": 0.0001992320579737045,
"loss": 0.6061,
"step": 149
},
{
"epoch": 0.40955631399317405,
"grad_norm": 0.1961037963628769,
"learning_rate": 0.00019919222522217996,
"loss": 0.622,
"step": 150
},
{
"epoch": 0.41228668941979524,
"grad_norm": 0.17400594055652618,
"learning_rate": 0.00019915138955797272,
"loss": 0.6138,
"step": 151
},
{
"epoch": 0.4150170648464164,
"grad_norm": 0.17892149090766907,
"learning_rate": 0.00019910955139396396,
"loss": 0.6242,
"step": 152
},
{
"epoch": 0.41774744027303756,
"grad_norm": 0.21851663291454315,
"learning_rate": 0.000199066711153171,
"loss": 0.5913,
"step": 153
},
{
"epoch": 0.4204778156996587,
"grad_norm": 0.1468774825334549,
"learning_rate": 0.0001990228692687429,
"loss": 0.6025,
"step": 154
},
{
"epoch": 0.4232081911262799,
"grad_norm": 0.1920468658208847,
"learning_rate": 0.00019897802618395614,
"loss": 0.6127,
"step": 155
},
{
"epoch": 0.425938566552901,
"grad_norm": 0.17375442385673523,
"learning_rate": 0.00019893218235221015,
"loss": 0.6211,
"step": 156
},
{
"epoch": 0.4286689419795222,
"grad_norm": 0.15414904057979584,
"learning_rate": 0.00019888533823702277,
"loss": 0.6183,
"step": 157
},
{
"epoch": 0.43139931740614335,
"grad_norm": 0.2245103418827057,
"learning_rate": 0.0001988374943120254,
"loss": 0.6248,
"step": 158
},
{
"epoch": 0.4341296928327645,
"grad_norm": 0.17193332314491272,
"learning_rate": 0.00019878865106095835,
"loss": 0.5969,
"step": 159
},
{
"epoch": 0.43686006825938567,
"grad_norm": 0.13767646253108978,
"learning_rate": 0.00019873880897766598,
"loss": 0.5943,
"step": 160
},
{
"epoch": 0.4395904436860068,
"grad_norm": 0.1449906826019287,
"learning_rate": 0.00019868796856609152,
"loss": 0.573,
"step": 161
},
{
"epoch": 0.442320819112628,
"grad_norm": 0.1392473578453064,
"learning_rate": 0.00019863613034027224,
"loss": 0.5926,
"step": 162
},
{
"epoch": 0.44505119453924913,
"grad_norm": 0.1772463023662567,
"learning_rate": 0.00019858329482433403,
"loss": 0.6007,
"step": 163
},
{
"epoch": 0.4477815699658703,
"grad_norm": 0.13768768310546875,
"learning_rate": 0.0001985294625524861,
"loss": 0.5901,
"step": 164
},
{
"epoch": 0.45051194539249145,
"grad_norm": 0.1631435751914978,
"learning_rate": 0.00019847463406901588,
"loss": 0.5907,
"step": 165
},
{
"epoch": 0.45324232081911264,
"grad_norm": 0.14781758189201355,
"learning_rate": 0.00019841880992828306,
"loss": 0.5903,
"step": 166
},
{
"epoch": 0.4559726962457338,
"grad_norm": 0.13440802693367004,
"learning_rate": 0.00019836199069471437,
"loss": 0.5884,
"step": 167
},
{
"epoch": 0.45870307167235497,
"grad_norm": 0.1414463371038437,
"learning_rate": 0.00019830417694279766,
"loss": 0.598,
"step": 168
},
{
"epoch": 0.4614334470989761,
"grad_norm": 0.13185666501522064,
"learning_rate": 0.0001982453692570762,
"loss": 0.621,
"step": 169
},
{
"epoch": 0.4641638225255973,
"grad_norm": 0.14422471821308136,
"learning_rate": 0.00019818556823214268,
"loss": 0.6065,
"step": 170
},
{
"epoch": 0.4668941979522184,
"grad_norm": 0.13765788078308105,
"learning_rate": 0.00019812477447263326,
"loss": 0.6073,
"step": 171
},
{
"epoch": 0.4696245733788396,
"grad_norm": 0.16028070449829102,
"learning_rate": 0.0001980629885932214,
"loss": 0.5767,
"step": 172
},
{
"epoch": 0.47235494880546075,
"grad_norm": 0.14638394117355347,
"learning_rate": 0.00019800021121861182,
"loss": 0.5971,
"step": 173
},
{
"epoch": 0.4750853242320819,
"grad_norm": 0.14843404293060303,
"learning_rate": 0.0001979364429835339,
"loss": 0.5894,
"step": 174
},
{
"epoch": 0.4778156996587031,
"grad_norm": 0.13411492109298706,
"learning_rate": 0.00019787168453273544,
"loss": 0.5757,
"step": 175
},
{
"epoch": 0.4805460750853242,
"grad_norm": 0.14304684102535248,
"learning_rate": 0.0001978059365209762,
"loss": 0.5846,
"step": 176
},
{
"epoch": 0.4832764505119454,
"grad_norm": 0.13569754362106323,
"learning_rate": 0.00019773919961302113,
"loss": 0.5872,
"step": 177
},
{
"epoch": 0.48600682593856653,
"grad_norm": 0.14318887889385223,
"learning_rate": 0.00019767147448363366,
"loss": 0.5804,
"step": 178
},
{
"epoch": 0.4887372013651877,
"grad_norm": 0.1457952857017517,
"learning_rate": 0.00019760276181756903,
"loss": 0.5973,
"step": 179
},
{
"epoch": 0.49146757679180886,
"grad_norm": 0.13820476830005646,
"learning_rate": 0.00019753306230956718,
"loss": 0.569,
"step": 180
},
{
"epoch": 0.49419795221843005,
"grad_norm": 0.19338561594486237,
"learning_rate": 0.00019746237666434587,
"loss": 0.5723,
"step": 181
},
{
"epoch": 0.4969283276450512,
"grad_norm": 0.17352697253227234,
"learning_rate": 0.00019739070559659347,
"loss": 0.578,
"step": 182
},
{
"epoch": 0.49965870307167237,
"grad_norm": 0.15502339601516724,
"learning_rate": 0.00019731804983096177,
"loss": 0.5953,
"step": 183
},
{
"epoch": 0.5023890784982935,
"grad_norm": 0.18948784470558167,
"learning_rate": 0.00019724441010205863,
"loss": 0.5883,
"step": 184
},
{
"epoch": 0.5051194539249146,
"grad_norm": 0.17587606608867645,
"learning_rate": 0.00019716978715444056,
"loss": 0.5723,
"step": 185
},
{
"epoch": 0.5078498293515359,
"grad_norm": 0.1599951833486557,
"learning_rate": 0.0001970941817426052,
"loss": 0.5799,
"step": 186
},
{
"epoch": 0.510580204778157,
"grad_norm": 0.1717846840620041,
"learning_rate": 0.00019701759463098374,
"loss": 0.5543,
"step": 187
},
{
"epoch": 0.5133105802047782,
"grad_norm": 0.14032602310180664,
"learning_rate": 0.00019694002659393305,
"loss": 0.5845,
"step": 188
},
{
"epoch": 0.5160409556313993,
"grad_norm": 0.17668449878692627,
"learning_rate": 0.000196861478415728,
"loss": 0.6026,
"step": 189
},
{
"epoch": 0.5187713310580204,
"grad_norm": 0.17806965112686157,
"learning_rate": 0.00019678195089055346,
"loss": 0.5681,
"step": 190
},
{
"epoch": 0.5215017064846417,
"grad_norm": 0.13321803510189056,
"learning_rate": 0.00019670144482249627,
"loss": 0.5586,
"step": 191
},
{
"epoch": 0.5242320819112628,
"grad_norm": 0.14684653282165527,
"learning_rate": 0.00019661996102553718,
"loss": 0.5589,
"step": 192
},
{
"epoch": 0.5269624573378839,
"grad_norm": 0.1308140754699707,
"learning_rate": 0.0001965375003235424,
"loss": 0.568,
"step": 193
},
{
"epoch": 0.5296928327645051,
"grad_norm": 0.17461615800857544,
"learning_rate": 0.00019645406355025565,
"loss": 0.5757,
"step": 194
},
{
"epoch": 0.5324232081911263,
"grad_norm": 0.15591022372245789,
"learning_rate": 0.0001963696515492893,
"loss": 0.5946,
"step": 195
},
{
"epoch": 0.5351535836177475,
"grad_norm": 0.14174342155456543,
"learning_rate": 0.00019628426517411625,
"loss": 0.5839,
"step": 196
},
{
"epoch": 0.5378839590443686,
"grad_norm": 0.15242989361286163,
"learning_rate": 0.0001961979052880609,
"loss": 0.5567,
"step": 197
},
{
"epoch": 0.5406143344709897,
"grad_norm": 0.16651766002178192,
"learning_rate": 0.00019611057276429085,
"loss": 0.5593,
"step": 198
},
{
"epoch": 0.543344709897611,
"grad_norm": 0.14858382940292358,
"learning_rate": 0.00019602226848580763,
"loss": 0.5848,
"step": 199
},
{
"epoch": 0.5460750853242321,
"grad_norm": 0.14774656295776367,
"learning_rate": 0.00019593299334543808,
"loss": 0.563,
"step": 200
},
{
"epoch": 0.5488054607508532,
"grad_norm": 0.13993892073631287,
"learning_rate": 0.0001958427482458253,
"loss": 0.5742,
"step": 201
},
{
"epoch": 0.5515358361774744,
"grad_norm": 0.15201717615127563,
"learning_rate": 0.0001957515340994193,
"loss": 0.5726,
"step": 202
},
{
"epoch": 0.5542662116040956,
"grad_norm": 0.1567879170179367,
"learning_rate": 0.00019565935182846802,
"loss": 0.5707,
"step": 203
},
{
"epoch": 0.5569965870307167,
"grad_norm": 0.13955365121364594,
"learning_rate": 0.00019556620236500793,
"loss": 0.5339,
"step": 204
},
{
"epoch": 0.5597269624573379,
"grad_norm": 0.1425381898880005,
"learning_rate": 0.00019547208665085457,
"loss": 0.5698,
"step": 205
},
{
"epoch": 0.562457337883959,
"grad_norm": 0.14695167541503906,
"learning_rate": 0.00019537700563759304,
"loss": 0.578,
"step": 206
},
{
"epoch": 0.5651877133105802,
"grad_norm": 0.15581448376178741,
"learning_rate": 0.00019528096028656832,
"loss": 0.5552,
"step": 207
},
{
"epoch": 0.5679180887372014,
"grad_norm": 0.13141174614429474,
"learning_rate": 0.00019518395156887576,
"loss": 0.5598,
"step": 208
},
{
"epoch": 0.5706484641638225,
"grad_norm": 0.16678418219089508,
"learning_rate": 0.00019508598046535095,
"loss": 0.5485,
"step": 209
},
{
"epoch": 0.5733788395904437,
"grad_norm": 0.1717272400856018,
"learning_rate": 0.00019498704796656018,
"loss": 0.5849,
"step": 210
},
{
"epoch": 0.5761092150170648,
"grad_norm": 0.14453086256980896,
"learning_rate": 0.00019488715507278998,
"loss": 0.5757,
"step": 211
},
{
"epoch": 0.578839590443686,
"grad_norm": 0.16329538822174072,
"learning_rate": 0.0001947863027940374,
"loss": 0.5632,
"step": 212
},
{
"epoch": 0.5815699658703072,
"grad_norm": 0.15865112841129303,
"learning_rate": 0.00019468449214999955,
"loss": 0.5728,
"step": 213
},
{
"epoch": 0.5843003412969283,
"grad_norm": 0.15316785871982574,
"learning_rate": 0.00019458172417006347,
"loss": 0.5556,
"step": 214
},
{
"epoch": 0.5870307167235495,
"grad_norm": 0.16666734218597412,
"learning_rate": 0.00019447799989329555,
"loss": 0.5759,
"step": 215
},
{
"epoch": 0.5897610921501707,
"grad_norm": 0.16525249183177948,
"learning_rate": 0.00019437332036843118,
"loss": 0.5667,
"step": 216
},
{
"epoch": 0.5924914675767918,
"grad_norm": 0.14022761583328247,
"learning_rate": 0.00019426768665386398,
"loss": 0.5611,
"step": 217
},
{
"epoch": 0.595221843003413,
"grad_norm": 0.15930500626564026,
"learning_rate": 0.00019416109981763526,
"loss": 0.5414,
"step": 218
},
{
"epoch": 0.5979522184300341,
"grad_norm": 0.141464963555336,
"learning_rate": 0.00019405356093742313,
"loss": 0.5363,
"step": 219
},
{
"epoch": 0.6006825938566553,
"grad_norm": 0.1541200429201126,
"learning_rate": 0.0001939450711005316,
"loss": 0.5487,
"step": 220
},
{
"epoch": 0.6034129692832765,
"grad_norm": 0.13717712461948395,
"learning_rate": 0.00019383563140387965,
"loss": 0.5564,
"step": 221
},
{
"epoch": 0.6061433447098976,
"grad_norm": 0.14139863848686218,
"learning_rate": 0.00019372524295399013,
"loss": 0.5592,
"step": 222
},
{
"epoch": 0.6088737201365187,
"grad_norm": 0.13494791090488434,
"learning_rate": 0.00019361390686697846,
"loss": 0.5452,
"step": 223
},
{
"epoch": 0.6116040955631399,
"grad_norm": 0.1512797623872757,
"learning_rate": 0.0001935016242685415,
"loss": 0.5595,
"step": 224
},
{
"epoch": 0.6143344709897611,
"grad_norm": 0.1422545164823532,
"learning_rate": 0.00019338839629394605,
"loss": 0.5602,
"step": 225
},
{
"epoch": 0.6170648464163823,
"grad_norm": 0.14444862306118011,
"learning_rate": 0.00019327422408801744,
"loss": 0.5574,
"step": 226
},
{
"epoch": 0.6197952218430034,
"grad_norm": 0.15434536337852478,
"learning_rate": 0.0001931591088051279,
"loss": 0.5467,
"step": 227
},
{
"epoch": 0.6225255972696245,
"grad_norm": 0.1420368254184723,
"learning_rate": 0.000193043051609185,
"loss": 0.5519,
"step": 228
},
{
"epoch": 0.6252559726962458,
"grad_norm": 0.12724490463733673,
"learning_rate": 0.00019292605367361978,
"loss": 0.5779,
"step": 229
},
{
"epoch": 0.6279863481228669,
"grad_norm": 0.13820625841617584,
"learning_rate": 0.00019280811618137484,
"loss": 0.5468,
"step": 230
},
{
"epoch": 0.630716723549488,
"grad_norm": 0.1629246473312378,
"learning_rate": 0.00019268924032489248,
"loss": 0.5721,
"step": 231
},
{
"epoch": 0.6334470989761092,
"grad_norm": 0.193836510181427,
"learning_rate": 0.00019256942730610268,
"loss": 0.5392,
"step": 232
},
{
"epoch": 0.6361774744027304,
"grad_norm": 0.18511579930782318,
"learning_rate": 0.0001924486783364108,
"loss": 0.5718,
"step": 233
},
{
"epoch": 0.6389078498293516,
"grad_norm": 0.14278100430965424,
"learning_rate": 0.00019232699463668542,
"loss": 0.5589,
"step": 234
},
{
"epoch": 0.6416382252559727,
"grad_norm": 0.14693580567836761,
"learning_rate": 0.00019220437743724605,
"loss": 0.515,
"step": 235
},
{
"epoch": 0.6443686006825938,
"grad_norm": 0.20018483698368073,
"learning_rate": 0.00019208082797785055,
"loss": 0.5484,
"step": 236
},
{
"epoch": 0.647098976109215,
"grad_norm": 0.1597984880208969,
"learning_rate": 0.00019195634750768275,
"loss": 0.539,
"step": 237
},
{
"epoch": 0.6498293515358362,
"grad_norm": 0.14270828664302826,
"learning_rate": 0.00019183093728533966,
"loss": 0.5496,
"step": 238
},
{
"epoch": 0.6525597269624573,
"grad_norm": 0.1788954734802246,
"learning_rate": 0.0001917045985788189,
"loss": 0.5629,
"step": 239
},
{
"epoch": 0.6552901023890785,
"grad_norm": 0.15234531462192535,
"learning_rate": 0.00019157733266550575,
"loss": 0.5454,
"step": 240
},
{
"epoch": 0.6580204778156996,
"grad_norm": 0.17805363237857819,
"learning_rate": 0.00019144914083216034,
"loss": 0.5599,
"step": 241
},
{
"epoch": 0.6607508532423209,
"grad_norm": 0.18313759565353394,
"learning_rate": 0.00019132002437490458,
"loss": 0.5544,
"step": 242
},
{
"epoch": 0.663481228668942,
"grad_norm": 0.16873271763324738,
"learning_rate": 0.00019118998459920902,
"loss": 0.5445,
"step": 243
},
{
"epoch": 0.6662116040955631,
"grad_norm": 0.23297423124313354,
"learning_rate": 0.00019105902281987976,
"loss": 0.5623,
"step": 244
},
{
"epoch": 0.6689419795221843,
"grad_norm": 0.13279114663600922,
"learning_rate": 0.00019092714036104508,
"loss": 0.5354,
"step": 245
},
{
"epoch": 0.6716723549488055,
"grad_norm": 0.18942788243293762,
"learning_rate": 0.00019079433855614201,
"loss": 0.5574,
"step": 246
},
{
"epoch": 0.6744027303754266,
"grad_norm": 0.15293289721012115,
"learning_rate": 0.00019066061874790303,
"loss": 0.5473,
"step": 247
},
{
"epoch": 0.6771331058020478,
"grad_norm": 0.1669221818447113,
"learning_rate": 0.00019052598228834217,
"loss": 0.5272,
"step": 248
},
{
"epoch": 0.6798634812286689,
"grad_norm": 0.16368111968040466,
"learning_rate": 0.00019039043053874175,
"loss": 0.5387,
"step": 249
},
{
"epoch": 0.6825938566552902,
"grad_norm": 0.15945561230182648,
"learning_rate": 0.00019025396486963827,
"loss": 0.548,
"step": 250
},
{
"epoch": 0.6853242320819113,
"grad_norm": 0.1538572758436203,
"learning_rate": 0.00019011658666080874,
"loss": 0.5419,
"step": 251
},
{
"epoch": 0.6880546075085324,
"grad_norm": 0.13356320559978485,
"learning_rate": 0.00018997829730125663,
"loss": 0.5271,
"step": 252
},
{
"epoch": 0.6907849829351536,
"grad_norm": 0.1389850527048111,
"learning_rate": 0.0001898390981891979,
"loss": 0.5489,
"step": 253
},
{
"epoch": 0.6935153583617747,
"grad_norm": 0.14726200699806213,
"learning_rate": 0.00018969899073204686,
"loss": 0.554,
"step": 254
},
{
"epoch": 0.6962457337883959,
"grad_norm": 0.13865283131599426,
"learning_rate": 0.0001895579763464019,
"loss": 0.5486,
"step": 255
},
{
"epoch": 0.6989761092150171,
"grad_norm": 0.15216147899627686,
"learning_rate": 0.00018941605645803115,
"loss": 0.5156,
"step": 256
},
{
"epoch": 0.7017064846416382,
"grad_norm": 0.13091793656349182,
"learning_rate": 0.00018927323250185815,
"loss": 0.5359,
"step": 257
},
{
"epoch": 0.7044368600682593,
"grad_norm": 0.14097946882247925,
"learning_rate": 0.0001891295059219472,
"loss": 0.5367,
"step": 258
},
{
"epoch": 0.7071672354948806,
"grad_norm": 0.1434548795223236,
"learning_rate": 0.00018898487817148898,
"loss": 0.5395,
"step": 259
},
{
"epoch": 0.7098976109215017,
"grad_norm": 0.15296564996242523,
"learning_rate": 0.0001888393507127856,
"loss": 0.5552,
"step": 260
},
{
"epoch": 0.7126279863481229,
"grad_norm": 0.15522688627243042,
"learning_rate": 0.000188692925017236,
"loss": 0.5433,
"step": 261
},
{
"epoch": 0.715358361774744,
"grad_norm": 0.1461726576089859,
"learning_rate": 0.000188545602565321,
"loss": 0.5291,
"step": 262
},
{
"epoch": 0.7180887372013652,
"grad_norm": 0.14327403903007507,
"learning_rate": 0.00018839738484658836,
"loss": 0.5502,
"step": 263
},
{
"epoch": 0.7208191126279864,
"grad_norm": 0.1613275706768036,
"learning_rate": 0.00018824827335963765,
"loss": 0.5565,
"step": 264
},
{
"epoch": 0.7235494880546075,
"grad_norm": 0.2040315866470337,
"learning_rate": 0.00018809826961210525,
"loss": 0.5324,
"step": 265
},
{
"epoch": 0.7262798634812286,
"grad_norm": 0.13186219334602356,
"learning_rate": 0.0001879473751206489,
"loss": 0.5395,
"step": 266
},
{
"epoch": 0.7290102389078499,
"grad_norm": 0.15462934970855713,
"learning_rate": 0.00018779559141093258,
"loss": 0.527,
"step": 267
},
{
"epoch": 0.731740614334471,
"grad_norm": 0.155660942196846,
"learning_rate": 0.0001876429200176108,
"loss": 0.5429,
"step": 268
},
{
"epoch": 0.7344709897610922,
"grad_norm": 0.15280231833457947,
"learning_rate": 0.00018748936248431353,
"loss": 0.5571,
"step": 269
},
{
"epoch": 0.7372013651877133,
"grad_norm": 0.19389967620372772,
"learning_rate": 0.00018733492036363005,
"loss": 0.5325,
"step": 270
},
{
"epoch": 0.7399317406143344,
"grad_norm": 0.14486828446388245,
"learning_rate": 0.0001871795952170937,
"loss": 0.5229,
"step": 271
},
{
"epoch": 0.7426621160409557,
"grad_norm": 0.18223008513450623,
"learning_rate": 0.00018702338861516587,
"loss": 0.538,
"step": 272
},
{
"epoch": 0.7453924914675768,
"grad_norm": 0.1434670090675354,
"learning_rate": 0.00018686630213722016,
"loss": 0.5353,
"step": 273
},
{
"epoch": 0.7481228668941979,
"grad_norm": 0.1796412169933319,
"learning_rate": 0.0001867083373715264,
"loss": 0.5221,
"step": 274
},
{
"epoch": 0.7508532423208191,
"grad_norm": 0.14365307986736298,
"learning_rate": 0.00018654949591523465,
"loss": 0.5211,
"step": 275
},
{
"epoch": 0.7535836177474403,
"grad_norm": 0.15224212408065796,
"learning_rate": 0.000186389779374359,
"loss": 0.5353,
"step": 276
},
{
"epoch": 0.7563139931740614,
"grad_norm": 0.1557937115430832,
"learning_rate": 0.00018622918936376132,
"loss": 0.5359,
"step": 277
},
{
"epoch": 0.7590443686006826,
"grad_norm": 0.13806480169296265,
"learning_rate": 0.00018606772750713504,
"loss": 0.5404,
"step": 278
},
{
"epoch": 0.7617747440273037,
"grad_norm": 0.14786981046199799,
"learning_rate": 0.00018590539543698854,
"loss": 0.5377,
"step": 279
},
{
"epoch": 0.764505119453925,
"grad_norm": 0.14674220979213715,
"learning_rate": 0.00018574219479462878,
"loss": 0.531,
"step": 280
},
{
"epoch": 0.7672354948805461,
"grad_norm": 0.146319180727005,
"learning_rate": 0.00018557812723014476,
"loss": 0.5244,
"step": 281
},
{
"epoch": 0.7699658703071672,
"grad_norm": 0.15404731035232544,
"learning_rate": 0.00018541319440239066,
"loss": 0.5419,
"step": 282
},
{
"epoch": 0.7726962457337884,
"grad_norm": 0.13534784317016602,
"learning_rate": 0.00018524739797896923,
"loss": 0.5162,
"step": 283
},
{
"epoch": 0.7754266211604095,
"grad_norm": 0.14462386071681976,
"learning_rate": 0.0001850807396362148,
"loss": 0.5338,
"step": 284
},
{
"epoch": 0.7781569965870307,
"grad_norm": 0.126676544547081,
"learning_rate": 0.00018491322105917645,
"loss": 0.5291,
"step": 285
},
{
"epoch": 0.7808873720136519,
"grad_norm": 0.1487646847963333,
"learning_rate": 0.0001847448439416009,
"loss": 0.5235,
"step": 286
},
{
"epoch": 0.783617747440273,
"grad_norm": 0.14197687804698944,
"learning_rate": 0.00018457560998591538,
"loss": 0.518,
"step": 287
},
{
"epoch": 0.7863481228668942,
"grad_norm": 0.13646024465560913,
"learning_rate": 0.00018440552090321047,
"loss": 0.5425,
"step": 288
},
{
"epoch": 0.7890784982935154,
"grad_norm": 0.15339186787605286,
"learning_rate": 0.00018423457841322273,
"loss": 0.5203,
"step": 289
},
{
"epoch": 0.7918088737201365,
"grad_norm": 0.14754898846149445,
"learning_rate": 0.00018406278424431736,
"loss": 0.5259,
"step": 290
},
{
"epoch": 0.7945392491467577,
"grad_norm": 0.1315944641828537,
"learning_rate": 0.00018389014013347078,
"loss": 0.5248,
"step": 291
},
{
"epoch": 0.7972696245733788,
"grad_norm": 0.14218468964099884,
"learning_rate": 0.00018371664782625287,
"loss": 0.5328,
"step": 292
},
{
"epoch": 0.8,
"grad_norm": 0.1456083059310913,
"learning_rate": 0.00018354230907680958,
"loss": 0.5352,
"step": 293
},
{
"epoch": 0.8027303754266212,
"grad_norm": 0.14852645993232727,
"learning_rate": 0.00018336712564784503,
"loss": 0.5379,
"step": 294
},
{
"epoch": 0.8054607508532423,
"grad_norm": 0.17426247894763947,
"learning_rate": 0.0001831910993106037,
"loss": 0.5358,
"step": 295
},
{
"epoch": 0.8081911262798634,
"grad_norm": 0.14848141372203827,
"learning_rate": 0.0001830142318448525,
"loss": 0.5377,
"step": 296
},
{
"epoch": 0.8109215017064847,
"grad_norm": 0.14610137045383453,
"learning_rate": 0.000182836525038863,
"loss": 0.5331,
"step": 297
},
{
"epoch": 0.8136518771331058,
"grad_norm": 0.1565829962491989,
"learning_rate": 0.00018265798068939294,
"loss": 0.5162,
"step": 298
},
{
"epoch": 0.816382252559727,
"grad_norm": 0.1328669935464859,
"learning_rate": 0.0001824786006016685,
"loss": 0.5045,
"step": 299
},
{
"epoch": 0.8191126279863481,
"grad_norm": 0.14329080283641815,
"learning_rate": 0.00018229838658936564,
"loss": 0.5083,
"step": 300
},
{
"epoch": 0.8218430034129692,
"grad_norm": 0.13927890360355377,
"learning_rate": 0.00018211734047459216,
"loss": 0.5221,
"step": 301
},
{
"epoch": 0.8245733788395905,
"grad_norm": 0.14164070785045624,
"learning_rate": 0.00018193546408786898,
"loss": 0.5243,
"step": 302
},
{
"epoch": 0.8273037542662116,
"grad_norm": 0.14735499024391174,
"learning_rate": 0.00018175275926811174,
"loss": 0.5136,
"step": 303
},
{
"epoch": 0.8300341296928327,
"grad_norm": 0.1411520093679428,
"learning_rate": 0.00018156922786261216,
"loss": 0.5164,
"step": 304
},
{
"epoch": 0.8327645051194539,
"grad_norm": 0.14603470265865326,
"learning_rate": 0.0001813848717270195,
"loss": 0.5306,
"step": 305
},
{
"epoch": 0.8354948805460751,
"grad_norm": 0.13602910935878754,
"learning_rate": 0.00018119969272532166,
"loss": 0.5014,
"step": 306
},
{
"epoch": 0.8382252559726963,
"grad_norm": 0.15259377658367157,
"learning_rate": 0.00018101369272982632,
"loss": 0.5309,
"step": 307
},
{
"epoch": 0.8409556313993174,
"grad_norm": 0.13250286877155304,
"learning_rate": 0.00018082687362114212,
"loss": 0.5034,
"step": 308
},
{
"epoch": 0.8436860068259385,
"grad_norm": 0.15984928607940674,
"learning_rate": 0.00018063923728815957,
"loss": 0.5167,
"step": 309
},
{
"epoch": 0.8464163822525598,
"grad_norm": 0.14640702307224274,
"learning_rate": 0.00018045078562803203,
"loss": 0.5361,
"step": 310
},
{
"epoch": 0.8491467576791809,
"grad_norm": 0.1568063646554947,
"learning_rate": 0.00018026152054615634,
"loss": 0.5221,
"step": 311
},
{
"epoch": 0.851877133105802,
"grad_norm": 0.14193738996982574,
"learning_rate": 0.0001800714439561538,
"loss": 0.5151,
"step": 312
},
{
"epoch": 0.8546075085324232,
"grad_norm": 0.15847285091876984,
"learning_rate": 0.00017988055777985067,
"loss": 0.4923,
"step": 313
},
{
"epoch": 0.8573378839590444,
"grad_norm": 0.16404108703136444,
"learning_rate": 0.00017968886394725874,
"loss": 0.5114,
"step": 314
},
{
"epoch": 0.8600682593856656,
"grad_norm": 0.22749370336532593,
"learning_rate": 0.00017949636439655592,
"loss": 0.5225,
"step": 315
},
{
"epoch": 0.8627986348122867,
"grad_norm": 0.17752817273139954,
"learning_rate": 0.00017930306107406653,
"loss": 0.5138,
"step": 316
},
{
"epoch": 0.8655290102389078,
"grad_norm": 0.16554813086986542,
"learning_rate": 0.00017910895593424163,
"loss": 0.5087,
"step": 317
},
{
"epoch": 0.868259385665529,
"grad_norm": 0.1888076812028885,
"learning_rate": 0.00017891405093963938,
"loss": 0.5163,
"step": 318
},
{
"epoch": 0.8709897610921502,
"grad_norm": 0.15817560255527496,
"learning_rate": 0.00017871834806090501,
"loss": 0.4918,
"step": 319
},
{
"epoch": 0.8737201365187713,
"grad_norm": 0.15167172253131866,
"learning_rate": 0.00017852184927675112,
"loss": 0.5248,
"step": 320
},
{
"epoch": 0.8764505119453925,
"grad_norm": 0.15847422182559967,
"learning_rate": 0.00017832455657393746,
"loss": 0.5205,
"step": 321
},
{
"epoch": 0.8791808873720136,
"grad_norm": 0.1436643898487091,
"learning_rate": 0.00017812647194725094,
"loss": 0.4869,
"step": 322
},
{
"epoch": 0.8819112627986349,
"grad_norm": 0.1533607393503189,
"learning_rate": 0.00017792759739948546,
"loss": 0.5141,
"step": 323
},
{
"epoch": 0.884641638225256,
"grad_norm": 0.1520746648311615,
"learning_rate": 0.00017772793494142167,
"loss": 0.5158,
"step": 324
},
{
"epoch": 0.8873720136518771,
"grad_norm": 0.1398312896490097,
"learning_rate": 0.0001775274865918066,
"loss": 0.5126,
"step": 325
},
{
"epoch": 0.8901023890784983,
"grad_norm": 0.14314454793930054,
"learning_rate": 0.00017732625437733335,
"loss": 0.5187,
"step": 326
},
{
"epoch": 0.8928327645051195,
"grad_norm": 0.1434595286846161,
"learning_rate": 0.00017712424033262042,
"loss": 0.5018,
"step": 327
},
{
"epoch": 0.8955631399317406,
"grad_norm": 0.13712714612483978,
"learning_rate": 0.00017692144650019125,
"loss": 0.5044,
"step": 328
},
{
"epoch": 0.8982935153583618,
"grad_norm": 0.13560262322425842,
"learning_rate": 0.00017671787493045356,
"loss": 0.5153,
"step": 329
},
{
"epoch": 0.9010238907849829,
"grad_norm": 0.15035240352153778,
"learning_rate": 0.0001765135276816787,
"loss": 0.5023,
"step": 330
},
{
"epoch": 0.903754266211604,
"grad_norm": 0.14878690242767334,
"learning_rate": 0.00017630840681998066,
"loss": 0.4916,
"step": 331
},
{
"epoch": 0.9064846416382253,
"grad_norm": 0.14930297434329987,
"learning_rate": 0.00017610251441929533,
"loss": 0.5092,
"step": 332
},
{
"epoch": 0.9092150170648464,
"grad_norm": 0.14058218896389008,
"learning_rate": 0.0001758958525613594,
"loss": 0.5091,
"step": 333
},
{
"epoch": 0.9119453924914676,
"grad_norm": 0.18505944311618805,
"learning_rate": 0.00017568842333568952,
"loss": 0.5047,
"step": 334
},
{
"epoch": 0.9146757679180887,
"grad_norm": 0.16390664875507355,
"learning_rate": 0.0001754802288395609,
"loss": 0.5115,
"step": 335
},
{
"epoch": 0.9174061433447099,
"grad_norm": 0.16126009821891785,
"learning_rate": 0.00017527127117798635,
"loss": 0.5017,
"step": 336
},
{
"epoch": 0.9201365187713311,
"grad_norm": 0.16674400866031647,
"learning_rate": 0.0001750615524636948,
"loss": 0.5193,
"step": 337
},
{
"epoch": 0.9228668941979522,
"grad_norm": 0.13656651973724365,
"learning_rate": 0.00017485107481711012,
"loss": 0.4909,
"step": 338
},
{
"epoch": 0.9255972696245733,
"grad_norm": 0.15961140394210815,
"learning_rate": 0.00017463984036632954,
"loss": 0.5145,
"step": 339
},
{
"epoch": 0.9283276450511946,
"grad_norm": 0.13122031092643738,
"learning_rate": 0.00017442785124710227,
"loss": 0.4854,
"step": 340
},
{
"epoch": 0.9310580204778157,
"grad_norm": 0.14256154000759125,
"learning_rate": 0.0001742151096028076,
"loss": 0.4852,
"step": 341
},
{
"epoch": 0.9337883959044369,
"grad_norm": 0.14853668212890625,
"learning_rate": 0.00017400161758443375,
"loss": 0.5202,
"step": 342
},
{
"epoch": 0.936518771331058,
"grad_norm": 0.16885364055633545,
"learning_rate": 0.00017378737735055562,
"loss": 0.5136,
"step": 343
},
{
"epoch": 0.9392491467576792,
"grad_norm": 0.1398458182811737,
"learning_rate": 0.00017357239106731317,
"loss": 0.4908,
"step": 344
},
{
"epoch": 0.9419795221843004,
"grad_norm": 0.1871606856584549,
"learning_rate": 0.00017335666090838965,
"loss": 0.5076,
"step": 345
},
{
"epoch": 0.9447098976109215,
"grad_norm": 0.17375800013542175,
"learning_rate": 0.00017314018905498931,
"loss": 0.5032,
"step": 346
},
{
"epoch": 0.9474402730375426,
"grad_norm": 0.14859388768672943,
"learning_rate": 0.00017292297769581571,
"loss": 0.5119,
"step": 347
},
{
"epoch": 0.9501706484641638,
"grad_norm": 0.18141716718673706,
"learning_rate": 0.00017270502902704926,
"loss": 0.4982,
"step": 348
},
{
"epoch": 0.952901023890785,
"grad_norm": 0.17015349864959717,
"learning_rate": 0.00017248634525232523,
"loss": 0.4835,
"step": 349
},
{
"epoch": 0.9556313993174061,
"grad_norm": 0.14770372211933136,
"learning_rate": 0.00017226692858271134,
"loss": 0.4851,
"step": 350
},
{
"epoch": 0.9583617747440273,
"grad_norm": 0.19386035203933716,
"learning_rate": 0.00017204678123668556,
"loss": 0.5172,
"step": 351
},
{
"epoch": 0.9610921501706484,
"grad_norm": 0.15225833654403687,
"learning_rate": 0.00017182590544011347,
"loss": 0.4972,
"step": 352
},
{
"epoch": 0.9638225255972697,
"grad_norm": 0.1522843837738037,
"learning_rate": 0.00017160430342622596,
"loss": 0.493,
"step": 353
},
{
"epoch": 0.9665529010238908,
"grad_norm": 0.13483871519565582,
"learning_rate": 0.00017138197743559654,
"loss": 0.4882,
"step": 354
},
{
"epoch": 0.9692832764505119,
"grad_norm": 0.14176194369792938,
"learning_rate": 0.00017115892971611863,
"loss": 0.4921,
"step": 355
},
{
"epoch": 0.9720136518771331,
"grad_norm": 0.15088023245334625,
"learning_rate": 0.00017093516252298296,
"loss": 0.5073,
"step": 356
},
{
"epoch": 0.9747440273037543,
"grad_norm": 0.14758853614330292,
"learning_rate": 0.00017071067811865476,
"loss": 0.5106,
"step": 357
},
{
"epoch": 0.9774744027303754,
"grad_norm": 0.14779959619045258,
"learning_rate": 0.00017048547877285077,
"loss": 0.5066,
"step": 358
},
{
"epoch": 0.9802047781569966,
"grad_norm": 0.14550894498825073,
"learning_rate": 0.00017025956676251636,
"loss": 0.484,
"step": 359
},
{
"epoch": 0.9829351535836177,
"grad_norm": 0.14111794531345367,
"learning_rate": 0.00017003294437180255,
"loss": 0.4719,
"step": 360
},
{
"epoch": 0.985665529010239,
"grad_norm": 0.15354299545288086,
"learning_rate": 0.00016980561389204284,
"loss": 0.4874,
"step": 361
},
{
"epoch": 0.9883959044368601,
"grad_norm": 0.14522601664066315,
"learning_rate": 0.0001695775776217301,
"loss": 0.4722,
"step": 362
},
{
"epoch": 0.9911262798634812,
"grad_norm": 0.143354594707489,
"learning_rate": 0.00016934883786649333,
"loss": 0.5038,
"step": 363
},
{
"epoch": 0.9938566552901024,
"grad_norm": 0.14298783242702484,
"learning_rate": 0.0001691193969390742,
"loss": 0.5007,
"step": 364
},
{
"epoch": 0.9965870307167235,
"grad_norm": 0.14418889582157135,
"learning_rate": 0.00016888925715930394,
"loss": 0.5054,
"step": 365
},
{
"epoch": 0.9993174061433447,
"grad_norm": 0.16697633266448975,
"learning_rate": 0.0001686584208540797,
"loss": 0.4948,
"step": 366
},
{
"epoch": 1.0020477815699658,
"grad_norm": 0.4087926745414734,
"learning_rate": 0.000168426890357341,
"loss": 0.8762,
"step": 367
},
{
"epoch": 1.004778156996587,
"grad_norm": 0.17126424610614777,
"learning_rate": 0.00016819466801004621,
"loss": 0.4964,
"step": 368
},
{
"epoch": 1.0075085324232083,
"grad_norm": 0.18887566030025482,
"learning_rate": 0.00016796175616014893,
"loss": 0.4842,
"step": 369
},
{
"epoch": 1.0102389078498293,
"grad_norm": 0.16404148936271667,
"learning_rate": 0.00016772815716257412,
"loss": 0.4761,
"step": 370
},
{
"epoch": 1.0129692832764505,
"grad_norm": 0.2001708298921585,
"learning_rate": 0.00016749387337919433,
"loss": 0.5005,
"step": 371
},
{
"epoch": 1.0156996587030718,
"grad_norm": 0.16301579773426056,
"learning_rate": 0.0001672589071788059,
"loss": 0.497,
"step": 372
},
{
"epoch": 1.0184300341296928,
"grad_norm": 0.15777342021465302,
"learning_rate": 0.0001670232609371049,
"loss": 0.4771,
"step": 373
},
{
"epoch": 1.021160409556314,
"grad_norm": 0.14179396629333496,
"learning_rate": 0.00016678693703666325,
"loss": 0.4491,
"step": 374
},
{
"epoch": 1.023890784982935,
"grad_norm": 0.16155582666397095,
"learning_rate": 0.00016654993786690444,
"loss": 0.4846,
"step": 375
},
{
"epoch": 1.0266211604095563,
"grad_norm": 0.15554013848304749,
"learning_rate": 0.00016631226582407952,
"loss": 0.4663,
"step": 376
},
{
"epoch": 1.0293515358361776,
"grad_norm": 0.15946047008037567,
"learning_rate": 0.00016607392331124282,
"loss": 0.4978,
"step": 377
},
{
"epoch": 1.0320819112627986,
"grad_norm": 0.14758449792861938,
"learning_rate": 0.00016583491273822765,
"loss": 0.4868,
"step": 378
},
{
"epoch": 1.0348122866894198,
"grad_norm": 0.16146820783615112,
"learning_rate": 0.0001655952365216219,
"loss": 0.5056,
"step": 379
},
{
"epoch": 1.0375426621160408,
"grad_norm": 0.14648115634918213,
"learning_rate": 0.0001653548970847438,
"loss": 0.5028,
"step": 380
},
{
"epoch": 1.040273037542662,
"grad_norm": 0.14611080288887024,
"learning_rate": 0.00016511389685761708,
"loss": 0.4975,
"step": 381
},
{
"epoch": 1.0430034129692833,
"grad_norm": 0.13934309780597687,
"learning_rate": 0.00016487223827694672,
"loss": 0.4881,
"step": 382
},
{
"epoch": 1.0457337883959044,
"grad_norm": 0.1680765300989151,
"learning_rate": 0.00016462992378609407,
"loss": 0.4947,
"step": 383
},
{
"epoch": 1.0484641638225256,
"grad_norm": 0.16694432497024536,
"learning_rate": 0.00016438695583505242,
"loss": 0.5,
"step": 384
},
{
"epoch": 1.0511945392491469,
"grad_norm": 0.13027995824813843,
"learning_rate": 0.00016414333688042186,
"loss": 0.4783,
"step": 385
},
{
"epoch": 1.0539249146757679,
"grad_norm": 0.15040864050388336,
"learning_rate": 0.0001638990693853848,
"loss": 0.4743,
"step": 386
},
{
"epoch": 1.0566552901023891,
"grad_norm": 0.17924848198890686,
"learning_rate": 0.00016365415581968084,
"loss": 0.493,
"step": 387
},
{
"epoch": 1.0593856655290101,
"grad_norm": 0.14788490533828735,
"learning_rate": 0.0001634085986595819,
"loss": 0.4932,
"step": 388
},
{
"epoch": 1.0621160409556314,
"grad_norm": 0.1745985448360443,
"learning_rate": 0.00016316240038786718,
"loss": 0.5192,
"step": 389
},
{
"epoch": 1.0648464163822526,
"grad_norm": 0.1356726437807083,
"learning_rate": 0.00016291556349379795,
"loss": 0.4829,
"step": 390
},
{
"epoch": 1.0675767918088737,
"grad_norm": 0.1533443182706833,
"learning_rate": 0.00016266809047309251,
"loss": 0.4881,
"step": 391
},
{
"epoch": 1.070307167235495,
"grad_norm": 0.16772933304309845,
"learning_rate": 0.00016241998382790095,
"loss": 0.4967,
"step": 392
},
{
"epoch": 1.073037542662116,
"grad_norm": 0.1441749781370163,
"learning_rate": 0.0001621712460667797,
"loss": 0.4867,
"step": 393
},
{
"epoch": 1.0757679180887372,
"grad_norm": 0.15063367784023285,
"learning_rate": 0.00016192187970466644,
"loss": 0.4916,
"step": 394
},
{
"epoch": 1.0784982935153584,
"grad_norm": 0.14568567276000977,
"learning_rate": 0.00016167188726285434,
"loss": 0.488,
"step": 395
},
{
"epoch": 1.0812286689419794,
"grad_norm": 0.14870832860469818,
"learning_rate": 0.0001614212712689668,
"loss": 0.4867,
"step": 396
},
{
"epoch": 1.0839590443686007,
"grad_norm": 0.15368396043777466,
"learning_rate": 0.00016117003425693188,
"loss": 0.4766,
"step": 397
},
{
"epoch": 1.086689419795222,
"grad_norm": 0.163841113448143,
"learning_rate": 0.00016091817876695655,
"loss": 0.4803,
"step": 398
},
{
"epoch": 1.089419795221843,
"grad_norm": 0.17485839128494263,
"learning_rate": 0.0001606657073455012,
"loss": 0.4875,
"step": 399
},
{
"epoch": 1.0921501706484642,
"grad_norm": 0.14765095710754395,
"learning_rate": 0.00016041262254525362,
"loss": 0.5058,
"step": 400
},
{
"epoch": 1.0948805460750852,
"grad_norm": 0.1618352234363556,
"learning_rate": 0.0001601589269251035,
"loss": 0.4903,
"step": 401
},
{
"epoch": 1.0976109215017065,
"grad_norm": 0.1456744223833084,
"learning_rate": 0.0001599046230501163,
"loss": 0.4822,
"step": 402
},
{
"epoch": 1.1003412969283277,
"grad_norm": 0.15738914906978607,
"learning_rate": 0.00015964971349150746,
"loss": 0.4944,
"step": 403
},
{
"epoch": 1.1030716723549487,
"grad_norm": 0.1475927084684372,
"learning_rate": 0.0001593942008266164,
"loss": 0.4971,
"step": 404
},
{
"epoch": 1.10580204778157,
"grad_norm": 0.14080214500427246,
"learning_rate": 0.00015913808763888039,
"loss": 0.4864,
"step": 405
},
{
"epoch": 1.108532423208191,
"grad_norm": 0.16853410005569458,
"learning_rate": 0.00015888137651780845,
"loss": 0.4891,
"step": 406
},
{
"epoch": 1.1112627986348123,
"grad_norm": 0.13720810413360596,
"learning_rate": 0.00015862407005895522,
"loss": 0.4822,
"step": 407
},
{
"epoch": 1.1139931740614335,
"grad_norm": 0.16895629465579987,
"learning_rate": 0.00015836617086389468,
"loss": 0.4598,
"step": 408
},
{
"epoch": 1.1167235494880545,
"grad_norm": 0.15242214500904083,
"learning_rate": 0.00015810768154019385,
"loss": 0.4761,
"step": 409
},
{
"epoch": 1.1194539249146758,
"grad_norm": 0.1633676141500473,
"learning_rate": 0.00015784860470138633,
"loss": 0.4912,
"step": 410
},
{
"epoch": 1.122184300341297,
"grad_norm": 0.16565294563770294,
"learning_rate": 0.00015758894296694615,
"loss": 0.4836,
"step": 411
},
{
"epoch": 1.124914675767918,
"grad_norm": 0.160204216837883,
"learning_rate": 0.00015732869896226094,
"loss": 0.4694,
"step": 412
},
{
"epoch": 1.1276450511945393,
"grad_norm": 0.19067788124084473,
"learning_rate": 0.00015706787531860557,
"loss": 0.4853,
"step": 413
},
{
"epoch": 1.1303754266211605,
"grad_norm": 0.14547327160835266,
"learning_rate": 0.00015680647467311557,
"loss": 0.4709,
"step": 414
},
{
"epoch": 1.1331058020477816,
"grad_norm": 0.18047383427619934,
"learning_rate": 0.0001565444996687605,
"loss": 0.4738,
"step": 415
},
{
"epoch": 1.1358361774744028,
"grad_norm": 0.16099834442138672,
"learning_rate": 0.00015628195295431697,
"loss": 0.4805,
"step": 416
},
{
"epoch": 1.1385665529010238,
"grad_norm": 0.1426517814397812,
"learning_rate": 0.00015601883718434207,
"loss": 0.4573,
"step": 417
},
{
"epoch": 1.141296928327645,
"grad_norm": 0.14879460632801056,
"learning_rate": 0.00015575515501914668,
"loss": 0.4778,
"step": 418
},
{
"epoch": 1.144027303754266,
"grad_norm": 0.1393750160932541,
"learning_rate": 0.0001554909091247682,
"loss": 0.447,
"step": 419
},
{
"epoch": 1.1467576791808873,
"grad_norm": 0.17599152028560638,
"learning_rate": 0.00015522610217294375,
"loss": 0.4969,
"step": 420
},
{
"epoch": 1.1494880546075086,
"grad_norm": 0.17399606108665466,
"learning_rate": 0.0001549607368410834,
"loss": 0.4933,
"step": 421
},
{
"epoch": 1.1522184300341296,
"grad_norm": 0.16555064916610718,
"learning_rate": 0.00015469481581224272,
"loss": 0.475,
"step": 422
},
{
"epoch": 1.1549488054607508,
"grad_norm": 0.20848453044891357,
"learning_rate": 0.00015442834177509582,
"loss": 0.4872,
"step": 423
},
{
"epoch": 1.157679180887372,
"grad_norm": 0.15129271149635315,
"learning_rate": 0.00015416131742390827,
"loss": 0.4963,
"step": 424
},
{
"epoch": 1.1604095563139931,
"grad_norm": 0.1703735888004303,
"learning_rate": 0.00015389374545850973,
"loss": 0.479,
"step": 425
},
{
"epoch": 1.1631399317406144,
"grad_norm": 0.152608722448349,
"learning_rate": 0.00015362562858426654,
"loss": 0.4831,
"step": 426
},
{
"epoch": 1.1658703071672356,
"grad_norm": 0.14749537408351898,
"learning_rate": 0.0001533569695120547,
"loss": 0.4784,
"step": 427
},
{
"epoch": 1.1686006825938566,
"grad_norm": 0.15642118453979492,
"learning_rate": 0.0001530877709582321,
"loss": 0.4679,
"step": 428
},
{
"epoch": 1.1713310580204779,
"grad_norm": 0.13505250215530396,
"learning_rate": 0.00015281803564461135,
"loss": 0.4779,
"step": 429
},
{
"epoch": 1.174061433447099,
"grad_norm": 0.14518040418624878,
"learning_rate": 0.00015254776629843205,
"loss": 0.448,
"step": 430
},
{
"epoch": 1.1767918088737201,
"grad_norm": 0.12947289645671844,
"learning_rate": 0.0001522769656523333,
"loss": 0.4735,
"step": 431
},
{
"epoch": 1.1795221843003414,
"grad_norm": 0.15066657960414886,
"learning_rate": 0.00015200563644432612,
"loss": 0.4791,
"step": 432
},
{
"epoch": 1.1822525597269624,
"grad_norm": 0.13076473772525787,
"learning_rate": 0.00015173378141776568,
"loss": 0.4582,
"step": 433
},
{
"epoch": 1.1849829351535837,
"grad_norm": 0.15804897248744965,
"learning_rate": 0.00015146140332132358,
"loss": 0.482,
"step": 434
},
{
"epoch": 1.1877133105802047,
"grad_norm": 0.1330975741147995,
"learning_rate": 0.00015118850490896012,
"loss": 0.4736,
"step": 435
},
{
"epoch": 1.190443686006826,
"grad_norm": 0.16358092427253723,
"learning_rate": 0.00015091508893989633,
"loss": 0.4992,
"step": 436
},
{
"epoch": 1.1931740614334472,
"grad_norm": 0.14177009463310242,
"learning_rate": 0.00015064115817858622,
"loss": 0.4646,
"step": 437
},
{
"epoch": 1.1959044368600682,
"grad_norm": 0.1569090485572815,
"learning_rate": 0.00015036671539468878,
"loss": 0.4683,
"step": 438
},
{
"epoch": 1.1986348122866894,
"grad_norm": 0.15897628664970398,
"learning_rate": 0.00015009176336303986,
"loss": 0.4829,
"step": 439
},
{
"epoch": 1.2013651877133107,
"grad_norm": 0.1482827216386795,
"learning_rate": 0.00014981630486362435,
"loss": 0.4552,
"step": 440
},
{
"epoch": 1.2040955631399317,
"grad_norm": 0.1546843945980072,
"learning_rate": 0.00014954034268154778,
"loss": 0.4702,
"step": 441
},
{
"epoch": 1.206825938566553,
"grad_norm": 0.1493782103061676,
"learning_rate": 0.00014926387960700842,
"loss": 0.4708,
"step": 442
},
{
"epoch": 1.209556313993174,
"grad_norm": 0.15456125140190125,
"learning_rate": 0.00014898691843526899,
"loss": 0.4738,
"step": 443
},
{
"epoch": 1.2122866894197952,
"grad_norm": 0.17920009791851044,
"learning_rate": 0.00014870946196662822,
"loss": 0.4616,
"step": 444
},
{
"epoch": 1.2150170648464165,
"grad_norm": 0.15904481709003448,
"learning_rate": 0.00014843151300639282,
"loss": 0.4791,
"step": 445
},
{
"epoch": 1.2177474402730375,
"grad_norm": 0.18129463493824005,
"learning_rate": 0.00014815307436484898,
"loss": 0.4789,
"step": 446
},
{
"epoch": 1.2204778156996587,
"grad_norm": 0.14939218759536743,
"learning_rate": 0.00014787414885723385,
"loss": 0.4774,
"step": 447
},
{
"epoch": 1.2232081911262798,
"grad_norm": 0.17625784873962402,
"learning_rate": 0.00014759473930370736,
"loss": 0.4785,
"step": 448
},
{
"epoch": 1.225938566552901,
"grad_norm": 0.158066987991333,
"learning_rate": 0.0001473148485293234,
"loss": 0.4588,
"step": 449
},
{
"epoch": 1.2286689419795223,
"grad_norm": 0.16107094287872314,
"learning_rate": 0.00014703447936400134,
"loss": 0.4734,
"step": 450
},
{
"epoch": 1.2313993174061433,
"grad_norm": 0.1387171894311905,
"learning_rate": 0.00014675363464249763,
"loss": 0.4733,
"step": 451
},
{
"epoch": 1.2341296928327645,
"grad_norm": 0.1561625450849533,
"learning_rate": 0.00014647231720437686,
"loss": 0.4502,
"step": 452
},
{
"epoch": 1.2368600682593858,
"grad_norm": 0.1419583112001419,
"learning_rate": 0.00014619052989398322,
"loss": 0.4674,
"step": 453
},
{
"epoch": 1.2395904436860068,
"grad_norm": 0.1578853279352188,
"learning_rate": 0.00014590827556041158,
"loss": 0.4635,
"step": 454
},
{
"epoch": 1.242320819112628,
"grad_norm": 0.1325417309999466,
"learning_rate": 0.00014562555705747892,
"loss": 0.4557,
"step": 455
},
{
"epoch": 1.245051194539249,
"grad_norm": 0.17075014114379883,
"learning_rate": 0.00014534237724369534,
"loss": 0.4928,
"step": 456
},
{
"epoch": 1.2477815699658703,
"grad_norm": 0.136972576379776,
"learning_rate": 0.00014505873898223496,
"loss": 0.4699,
"step": 457
},
{
"epoch": 1.2505119453924913,
"grad_norm": 0.16119800508022308,
"learning_rate": 0.00014477464514090743,
"loss": 0.4807,
"step": 458
},
{
"epoch": 1.2532423208191126,
"grad_norm": 0.13321495056152344,
"learning_rate": 0.00014449009859212857,
"loss": 0.4806,
"step": 459
},
{
"epoch": 1.2559726962457338,
"grad_norm": 0.14326980710029602,
"learning_rate": 0.00014420510221289137,
"loss": 0.4491,
"step": 460
},
{
"epoch": 1.2587030716723548,
"grad_norm": 0.13098248839378357,
"learning_rate": 0.00014391965888473703,
"loss": 0.4653,
"step": 461
},
{
"epoch": 1.261433447098976,
"grad_norm": 0.14561250805854797,
"learning_rate": 0.00014363377149372584,
"loss": 0.479,
"step": 462
},
{
"epoch": 1.2641638225255973,
"grad_norm": 0.1377183198928833,
"learning_rate": 0.0001433474429304077,
"loss": 0.4607,
"step": 463
},
{
"epoch": 1.2668941979522184,
"grad_norm": 0.14758490025997162,
"learning_rate": 0.0001430606760897934,
"loss": 0.472,
"step": 464
},
{
"epoch": 1.2696245733788396,
"grad_norm": 0.15359081327915192,
"learning_rate": 0.0001427734738713248,
"loss": 0.469,
"step": 465
},
{
"epoch": 1.2723549488054609,
"grad_norm": 0.16317234933376312,
"learning_rate": 0.00014248583917884594,
"loss": 0.4749,
"step": 466
},
{
"epoch": 1.2750853242320819,
"grad_norm": 0.1310819685459137,
"learning_rate": 0.00014219777492057348,
"loss": 0.4722,
"step": 467
},
{
"epoch": 1.2778156996587031,
"grad_norm": 0.14236976206302643,
"learning_rate": 0.0001419092840090673,
"loss": 0.4704,
"step": 468
},
{
"epoch": 1.2805460750853244,
"grad_norm": 0.1274275928735733,
"learning_rate": 0.00014162036936120115,
"loss": 0.4687,
"step": 469
},
{
"epoch": 1.2832764505119454,
"grad_norm": 0.13622865080833435,
"learning_rate": 0.00014133103389813302,
"loss": 0.4649,
"step": 470
},
{
"epoch": 1.2860068259385666,
"grad_norm": 0.14750082790851593,
"learning_rate": 0.0001410412805452757,
"loss": 0.4654,
"step": 471
},
{
"epoch": 1.2887372013651877,
"grad_norm": 0.14838138222694397,
"learning_rate": 0.0001407511122322672,
"loss": 0.4626,
"step": 472
},
{
"epoch": 1.291467576791809,
"grad_norm": 0.1439974308013916,
"learning_rate": 0.00014046053189294112,
"loss": 0.4499,
"step": 473
},
{
"epoch": 1.29419795221843,
"grad_norm": 0.14686186611652374,
"learning_rate": 0.00014016954246529696,
"loss": 0.4676,
"step": 474
},
{
"epoch": 1.2969283276450512,
"grad_norm": 0.15870115160942078,
"learning_rate": 0.00013987814689147041,
"loss": 0.4448,
"step": 475
},
{
"epoch": 1.2996587030716724,
"grad_norm": 0.16382190585136414,
"learning_rate": 0.0001395863481177036,
"loss": 0.4476,
"step": 476
},
{
"epoch": 1.3023890784982934,
"grad_norm": 0.21564428508281708,
"learning_rate": 0.00013929414909431544,
"loss": 0.4751,
"step": 477
},
{
"epoch": 1.3051194539249147,
"grad_norm": 0.15450774133205414,
"learning_rate": 0.00013900155277567157,
"loss": 0.4776,
"step": 478
},
{
"epoch": 1.307849829351536,
"grad_norm": 0.15922100841999054,
"learning_rate": 0.00013870856212015468,
"loss": 0.4693,
"step": 479
},
{
"epoch": 1.310580204778157,
"grad_norm": 0.13661661744117737,
"learning_rate": 0.00013841518009013445,
"loss": 0.454,
"step": 480
},
{
"epoch": 1.3133105802047782,
"grad_norm": 0.1507856249809265,
"learning_rate": 0.00013812140965193773,
"loss": 0.4701,
"step": 481
},
{
"epoch": 1.3160409556313994,
"grad_norm": 0.14306284487247467,
"learning_rate": 0.00013782725377581848,
"loss": 0.4842,
"step": 482
},
{
"epoch": 1.3187713310580205,
"grad_norm": 0.15984225273132324,
"learning_rate": 0.00013753271543592773,
"loss": 0.4628,
"step": 483
},
{
"epoch": 1.3215017064846417,
"grad_norm": 0.15242429077625275,
"learning_rate": 0.00013723779761028347,
"loss": 0.4708,
"step": 484
},
{
"epoch": 1.3242320819112627,
"grad_norm": 0.14520719647407532,
"learning_rate": 0.0001369425032807407,
"loss": 0.4608,
"step": 485
},
{
"epoch": 1.326962457337884,
"grad_norm": 0.16152748465538025,
"learning_rate": 0.00013664683543296112,
"loss": 0.4729,
"step": 486
},
{
"epoch": 1.329692832764505,
"grad_norm": 0.15082891285419464,
"learning_rate": 0.00013635079705638298,
"loss": 0.4593,
"step": 487
},
{
"epoch": 1.3324232081911263,
"grad_norm": 0.16038447618484497,
"learning_rate": 0.00013605439114419094,
"loss": 0.4718,
"step": 488
},
{
"epoch": 1.3351535836177475,
"grad_norm": 0.1532922238111496,
"learning_rate": 0.00013575762069328566,
"loss": 0.4758,
"step": 489
},
{
"epoch": 1.3378839590443685,
"grad_norm": 0.12969861924648285,
"learning_rate": 0.00013546048870425356,
"loss": 0.4582,
"step": 490
},
{
"epoch": 1.3406143344709898,
"grad_norm": 0.13830237090587616,
"learning_rate": 0.00013516299818133664,
"loss": 0.4585,
"step": 491
},
{
"epoch": 1.343344709897611,
"grad_norm": 0.148755744099617,
"learning_rate": 0.00013486515213240188,
"loss": 0.4803,
"step": 492
},
{
"epoch": 1.346075085324232,
"grad_norm": 0.13623669743537903,
"learning_rate": 0.0001345669535689108,
"loss": 0.4562,
"step": 493
},
{
"epoch": 1.3488054607508533,
"grad_norm": 0.14996616542339325,
"learning_rate": 0.00013426840550588933,
"loss": 0.4546,
"step": 494
},
{
"epoch": 1.3515358361774745,
"grad_norm": 0.1468917429447174,
"learning_rate": 0.000133969510961897,
"loss": 0.4712,
"step": 495
},
{
"epoch": 1.3542662116040955,
"grad_norm": 0.14525148272514343,
"learning_rate": 0.0001336702729589965,
"loss": 0.4614,
"step": 496
},
{
"epoch": 1.3569965870307168,
"grad_norm": 0.13571806252002716,
"learning_rate": 0.00013337069452272333,
"loss": 0.4601,
"step": 497
},
{
"epoch": 1.3597269624573378,
"grad_norm": 0.15235814452171326,
"learning_rate": 0.00013307077868205487,
"loss": 0.4785,
"step": 498
},
{
"epoch": 1.362457337883959,
"grad_norm": 0.14220909774303436,
"learning_rate": 0.00013277052846937996,
"loss": 0.4561,
"step": 499
},
{
"epoch": 1.36518771331058,
"grad_norm": 0.14699751138687134,
"learning_rate": 0.00013246994692046836,
"loss": 0.446,
"step": 500
},
{
"epoch": 1.3679180887372013,
"grad_norm": 0.1636335849761963,
"learning_rate": 0.00013216903707443967,
"loss": 0.4614,
"step": 501
},
{
"epoch": 1.3706484641638226,
"grad_norm": 0.1447010040283203,
"learning_rate": 0.00013186780197373306,
"loss": 0.4573,
"step": 502
},
{
"epoch": 1.3733788395904436,
"grad_norm": 0.17758530378341675,
"learning_rate": 0.0001315662446640761,
"loss": 0.4701,
"step": 503
},
{
"epoch": 1.3761092150170648,
"grad_norm": 0.1493985950946808,
"learning_rate": 0.00013126436819445422,
"loss": 0.4671,
"step": 504
},
{
"epoch": 1.378839590443686,
"grad_norm": 0.16412951052188873,
"learning_rate": 0.0001309621756170799,
"loss": 0.4705,
"step": 505
},
{
"epoch": 1.3815699658703071,
"grad_norm": 0.14819127321243286,
"learning_rate": 0.00013065966998736155,
"loss": 0.4579,
"step": 506
},
{
"epoch": 1.3843003412969284,
"grad_norm": 0.1500328630208969,
"learning_rate": 0.00013035685436387298,
"loss": 0.4484,
"step": 507
},
{
"epoch": 1.3870307167235496,
"grad_norm": 0.14388103783130646,
"learning_rate": 0.0001300537318083221,
"loss": 0.4325,
"step": 508
},
{
"epoch": 1.3897610921501706,
"grad_norm": 0.17138421535491943,
"learning_rate": 0.00012975030538552032,
"loss": 0.4717,
"step": 509
},
{
"epoch": 1.3924914675767919,
"grad_norm": 0.15119241178035736,
"learning_rate": 0.00012944657816335123,
"loss": 0.4279,
"step": 510
},
{
"epoch": 1.395221843003413,
"grad_norm": 0.15165849030017853,
"learning_rate": 0.00012914255321273986,
"loss": 0.4716,
"step": 511
},
{
"epoch": 1.3979522184300341,
"grad_norm": 0.14222781360149384,
"learning_rate": 0.0001288382336076215,
"loss": 0.4393,
"step": 512
},
{
"epoch": 1.4006825938566552,
"grad_norm": 0.1435043066740036,
"learning_rate": 0.00012853362242491053,
"loss": 0.4589,
"step": 513
},
{
"epoch": 1.4034129692832764,
"grad_norm": 0.15017302334308624,
"learning_rate": 0.00012822872274446958,
"loss": 0.4492,
"step": 514
},
{
"epoch": 1.4061433447098977,
"grad_norm": 0.14196786284446716,
"learning_rate": 0.00012792353764907804,
"loss": 0.4588,
"step": 515
},
{
"epoch": 1.4088737201365187,
"grad_norm": 0.144223153591156,
"learning_rate": 0.0001276180702244012,
"loss": 0.4678,
"step": 516
},
{
"epoch": 1.41160409556314,
"grad_norm": 0.12913116812705994,
"learning_rate": 0.0001273123235589589,
"loss": 0.4515,
"step": 517
},
{
"epoch": 1.4143344709897612,
"grad_norm": 0.1370343565940857,
"learning_rate": 0.00012700630074409427,
"loss": 0.444,
"step": 518
},
{
"epoch": 1.4170648464163822,
"grad_norm": 0.15820138156414032,
"learning_rate": 0.00012670000487394266,
"loss": 0.4693,
"step": 519
},
{
"epoch": 1.4197952218430034,
"grad_norm": 0.13780242204666138,
"learning_rate": 0.0001263934390454001,
"loss": 0.4621,
"step": 520
},
{
"epoch": 1.4225255972696247,
"grad_norm": 0.13959269225597382,
"learning_rate": 0.00012608660635809207,
"loss": 0.448,
"step": 521
},
{
"epoch": 1.4252559726962457,
"grad_norm": 0.14923076331615448,
"learning_rate": 0.00012577950991434248,
"loss": 0.4694,
"step": 522
},
{
"epoch": 1.427986348122867,
"grad_norm": 0.14546504616737366,
"learning_rate": 0.00012547215281914168,
"loss": 0.4679,
"step": 523
},
{
"epoch": 1.430716723549488,
"grad_norm": 0.13985708355903625,
"learning_rate": 0.00012516453818011566,
"loss": 0.469,
"step": 524
},
{
"epoch": 1.4334470989761092,
"grad_norm": 0.15217062830924988,
"learning_rate": 0.00012485666910749428,
"loss": 0.4698,
"step": 525
},
{
"epoch": 1.4361774744027302,
"grad_norm": 0.1293148696422577,
"learning_rate": 0.00012454854871407994,
"loss": 0.4448,
"step": 526
},
{
"epoch": 1.4389078498293515,
"grad_norm": 0.16416317224502563,
"learning_rate": 0.0001242401801152161,
"loss": 0.4589,
"step": 527
},
{
"epoch": 1.4416382252559727,
"grad_norm": 0.14671452343463898,
"learning_rate": 0.0001239315664287558,
"loss": 0.4498,
"step": 528
},
{
"epoch": 1.4443686006825938,
"grad_norm": 0.17271259427070618,
"learning_rate": 0.00012362271077503008,
"loss": 0.4492,
"step": 529
},
{
"epoch": 1.447098976109215,
"grad_norm": 0.15541909635066986,
"learning_rate": 0.00012331361627681645,
"loss": 0.4281,
"step": 530
},
{
"epoch": 1.4498293515358363,
"grad_norm": 0.14913444221019745,
"learning_rate": 0.00012300428605930736,
"loss": 0.4334,
"step": 531
},
{
"epoch": 1.4525597269624573,
"grad_norm": 0.167875275015831,
"learning_rate": 0.00012269472325007858,
"loss": 0.4736,
"step": 532
},
{
"epoch": 1.4552901023890785,
"grad_norm": 0.15073426067829132,
"learning_rate": 0.00012238493097905756,
"loss": 0.4744,
"step": 533
},
{
"epoch": 1.4580204778156998,
"grad_norm": 0.16593150794506073,
"learning_rate": 0.00012207491237849172,
"loss": 0.455,
"step": 534
},
{
"epoch": 1.4607508532423208,
"grad_norm": 0.13389617204666138,
"learning_rate": 0.00012176467058291699,
"loss": 0.4394,
"step": 535
},
{
"epoch": 1.463481228668942,
"grad_norm": 0.16609057784080505,
"learning_rate": 0.00012145420872912585,
"loss": 0.4562,
"step": 536
},
{
"epoch": 1.466211604095563,
"grad_norm": 0.1394118070602417,
"learning_rate": 0.00012114352995613582,
"loss": 0.4451,
"step": 537
},
{
"epoch": 1.4689419795221843,
"grad_norm": 0.17529746890068054,
"learning_rate": 0.00012083263740515765,
"loss": 0.4522,
"step": 538
},
{
"epoch": 1.4716723549488053,
"grad_norm": 0.15724115073680878,
"learning_rate": 0.00012052153421956342,
"loss": 0.4556,
"step": 539
},
{
"epoch": 1.4744027303754266,
"grad_norm": 0.17186792194843292,
"learning_rate": 0.00012021022354485514,
"loss": 0.4546,
"step": 540
},
{
"epoch": 1.4771331058020478,
"grad_norm": 0.1301499456167221,
"learning_rate": 0.00011989870852863254,
"loss": 0.4431,
"step": 541
},
{
"epoch": 1.4798634812286688,
"grad_norm": 0.1480223387479782,
"learning_rate": 0.00011958699232056134,
"loss": 0.4491,
"step": 542
},
{
"epoch": 1.48259385665529,
"grad_norm": 0.13150086998939514,
"learning_rate": 0.00011927507807234168,
"loss": 0.4568,
"step": 543
},
{
"epoch": 1.4853242320819113,
"grad_norm": 0.15769197046756744,
"learning_rate": 0.00011896296893767587,
"loss": 0.4501,
"step": 544
},
{
"epoch": 1.4880546075085324,
"grad_norm": 0.13996848464012146,
"learning_rate": 0.0001186506680722367,
"loss": 0.4673,
"step": 545
},
{
"epoch": 1.4907849829351536,
"grad_norm": 0.16406555473804474,
"learning_rate": 0.00011833817863363564,
"loss": 0.4634,
"step": 546
},
{
"epoch": 1.4935153583617748,
"grad_norm": 0.14007951319217682,
"learning_rate": 0.0001180255037813906,
"loss": 0.4466,
"step": 547
},
{
"epoch": 1.4962457337883959,
"grad_norm": 0.15525664389133453,
"learning_rate": 0.00011771264667689427,
"loss": 0.4323,
"step": 548
},
{
"epoch": 1.4989761092150171,
"grad_norm": 0.143234983086586,
"learning_rate": 0.00011739961048338213,
"loss": 0.4395,
"step": 549
},
{
"epoch": 1.5017064846416384,
"grad_norm": 0.15597446262836456,
"learning_rate": 0.00011708639836590023,
"loss": 0.4546,
"step": 550
},
{
"epoch": 1.5044368600682594,
"grad_norm": 0.13816912472248077,
"learning_rate": 0.00011677301349127348,
"loss": 0.4586,
"step": 551
},
{
"epoch": 1.5071672354948804,
"grad_norm": 0.16299140453338623,
"learning_rate": 0.00011645945902807341,
"loss": 0.4465,
"step": 552
},
{
"epoch": 1.5098976109215017,
"grad_norm": 0.15032370388507843,
"learning_rate": 0.00011614573814658629,
"loss": 0.4579,
"step": 553
},
{
"epoch": 1.512627986348123,
"grad_norm": 0.158245250582695,
"learning_rate": 0.00011583185401878101,
"loss": 0.4462,
"step": 554
},
{
"epoch": 1.515358361774744,
"grad_norm": 0.16943717002868652,
"learning_rate": 0.00011551780981827698,
"loss": 0.4572,
"step": 555
},
{
"epoch": 1.5180887372013652,
"grad_norm": 0.14559145271778107,
"learning_rate": 0.00011520360872031209,
"loss": 0.4693,
"step": 556
},
{
"epoch": 1.5208191126279864,
"grad_norm": 0.16138285398483276,
"learning_rate": 0.00011488925390171059,
"loss": 0.4623,
"step": 557
},
{
"epoch": 1.5235494880546074,
"grad_norm": 0.14859908819198608,
"learning_rate": 0.00011457474854085096,
"loss": 0.4684,
"step": 558
},
{
"epoch": 1.5262798634812287,
"grad_norm": 0.15695518255233765,
"learning_rate": 0.00011426009581763377,
"loss": 0.4619,
"step": 559
},
{
"epoch": 1.52901023890785,
"grad_norm": 0.14711041748523712,
"learning_rate": 0.00011394529891344958,
"loss": 0.4556,
"step": 560
},
{
"epoch": 1.531740614334471,
"grad_norm": 0.15271785855293274,
"learning_rate": 0.0001136303610111467,
"loss": 0.4557,
"step": 561
},
{
"epoch": 1.5344709897610922,
"grad_norm": 0.1541603058576584,
"learning_rate": 0.00011331528529499909,
"loss": 0.4644,
"step": 562
},
{
"epoch": 1.5372013651877134,
"grad_norm": 0.1686154007911682,
"learning_rate": 0.00011300007495067401,
"loss": 0.4446,
"step": 563
},
{
"epoch": 1.5399317406143345,
"grad_norm": 0.14042454957962036,
"learning_rate": 0.00011268473316520007,
"loss": 0.4275,
"step": 564
},
{
"epoch": 1.5426621160409555,
"grad_norm": 0.14575007557868958,
"learning_rate": 0.00011236926312693479,
"loss": 0.4579,
"step": 565
},
{
"epoch": 1.545392491467577,
"grad_norm": 0.16124123334884644,
"learning_rate": 0.0001120536680255323,
"loss": 0.441,
"step": 566
},
{
"epoch": 1.548122866894198,
"grad_norm": 0.14471302926540375,
"learning_rate": 0.00011173795105191145,
"loss": 0.4543,
"step": 567
},
{
"epoch": 1.550853242320819,
"grad_norm": 0.162650004029274,
"learning_rate": 0.00011142211539822318,
"loss": 0.4483,
"step": 568
},
{
"epoch": 1.5535836177474402,
"grad_norm": 0.14518044888973236,
"learning_rate": 0.00011110616425781833,
"loss": 0.4386,
"step": 569
},
{
"epoch": 1.5563139931740615,
"grad_norm": 0.1545732617378235,
"learning_rate": 0.00011079010082521557,
"loss": 0.4505,
"step": 570
},
{
"epoch": 1.5590443686006825,
"grad_norm": 0.12941716611385345,
"learning_rate": 0.00011047392829606876,
"loss": 0.4288,
"step": 571
},
{
"epoch": 1.5617747440273038,
"grad_norm": 0.15107029676437378,
"learning_rate": 0.0001101576498671349,
"loss": 0.4275,
"step": 572
},
{
"epoch": 1.564505119453925,
"grad_norm": 0.13776972889900208,
"learning_rate": 0.00010984126873624179,
"loss": 0.432,
"step": 573
},
{
"epoch": 1.567235494880546,
"grad_norm": 0.15370745956897736,
"learning_rate": 0.00010952478810225548,
"loss": 0.4523,
"step": 574
},
{
"epoch": 1.5699658703071673,
"grad_norm": 0.14505314826965332,
"learning_rate": 0.00010920821116504816,
"loss": 0.4444,
"step": 575
},
{
"epoch": 1.5726962457337885,
"grad_norm": 0.13053090870380402,
"learning_rate": 0.0001088915411254657,
"loss": 0.4492,
"step": 576
},
{
"epoch": 1.5754266211604095,
"grad_norm": 0.15613074600696564,
"learning_rate": 0.00010857478118529533,
"loss": 0.465,
"step": 577
},
{
"epoch": 1.5781569965870306,
"grad_norm": 0.164808988571167,
"learning_rate": 0.00010825793454723325,
"loss": 0.4498,
"step": 578
},
{
"epoch": 1.580887372013652,
"grad_norm": 0.14633600413799286,
"learning_rate": 0.0001079410044148522,
"loss": 0.4416,
"step": 579
},
{
"epoch": 1.583617747440273,
"grad_norm": 0.15268942713737488,
"learning_rate": 0.00010762399399256917,
"loss": 0.4431,
"step": 580
},
{
"epoch": 1.586348122866894,
"grad_norm": 0.16290055215358734,
"learning_rate": 0.00010730690648561292,
"loss": 0.465,
"step": 581
},
{
"epoch": 1.5890784982935153,
"grad_norm": 0.13567085564136505,
"learning_rate": 0.00010698974509999158,
"loss": 0.4397,
"step": 582
},
{
"epoch": 1.5918088737201366,
"grad_norm": 0.14896200597286224,
"learning_rate": 0.00010667251304246029,
"loss": 0.4458,
"step": 583
},
{
"epoch": 1.5945392491467576,
"grad_norm": 0.14076146483421326,
"learning_rate": 0.00010635521352048872,
"loss": 0.4476,
"step": 584
},
{
"epoch": 1.5972696245733788,
"grad_norm": 0.13334687054157257,
"learning_rate": 0.00010603784974222861,
"loss": 0.4283,
"step": 585
},
{
"epoch": 1.6,
"grad_norm": 0.16997142136096954,
"learning_rate": 0.00010572042491648149,
"loss": 0.4472,
"step": 586
},
{
"epoch": 1.6027303754266211,
"grad_norm": 0.15508471429347992,
"learning_rate": 0.00010540294225266607,
"loss": 0.4518,
"step": 587
},
{
"epoch": 1.6054607508532424,
"grad_norm": 0.1571933478116989,
"learning_rate": 0.0001050854049607858,
"loss": 0.4493,
"step": 588
},
{
"epoch": 1.6081911262798636,
"grad_norm": 0.15027360618114471,
"learning_rate": 0.00010476781625139656,
"loss": 0.4562,
"step": 589
},
{
"epoch": 1.6109215017064846,
"grad_norm": 0.14502452313899994,
"learning_rate": 0.00010445017933557404,
"loss": 0.4476,
"step": 590
},
{
"epoch": 1.6136518771331056,
"grad_norm": 0.1403171718120575,
"learning_rate": 0.00010413249742488131,
"loss": 0.4154,
"step": 591
},
{
"epoch": 1.6163822525597271,
"grad_norm": 0.14815428853034973,
"learning_rate": 0.00010381477373133652,
"loss": 0.4501,
"step": 592
},
{
"epoch": 1.6191126279863481,
"grad_norm": 0.16107513010501862,
"learning_rate": 0.00010349701146738007,
"loss": 0.4507,
"step": 593
},
{
"epoch": 1.6218430034129692,
"grad_norm": 0.14111128449440002,
"learning_rate": 0.00010317921384584244,
"loss": 0.4483,
"step": 594
},
{
"epoch": 1.6245733788395904,
"grad_norm": 0.1680098921060562,
"learning_rate": 0.0001028613840799117,
"loss": 0.4584,
"step": 595
},
{
"epoch": 1.6273037542662117,
"grad_norm": 0.13612088561058044,
"learning_rate": 0.00010254352538310075,
"loss": 0.4389,
"step": 596
},
{
"epoch": 1.6300341296928327,
"grad_norm": 0.16298632323741913,
"learning_rate": 0.00010222564096921505,
"loss": 0.452,
"step": 597
},
{
"epoch": 1.632764505119454,
"grad_norm": 0.14744146168231964,
"learning_rate": 0.00010190773405232024,
"loss": 0.4588,
"step": 598
},
{
"epoch": 1.6354948805460752,
"grad_norm": 0.1594633013010025,
"learning_rate": 0.00010158980784670927,
"loss": 0.4403,
"step": 599
},
{
"epoch": 1.6382252559726962,
"grad_norm": 0.15123943984508514,
"learning_rate": 0.00010127186556687019,
"loss": 0.4395,
"step": 600
},
{
"epoch": 1.6409556313993174,
"grad_norm": 0.1518649309873581,
"learning_rate": 0.00010095391042745361,
"loss": 0.4265,
"step": 601
},
{
"epoch": 1.6436860068259387,
"grad_norm": 0.1611323356628418,
"learning_rate": 0.00010063594564324012,
"loss": 0.444,
"step": 602
},
{
"epoch": 1.6464163822525597,
"grad_norm": 0.14762264490127563,
"learning_rate": 0.00010031797442910789,
"loss": 0.4409,
"step": 603
},
{
"epoch": 1.6491467576791807,
"grad_norm": 0.1546734869480133,
"learning_rate": 0.0001,
"loss": 0.4472,
"step": 604
},
{
"epoch": 1.6518771331058022,
"grad_norm": 0.1580485701560974,
"learning_rate": 9.968202557089212e-05,
"loss": 0.4497,
"step": 605
},
{
"epoch": 1.6546075085324232,
"grad_norm": 0.151153564453125,
"learning_rate": 9.93640543567599e-05,
"loss": 0.4495,
"step": 606
},
{
"epoch": 1.6573378839590442,
"grad_norm": 0.14238281548023224,
"learning_rate": 9.904608957254642e-05,
"loss": 0.4481,
"step": 607
},
{
"epoch": 1.6600682593856655,
"grad_norm": 0.13984693586826324,
"learning_rate": 9.872813443312984e-05,
"loss": 0.4262,
"step": 608
},
{
"epoch": 1.6627986348122867,
"grad_norm": 0.14657770097255707,
"learning_rate": 9.84101921532908e-05,
"loss": 0.4554,
"step": 609
},
{
"epoch": 1.6655290102389078,
"grad_norm": 0.15702606737613678,
"learning_rate": 9.809226594767978e-05,
"loss": 0.4553,
"step": 610
},
{
"epoch": 1.668259385665529,
"grad_norm": 0.15596133470535278,
"learning_rate": 9.777435903078494e-05,
"loss": 0.4353,
"step": 611
},
{
"epoch": 1.6709897610921502,
"grad_norm": 0.16193975508213043,
"learning_rate": 9.745647461689931e-05,
"loss": 0.4313,
"step": 612
},
{
"epoch": 1.6737201365187713,
"grad_norm": 0.15690681338310242,
"learning_rate": 9.713861592008833e-05,
"loss": 0.4498,
"step": 613
},
{
"epoch": 1.6764505119453925,
"grad_norm": 0.15277935564517975,
"learning_rate": 9.682078615415754e-05,
"loss": 0.4358,
"step": 614
},
{
"epoch": 1.6791808873720138,
"grad_norm": 0.1447397917509079,
"learning_rate": 9.650298853261997e-05,
"loss": 0.4264,
"step": 615
},
{
"epoch": 1.6819112627986348,
"grad_norm": 0.14292937517166138,
"learning_rate": 9.61852262686635e-05,
"loss": 0.4521,
"step": 616
},
{
"epoch": 1.6846416382252558,
"grad_norm": 0.15041732788085938,
"learning_rate": 9.586750257511867e-05,
"loss": 0.445,
"step": 617
},
{
"epoch": 1.6873720136518773,
"grad_norm": 0.14610610902309418,
"learning_rate": 9.5549820664426e-05,
"loss": 0.4493,
"step": 618
},
{
"epoch": 1.6901023890784983,
"grad_norm": 0.1494503617286682,
"learning_rate": 9.523218374860348e-05,
"loss": 0.4473,
"step": 619
},
{
"epoch": 1.6928327645051193,
"grad_norm": 0.14970283210277557,
"learning_rate": 9.491459503921421e-05,
"loss": 0.453,
"step": 620
},
{
"epoch": 1.6955631399317406,
"grad_norm": 0.15402431786060333,
"learning_rate": 9.459705774733396e-05,
"loss": 0.4435,
"step": 621
},
{
"epoch": 1.6982935153583618,
"grad_norm": 0.1583550125360489,
"learning_rate": 9.427957508351852e-05,
"loss": 0.4379,
"step": 622
},
{
"epoch": 1.7010238907849828,
"grad_norm": 0.15820656716823578,
"learning_rate": 9.396215025777139e-05,
"loss": 0.4416,
"step": 623
},
{
"epoch": 1.703754266211604,
"grad_norm": 0.14503423869609833,
"learning_rate": 9.364478647951133e-05,
"loss": 0.434,
"step": 624
},
{
"epoch": 1.7064846416382253,
"grad_norm": 0.1509648561477661,
"learning_rate": 9.332748695753973e-05,
"loss": 0.4516,
"step": 625
},
{
"epoch": 1.7092150170648464,
"grad_norm": 0.15958918631076813,
"learning_rate": 9.301025490000841e-05,
"loss": 0.4504,
"step": 626
},
{
"epoch": 1.7119453924914676,
"grad_norm": 0.1522430032491684,
"learning_rate": 9.269309351438711e-05,
"loss": 0.4339,
"step": 627
},
{
"epoch": 1.7146757679180888,
"grad_norm": 0.14348183572292328,
"learning_rate": 9.237600600743085e-05,
"loss": 0.4236,
"step": 628
},
{
"epoch": 1.7174061433447099,
"grad_norm": 0.1558932512998581,
"learning_rate": 9.20589955851478e-05,
"loss": 0.4514,
"step": 629
},
{
"epoch": 1.7201365187713311,
"grad_norm": 0.13524165749549866,
"learning_rate": 9.174206545276677e-05,
"loss": 0.4426,
"step": 630
},
{
"epoch": 1.7228668941979524,
"grad_norm": 0.1464926302433014,
"learning_rate": 9.142521881470469e-05,
"loss": 0.4397,
"step": 631
},
{
"epoch": 1.7255972696245734,
"grad_norm": 0.14485323429107666,
"learning_rate": 9.11084588745343e-05,
"loss": 0.4497,
"step": 632
},
{
"epoch": 1.7283276450511944,
"grad_norm": 0.139760822057724,
"learning_rate": 9.07917888349519e-05,
"loss": 0.4465,
"step": 633
},
{
"epoch": 1.7310580204778157,
"grad_norm": 0.16193385422229767,
"learning_rate": 9.047521189774455e-05,
"loss": 0.4377,
"step": 634
},
{
"epoch": 1.733788395904437,
"grad_norm": 0.17076647281646729,
"learning_rate": 9.015873126375822e-05,
"loss": 0.4632,
"step": 635
},
{
"epoch": 1.736518771331058,
"grad_norm": 0.14881980419158936,
"learning_rate": 8.984235013286511e-05,
"loss": 0.4401,
"step": 636
},
{
"epoch": 1.7392491467576792,
"grad_norm": 0.15158087015151978,
"learning_rate": 8.952607170393125e-05,
"loss": 0.4292,
"step": 637
},
{
"epoch": 1.7419795221843004,
"grad_norm": 0.14353424310684204,
"learning_rate": 8.920989917478447e-05,
"loss": 0.4439,
"step": 638
},
{
"epoch": 1.7447098976109214,
"grad_norm": 0.13718026876449585,
"learning_rate": 8.88938357421817e-05,
"loss": 0.4267,
"step": 639
},
{
"epoch": 1.7474402730375427,
"grad_norm": 0.15788930654525757,
"learning_rate": 8.857788460177686e-05,
"loss": 0.4241,
"step": 640
},
{
"epoch": 1.750170648464164,
"grad_norm": 0.14324265718460083,
"learning_rate": 8.826204894808855e-05,
"loss": 0.4472,
"step": 641
},
{
"epoch": 1.752901023890785,
"grad_norm": 0.15020030736923218,
"learning_rate": 8.79463319744677e-05,
"loss": 0.4339,
"step": 642
},
{
"epoch": 1.7556313993174062,
"grad_norm": 0.13715054094791412,
"learning_rate": 8.763073687306524e-05,
"loss": 0.4429,
"step": 643
},
{
"epoch": 1.7583617747440274,
"grad_norm": 0.16862636804580688,
"learning_rate": 8.731526683479992e-05,
"loss": 0.4509,
"step": 644
},
{
"epoch": 1.7610921501706485,
"grad_norm": 0.14781633019447327,
"learning_rate": 8.6999925049326e-05,
"loss": 0.4319,
"step": 645
},
{
"epoch": 1.7638225255972695,
"grad_norm": 0.15873977541923523,
"learning_rate": 8.668471470500095e-05,
"loss": 0.4486,
"step": 646
},
{
"epoch": 1.7665529010238907,
"grad_norm": 0.1477411389350891,
"learning_rate": 8.63696389888533e-05,
"loss": 0.4461,
"step": 647
},
{
"epoch": 1.769283276450512,
"grad_norm": 0.1804722100496292,
"learning_rate": 8.605470108655045e-05,
"loss": 0.4402,
"step": 648
},
{
"epoch": 1.772013651877133,
"grad_norm": 0.15354932844638824,
"learning_rate": 8.573990418236625e-05,
"loss": 0.4564,
"step": 649
},
{
"epoch": 1.7747440273037542,
"grad_norm": 0.1494126319885254,
"learning_rate": 8.542525145914905e-05,
"loss": 0.4377,
"step": 650
},
{
"epoch": 1.7774744027303755,
"grad_norm": 0.14122453331947327,
"learning_rate": 8.511074609828944e-05,
"loss": 0.4361,
"step": 651
},
{
"epoch": 1.7802047781569965,
"grad_norm": 0.16938751935958862,
"learning_rate": 8.479639127968792e-05,
"loss": 0.4593,
"step": 652
},
{
"epoch": 1.7829351535836178,
"grad_norm": 0.1502314805984497,
"learning_rate": 8.448219018172303e-05,
"loss": 0.4336,
"step": 653
},
{
"epoch": 1.785665529010239,
"grad_norm": 0.180609330534935,
"learning_rate": 8.4168145981219e-05,
"loss": 0.45,
"step": 654
},
{
"epoch": 1.78839590443686,
"grad_norm": 0.1395808309316635,
"learning_rate": 8.385426185341374e-05,
"loss": 0.4188,
"step": 655
},
{
"epoch": 1.7911262798634813,
"grad_norm": 0.1596853882074356,
"learning_rate": 8.35405409719266e-05,
"loss": 0.452,
"step": 656
},
{
"epoch": 1.7938566552901025,
"grad_norm": 0.15330305695533752,
"learning_rate": 8.322698650872656e-05,
"loss": 0.4355,
"step": 657
},
{
"epoch": 1.7965870307167235,
"grad_norm": 0.1545482873916626,
"learning_rate": 8.291360163409978e-05,
"loss": 0.4366,
"step": 658
},
{
"epoch": 1.7993174061433446,
"grad_norm": 0.13950030505657196,
"learning_rate": 8.260038951661787e-05,
"loss": 0.4169,
"step": 659
},
{
"epoch": 1.802047781569966,
"grad_norm": 0.14191307127475739,
"learning_rate": 8.228735332310575e-05,
"loss": 0.4471,
"step": 660
},
{
"epoch": 1.804778156996587,
"grad_norm": 0.14557993412017822,
"learning_rate": 8.197449621860943e-05,
"loss": 0.4028,
"step": 661
},
{
"epoch": 1.807508532423208,
"grad_norm": 0.13985979557037354,
"learning_rate": 8.16618213663644e-05,
"loss": 0.4293,
"step": 662
},
{
"epoch": 1.8102389078498293,
"grad_norm": 0.1420183628797531,
"learning_rate": 8.134933192776333e-05,
"loss": 0.4313,
"step": 663
},
{
"epoch": 1.8129692832764506,
"grad_norm": 0.1498919129371643,
"learning_rate": 8.103703106232416e-05,
"loss": 0.4315,
"step": 664
},
{
"epoch": 1.8156996587030716,
"grad_norm": 0.1570868194103241,
"learning_rate": 8.072492192765833e-05,
"loss": 0.4348,
"step": 665
},
{
"epoch": 1.8184300341296928,
"grad_norm": 0.15423277020454407,
"learning_rate": 8.041300767943867e-05,
"loss": 0.44,
"step": 666
},
{
"epoch": 1.821160409556314,
"grad_norm": 0.13805197179317474,
"learning_rate": 8.010129147136749e-05,
"loss": 0.4317,
"step": 667
},
{
"epoch": 1.823890784982935,
"grad_norm": 0.14628642797470093,
"learning_rate": 7.978977645514487e-05,
"loss": 0.4379,
"step": 668
},
{
"epoch": 1.8266211604095564,
"grad_norm": 0.15710268914699554,
"learning_rate": 7.947846578043659e-05,
"loss": 0.4557,
"step": 669
},
{
"epoch": 1.8293515358361776,
"grad_norm": 0.15317128598690033,
"learning_rate": 7.916736259484239e-05,
"loss": 0.4506,
"step": 670
},
{
"epoch": 1.8320819112627986,
"grad_norm": 0.15919502079486847,
"learning_rate": 7.88564700438642e-05,
"loss": 0.435,
"step": 671
},
{
"epoch": 1.8348122866894196,
"grad_norm": 0.1551041305065155,
"learning_rate": 7.854579127087417e-05,
"loss": 0.4283,
"step": 672
},
{
"epoch": 1.8375426621160411,
"grad_norm": 0.16782739758491516,
"learning_rate": 7.823532941708303e-05,
"loss": 0.4507,
"step": 673
},
{
"epoch": 1.8402730375426621,
"grad_norm": 0.15251149237155914,
"learning_rate": 7.792508762150833e-05,
"loss": 0.4255,
"step": 674
},
{
"epoch": 1.8430034129692832,
"grad_norm": 0.17279517650604248,
"learning_rate": 7.761506902094248e-05,
"loss": 0.4255,
"step": 675
},
{
"epoch": 1.8457337883959044,
"grad_norm": 0.13385094702243805,
"learning_rate": 7.730527674992143e-05,
"loss": 0.4229,
"step": 676
},
{
"epoch": 1.8484641638225257,
"grad_norm": 0.15888231992721558,
"learning_rate": 7.699571394069269e-05,
"loss": 0.4335,
"step": 677
},
{
"epoch": 1.8511945392491467,
"grad_norm": 0.1549587994813919,
"learning_rate": 7.668638372318359e-05,
"loss": 0.4083,
"step": 678
},
{
"epoch": 1.853924914675768,
"grad_norm": 0.16143332421779633,
"learning_rate": 7.637728922496996e-05,
"loss": 0.4367,
"step": 679
},
{
"epoch": 1.8566552901023892,
"grad_norm": 0.13735996186733246,
"learning_rate": 7.606843357124426e-05,
"loss": 0.4296,
"step": 680
},
{
"epoch": 1.8593856655290102,
"grad_norm": 0.14317500591278076,
"learning_rate": 7.575981988478392e-05,
"loss": 0.4419,
"step": 681
},
{
"epoch": 1.8621160409556314,
"grad_norm": 0.14451129734516144,
"learning_rate": 7.54514512859201e-05,
"loss": 0.4389,
"step": 682
},
{
"epoch": 1.8648464163822527,
"grad_norm": 0.14233650267124176,
"learning_rate": 7.514333089250577e-05,
"loss": 0.4258,
"step": 683
},
{
"epoch": 1.8675767918088737,
"grad_norm": 0.14275044202804565,
"learning_rate": 7.483546181988436e-05,
"loss": 0.4456,
"step": 684
},
{
"epoch": 1.8703071672354947,
"grad_norm": 0.15278606116771698,
"learning_rate": 7.452784718085833e-05,
"loss": 0.4344,
"step": 685
},
{
"epoch": 1.8730375426621162,
"grad_norm": 0.13797658681869507,
"learning_rate": 7.422049008565757e-05,
"loss": 0.4392,
"step": 686
},
{
"epoch": 1.8757679180887372,
"grad_norm": 0.1518598347902298,
"learning_rate": 7.391339364190794e-05,
"loss": 0.431,
"step": 687
},
{
"epoch": 1.8784982935153582,
"grad_norm": 0.1477964073419571,
"learning_rate": 7.360656095459995e-05,
"loss": 0.4449,
"step": 688
},
{
"epoch": 1.8812286689419795,
"grad_norm": 0.14295299351215363,
"learning_rate": 7.329999512605738e-05,
"loss": 0.4384,
"step": 689
},
{
"epoch": 1.8839590443686007,
"grad_norm": 0.1487056165933609,
"learning_rate": 7.299369925590574e-05,
"loss": 0.4311,
"step": 690
},
{
"epoch": 1.8866894197952218,
"grad_norm": 0.14108945429325104,
"learning_rate": 7.268767644104112e-05,
"loss": 0.4239,
"step": 691
},
{
"epoch": 1.889419795221843,
"grad_norm": 0.15047363936901093,
"learning_rate": 7.238192977559884e-05,
"loss": 0.4427,
"step": 692
},
{
"epoch": 1.8921501706484642,
"grad_norm": 0.14634403586387634,
"learning_rate": 7.2076462350922e-05,
"loss": 0.4416,
"step": 693
},
{
"epoch": 1.8948805460750853,
"grad_norm": 0.15272392332553864,
"learning_rate": 7.177127725553045e-05,
"loss": 0.43,
"step": 694
},
{
"epoch": 1.8976109215017065,
"grad_norm": 0.15047992765903473,
"learning_rate": 7.146637757508949e-05,
"loss": 0.4436,
"step": 695
},
{
"epoch": 1.9003412969283278,
"grad_norm": 0.14413118362426758,
"learning_rate": 7.116176639237852e-05,
"loss": 0.4169,
"step": 696
},
{
"epoch": 1.9030716723549488,
"grad_norm": 0.1437167525291443,
"learning_rate": 7.085744678726013e-05,
"loss": 0.4389,
"step": 697
},
{
"epoch": 1.9058020477815698,
"grad_norm": 0.14662359654903412,
"learning_rate": 7.05534218366488e-05,
"loss": 0.435,
"step": 698
},
{
"epoch": 1.9085324232081913,
"grad_norm": 0.14521794021129608,
"learning_rate": 7.024969461447972e-05,
"loss": 0.4505,
"step": 699
},
{
"epoch": 1.9112627986348123,
"grad_norm": 0.14155706763267517,
"learning_rate": 6.994626819167789e-05,
"loss": 0.4301,
"step": 700
},
{
"epoch": 1.9139931740614333,
"grad_norm": 0.14663158357143402,
"learning_rate": 6.964314563612708e-05,
"loss": 0.4262,
"step": 701
},
{
"epoch": 1.9167235494880546,
"grad_norm": 0.15034128725528717,
"learning_rate": 6.934033001263847e-05,
"loss": 0.4424,
"step": 702
},
{
"epoch": 1.9194539249146758,
"grad_norm": 0.15093255043029785,
"learning_rate": 6.903782438292015e-05,
"loss": 0.4425,
"step": 703
},
{
"epoch": 1.9221843003412968,
"grad_norm": 0.1556250900030136,
"learning_rate": 6.873563180554583e-05,
"loss": 0.4225,
"step": 704
},
{
"epoch": 1.924914675767918,
"grad_norm": 0.16173475980758667,
"learning_rate": 6.843375533592395e-05,
"loss": 0.4228,
"step": 705
},
{
"epoch": 1.9276450511945393,
"grad_norm": 0.1532420516014099,
"learning_rate": 6.813219802626698e-05,
"loss": 0.432,
"step": 706
},
{
"epoch": 1.9303754266211604,
"grad_norm": 0.1596469283103943,
"learning_rate": 6.783096292556035e-05,
"loss": 0.4232,
"step": 707
},
{
"epoch": 1.9331058020477816,
"grad_norm": 0.16538076102733612,
"learning_rate": 6.753005307953167e-05,
"loss": 0.45,
"step": 708
},
{
"epoch": 1.9358361774744028,
"grad_norm": 0.15187640488147736,
"learning_rate": 6.722947153062003e-05,
"loss": 0.4442,
"step": 709
},
{
"epoch": 1.9385665529010239,
"grad_norm": 0.14827731251716614,
"learning_rate": 6.692922131794517e-05,
"loss": 0.414,
"step": 710
},
{
"epoch": 1.9412969283276449,
"grad_norm": 0.16438645124435425,
"learning_rate": 6.662930547727668e-05,
"loss": 0.419,
"step": 711
},
{
"epoch": 1.9440273037542664,
"grad_norm": 0.15135832130908966,
"learning_rate": 6.632972704100349e-05,
"loss": 0.4155,
"step": 712
},
{
"epoch": 1.9467576791808874,
"grad_norm": 0.15094083547592163,
"learning_rate": 6.603048903810305e-05,
"loss": 0.4258,
"step": 713
},
{
"epoch": 1.9494880546075084,
"grad_norm": 0.14968033134937286,
"learning_rate": 6.57315944941107e-05,
"loss": 0.4395,
"step": 714
},
{
"epoch": 1.9522184300341296,
"grad_norm": 0.16519851982593536,
"learning_rate": 6.54330464310892e-05,
"loss": 0.4406,
"step": 715
},
{
"epoch": 1.954948805460751,
"grad_norm": 0.13855180144309998,
"learning_rate": 6.513484786759818e-05,
"loss": 0.43,
"step": 716
},
{
"epoch": 1.957679180887372,
"grad_norm": 0.1570328176021576,
"learning_rate": 6.483700181866337e-05,
"loss": 0.4288,
"step": 717
},
{
"epoch": 1.9604095563139932,
"grad_norm": 0.14928270876407623,
"learning_rate": 6.453951129574644e-05,
"loss": 0.4224,
"step": 718
},
{
"epoch": 1.9631399317406144,
"grad_norm": 0.16348999738693237,
"learning_rate": 6.42423793067144e-05,
"loss": 0.4379,
"step": 719
},
{
"epoch": 1.9658703071672354,
"grad_norm": 0.14947615563869476,
"learning_rate": 6.39456088558091e-05,
"loss": 0.445,
"step": 720
},
{
"epoch": 1.9686006825938567,
"grad_norm": 0.14767783880233765,
"learning_rate": 6.3649202943617e-05,
"loss": 0.4388,
"step": 721
},
{
"epoch": 1.971331058020478,
"grad_norm": 0.14223739504814148,
"learning_rate": 6.33531645670389e-05,
"loss": 0.433,
"step": 722
},
{
"epoch": 1.974061433447099,
"grad_norm": 0.1595824956893921,
"learning_rate": 6.305749671925931e-05,
"loss": 0.4325,
"step": 723
},
{
"epoch": 1.9767918088737202,
"grad_norm": 0.14971914887428284,
"learning_rate": 6.276220238971652e-05,
"loss": 0.4336,
"step": 724
},
{
"epoch": 1.9795221843003414,
"grad_norm": 0.15426860749721527,
"learning_rate": 6.24672845640723e-05,
"loss": 0.433,
"step": 725
},
{
"epoch": 1.9822525597269625,
"grad_norm": 0.1470557302236557,
"learning_rate": 6.217274622418153e-05,
"loss": 0.435,
"step": 726
},
{
"epoch": 1.9849829351535835,
"grad_norm": 0.1564924120903015,
"learning_rate": 6.187859034806224e-05,
"loss": 0.4371,
"step": 727
},
{
"epoch": 1.9877133105802047,
"grad_norm": 0.16420651972293854,
"learning_rate": 6.158481990986557e-05,
"loss": 0.4478,
"step": 728
},
{
"epoch": 1.990443686006826,
"grad_norm": 0.15158484876155853,
"learning_rate": 6.129143787984533e-05,
"loss": 0.4267,
"step": 729
},
{
"epoch": 1.993174061433447,
"grad_norm": 0.15700684487819672,
"learning_rate": 6.099844722432843e-05,
"loss": 0.4293,
"step": 730
},
{
"epoch": 1.9959044368600682,
"grad_norm": 0.14848262071609497,
"learning_rate": 6.070585090568459e-05,
"loss": 0.4339,
"step": 731
},
{
"epoch": 1.9986348122866895,
"grad_norm": 0.15596534311771393,
"learning_rate": 6.0413651882296406e-05,
"loss": 0.4245,
"step": 732
},
{
"epoch": 2.0013651877133105,
"grad_norm": 0.3338797092437744,
"learning_rate": 6.012185310852962e-05,
"loss": 0.7156,
"step": 733
},
{
"epoch": 2.0040955631399315,
"grad_norm": 0.1896699219942093,
"learning_rate": 5.983045753470308e-05,
"loss": 0.4228,
"step": 734
},
{
"epoch": 2.006825938566553,
"grad_norm": 0.20619529485702515,
"learning_rate": 5.953946810705888e-05,
"loss": 0.4244,
"step": 735
},
{
"epoch": 2.009556313993174,
"grad_norm": 0.17412033677101135,
"learning_rate": 5.924888776773281e-05,
"loss": 0.4186,
"step": 736
},
{
"epoch": 2.012286689419795,
"grad_norm": 0.1896408498287201,
"learning_rate": 5.8958719454724346e-05,
"loss": 0.4259,
"step": 737
},
{
"epoch": 2.0150170648464165,
"grad_norm": 0.16549214720726013,
"learning_rate": 5.8668966101867005e-05,
"loss": 0.3967,
"step": 738
},
{
"epoch": 2.0177474402730375,
"grad_norm": 0.16784432530403137,
"learning_rate": 5.837963063879884e-05,
"loss": 0.4347,
"step": 739
},
{
"epoch": 2.0204778156996586,
"grad_norm": 0.18551002442836761,
"learning_rate": 5.809071599093272e-05,
"loss": 0.4407,
"step": 740
},
{
"epoch": 2.02320819112628,
"grad_norm": 0.17308658361434937,
"learning_rate": 5.780222507942654e-05,
"loss": 0.4343,
"step": 741
},
{
"epoch": 2.025938566552901,
"grad_norm": 0.17182452976703644,
"learning_rate": 5.751416082115408e-05,
"loss": 0.422,
"step": 742
},
{
"epoch": 2.028668941979522,
"grad_norm": 0.17356833815574646,
"learning_rate": 5.722652612867523e-05,
"loss": 0.4168,
"step": 743
},
{
"epoch": 2.0313993174061435,
"grad_norm": 0.17314977943897247,
"learning_rate": 5.6939323910206645e-05,
"loss": 0.4285,
"step": 744
},
{
"epoch": 2.0341296928327646,
"grad_norm": 0.1868155151605606,
"learning_rate": 5.6652557069592304e-05,
"loss": 0.4211,
"step": 745
},
{
"epoch": 2.0368600682593856,
"grad_norm": 0.1524539738893509,
"learning_rate": 5.63662285062742e-05,
"loss": 0.4233,
"step": 746
},
{
"epoch": 2.0395904436860066,
"grad_norm": 0.17929619550704956,
"learning_rate": 5.608034111526298e-05,
"loss": 0.4305,
"step": 747
},
{
"epoch": 2.042320819112628,
"grad_norm": 0.1604132503271103,
"learning_rate": 5.579489778710867e-05,
"loss": 0.4214,
"step": 748
},
{
"epoch": 2.045051194539249,
"grad_norm": 0.16710160672664642,
"learning_rate": 5.550990140787147e-05,
"loss": 0.4369,
"step": 749
},
{
"epoch": 2.04778156996587,
"grad_norm": 0.16114738583564758,
"learning_rate": 5.522535485909257e-05,
"loss": 0.4216,
"step": 750
},
{
"epoch": 2.0505119453924916,
"grad_norm": 0.16698378324508667,
"learning_rate": 5.494126101776505e-05,
"loss": 0.4323,
"step": 751
},
{
"epoch": 2.0532423208191126,
"grad_norm": 0.174140065908432,
"learning_rate": 5.4657622756304704e-05,
"loss": 0.4135,
"step": 752
},
{
"epoch": 2.0559726962457336,
"grad_norm": 0.1496962457895279,
"learning_rate": 5.437444294252107e-05,
"loss": 0.438,
"step": 753
},
{
"epoch": 2.058703071672355,
"grad_norm": 0.1660911738872528,
"learning_rate": 5.409172443958843e-05,
"loss": 0.4262,
"step": 754
},
{
"epoch": 2.061433447098976,
"grad_norm": 0.18081265687942505,
"learning_rate": 5.380947010601681e-05,
"loss": 0.4172,
"step": 755
},
{
"epoch": 2.064163822525597,
"grad_norm": 0.14415475726127625,
"learning_rate": 5.3527682795623146e-05,
"loss": 0.4181,
"step": 756
},
{
"epoch": 2.0668941979522186,
"grad_norm": 0.20684713125228882,
"learning_rate": 5.324636535750238e-05,
"loss": 0.4291,
"step": 757
},
{
"epoch": 2.0696245733788396,
"grad_norm": 0.17085103690624237,
"learning_rate": 5.296552063599868e-05,
"loss": 0.4372,
"step": 758
},
{
"epoch": 2.0723549488054607,
"grad_norm": 0.17061397433280945,
"learning_rate": 5.2685151470676653e-05,
"loss": 0.4247,
"step": 759
},
{
"epoch": 2.0750853242320817,
"grad_norm": 0.17692053318023682,
"learning_rate": 5.240526069629265e-05,
"loss": 0.4261,
"step": 760
},
{
"epoch": 2.077815699658703,
"grad_norm": 0.15353117883205414,
"learning_rate": 5.212585114276614e-05,
"loss": 0.4272,
"step": 761
},
{
"epoch": 2.080546075085324,
"grad_norm": 0.17618128657341003,
"learning_rate": 5.1846925635151045e-05,
"loss": 0.4206,
"step": 762
},
{
"epoch": 2.083276450511945,
"grad_norm": 0.154897078871727,
"learning_rate": 5.156848699360719e-05,
"loss": 0.4086,
"step": 763
},
{
"epoch": 2.0860068259385667,
"grad_norm": 0.15779747068881989,
"learning_rate": 5.129053803337181e-05,
"loss": 0.4073,
"step": 764
},
{
"epoch": 2.0887372013651877,
"grad_norm": 0.186599463224411,
"learning_rate": 5.101308156473104e-05,
"loss": 0.4204,
"step": 765
},
{
"epoch": 2.0914675767918087,
"grad_norm": 0.15039357542991638,
"learning_rate": 5.073612039299157e-05,
"loss": 0.4277,
"step": 766
},
{
"epoch": 2.09419795221843,
"grad_norm": 0.16809961199760437,
"learning_rate": 5.0459657318452224e-05,
"loss": 0.4312,
"step": 767
},
{
"epoch": 2.096928327645051,
"grad_norm": 0.18085776269435883,
"learning_rate": 5.0183695136375664e-05,
"loss": 0.4233,
"step": 768
},
{
"epoch": 2.0996587030716722,
"grad_norm": 0.15629561245441437,
"learning_rate": 4.9908236636960126e-05,
"loss": 0.4252,
"step": 769
},
{
"epoch": 2.1023890784982937,
"grad_norm": 0.18560980260372162,
"learning_rate": 4.963328460531127e-05,
"loss": 0.4112,
"step": 770
},
{
"epoch": 2.1051194539249147,
"grad_norm": 0.16012516617774963,
"learning_rate": 4.935884182141377e-05,
"loss": 0.4108,
"step": 771
},
{
"epoch": 2.1078498293515358,
"grad_norm": 0.1524171382188797,
"learning_rate": 4.908491106010368e-05,
"loss": 0.4183,
"step": 772
},
{
"epoch": 2.1105802047781568,
"grad_norm": 0.173212930560112,
"learning_rate": 4.8811495091039926e-05,
"loss": 0.4058,
"step": 773
},
{
"epoch": 2.1133105802047782,
"grad_norm": 0.1540430635213852,
"learning_rate": 4.8538596678676406e-05,
"loss": 0.4016,
"step": 774
},
{
"epoch": 2.1160409556313993,
"grad_norm": 0.1477975845336914,
"learning_rate": 4.826621858223431e-05,
"loss": 0.3975,
"step": 775
},
{
"epoch": 2.1187713310580203,
"grad_norm": 0.17888864874839783,
"learning_rate": 4.79943635556739e-05,
"loss": 0.4208,
"step": 776
},
{
"epoch": 2.1215017064846418,
"grad_norm": 0.15199348330497742,
"learning_rate": 4.7723034347666696e-05,
"loss": 0.4304,
"step": 777
},
{
"epoch": 2.124232081911263,
"grad_norm": 0.15406261384487152,
"learning_rate": 4.745223370156797e-05,
"loss": 0.439,
"step": 778
},
{
"epoch": 2.126962457337884,
"grad_norm": 0.1821894645690918,
"learning_rate": 4.71819643553887e-05,
"loss": 0.4298,
"step": 779
},
{
"epoch": 2.1296928327645053,
"grad_norm": 0.15180355310440063,
"learning_rate": 4.691222904176791e-05,
"loss": 0.4136,
"step": 780
},
{
"epoch": 2.1324232081911263,
"grad_norm": 0.17140239477157593,
"learning_rate": 4.6643030487945326e-05,
"loss": 0.4292,
"step": 781
},
{
"epoch": 2.1351535836177473,
"grad_norm": 0.15253609418869019,
"learning_rate": 4.6374371415733496e-05,
"loss": 0.4393,
"step": 782
},
{
"epoch": 2.137883959044369,
"grad_norm": 0.15631216764450073,
"learning_rate": 4.6106254541490325e-05,
"loss": 0.4162,
"step": 783
},
{
"epoch": 2.14061433447099,
"grad_norm": 0.18031221628189087,
"learning_rate": 4.583868257609171e-05,
"loss": 0.415,
"step": 784
},
{
"epoch": 2.143344709897611,
"grad_norm": 0.15720027685165405,
"learning_rate": 4.55716582249042e-05,
"loss": 0.4288,
"step": 785
},
{
"epoch": 2.146075085324232,
"grad_norm": 0.15618009865283966,
"learning_rate": 4.530518418775733e-05,
"loss": 0.4238,
"step": 786
},
{
"epoch": 2.1488054607508533,
"grad_norm": 0.16991287469863892,
"learning_rate": 4.50392631589166e-05,
"loss": 0.4321,
"step": 787
},
{
"epoch": 2.1515358361774743,
"grad_norm": 0.16138002276420593,
"learning_rate": 4.477389782705628e-05,
"loss": 0.4172,
"step": 788
},
{
"epoch": 2.1542662116040954,
"grad_norm": 0.15910767018795013,
"learning_rate": 4.450909087523186e-05,
"loss": 0.4149,
"step": 789
},
{
"epoch": 2.156996587030717,
"grad_norm": 0.15495507419109344,
"learning_rate": 4.424484498085335e-05,
"loss": 0.4351,
"step": 790
},
{
"epoch": 2.159726962457338,
"grad_norm": 0.15659591555595398,
"learning_rate": 4.398116281565794e-05,
"loss": 0.4254,
"step": 791
},
{
"epoch": 2.162457337883959,
"grad_norm": 0.16147974133491516,
"learning_rate": 4.371804704568309e-05,
"loss": 0.4265,
"step": 792
},
{
"epoch": 2.1651877133105804,
"grad_norm": 0.14304347336292267,
"learning_rate": 4.345550033123954e-05,
"loss": 0.4211,
"step": 793
},
{
"epoch": 2.1679180887372014,
"grad_norm": 0.15479592978954315,
"learning_rate": 4.3193525326884435e-05,
"loss": 0.4002,
"step": 794
},
{
"epoch": 2.1706484641638224,
"grad_norm": 0.1610931009054184,
"learning_rate": 4.293212468139447e-05,
"loss": 0.4156,
"step": 795
},
{
"epoch": 2.173378839590444,
"grad_norm": 0.16558706760406494,
"learning_rate": 4.267130103773911e-05,
"loss": 0.4285,
"step": 796
},
{
"epoch": 2.176109215017065,
"grad_norm": 0.16455373167991638,
"learning_rate": 4.241105703305388e-05,
"loss": 0.4058,
"step": 797
},
{
"epoch": 2.178839590443686,
"grad_norm": 0.15886934101581573,
"learning_rate": 4.215139529861367e-05,
"loss": 0.4311,
"step": 798
},
{
"epoch": 2.181569965870307,
"grad_norm": 0.15277110040187836,
"learning_rate": 4.189231845980618e-05,
"loss": 0.4176,
"step": 799
},
{
"epoch": 2.1843003412969284,
"grad_norm": 0.15656784176826477,
"learning_rate": 4.163382913610533e-05,
"loss": 0.4205,
"step": 800
},
{
"epoch": 2.1870307167235494,
"grad_norm": 0.1564100980758667,
"learning_rate": 4.1375929941044786e-05,
"loss": 0.4313,
"step": 801
},
{
"epoch": 2.1897610921501705,
"grad_norm": 0.14734816551208496,
"learning_rate": 4.111862348219158e-05,
"loss": 0.4258,
"step": 802
},
{
"epoch": 2.192491467576792,
"grad_norm": 0.15179724991321564,
"learning_rate": 4.086191236111964e-05,
"loss": 0.4372,
"step": 803
},
{
"epoch": 2.195221843003413,
"grad_norm": 0.16015468537807465,
"learning_rate": 4.060579917338362e-05,
"loss": 0.4104,
"step": 804
},
{
"epoch": 2.197952218430034,
"grad_norm": 0.1574854701757431,
"learning_rate": 4.0350286508492554e-05,
"loss": 0.4068,
"step": 805
},
{
"epoch": 2.2006825938566554,
"grad_norm": 0.14426739513874054,
"learning_rate": 4.009537694988372e-05,
"loss": 0.3904,
"step": 806
},
{
"epoch": 2.2034129692832765,
"grad_norm": 0.14949829876422882,
"learning_rate": 3.9841073074896517e-05,
"loss": 0.3934,
"step": 807
},
{
"epoch": 2.2061433447098975,
"grad_norm": 0.1610165387392044,
"learning_rate": 3.958737745474638e-05,
"loss": 0.4207,
"step": 808
},
{
"epoch": 2.208873720136519,
"grad_norm": 0.15804022550582886,
"learning_rate": 3.933429265449882e-05,
"loss": 0.3968,
"step": 809
},
{
"epoch": 2.21160409556314,
"grad_norm": 0.15507763624191284,
"learning_rate": 3.9081821233043436e-05,
"loss": 0.4322,
"step": 810
},
{
"epoch": 2.214334470989761,
"grad_norm": 0.16331470012664795,
"learning_rate": 3.8829965743068174e-05,
"loss": 0.4317,
"step": 811
},
{
"epoch": 2.217064846416382,
"grad_norm": 0.14913159608840942,
"learning_rate": 3.857872873103322e-05,
"loss": 0.4098,
"step": 812
},
{
"epoch": 2.2197952218430035,
"grad_norm": 0.15193897485733032,
"learning_rate": 3.832811273714569e-05,
"loss": 0.4319,
"step": 813
},
{
"epoch": 2.2225255972696245,
"grad_norm": 0.15656188130378723,
"learning_rate": 3.807812029533362e-05,
"loss": 0.3962,
"step": 814
},
{
"epoch": 2.2252559726962455,
"grad_norm": 0.1457897126674652,
"learning_rate": 3.7828753933220295e-05,
"loss": 0.4044,
"step": 815
},
{
"epoch": 2.227986348122867,
"grad_norm": 0.15471549332141876,
"learning_rate": 3.758001617209906e-05,
"loss": 0.4251,
"step": 816
},
{
"epoch": 2.230716723549488,
"grad_norm": 0.16151392459869385,
"learning_rate": 3.733190952690753e-05,
"loss": 0.4278,
"step": 817
},
{
"epoch": 2.233447098976109,
"grad_norm": 0.15231560170650482,
"learning_rate": 3.708443650620206e-05,
"loss": 0.4286,
"step": 818
},
{
"epoch": 2.2361774744027305,
"grad_norm": 0.13981011509895325,
"learning_rate": 3.683759961213282e-05,
"loss": 0.4127,
"step": 819
},
{
"epoch": 2.2389078498293515,
"grad_norm": 0.16484162211418152,
"learning_rate": 3.6591401340418116e-05,
"loss": 0.4399,
"step": 820
},
{
"epoch": 2.2416382252559726,
"grad_norm": 0.15228329598903656,
"learning_rate": 3.634584418031915e-05,
"loss": 0.4247,
"step": 821
},
{
"epoch": 2.244368600682594,
"grad_norm": 0.14780929684638977,
"learning_rate": 3.6100930614615205e-05,
"loss": 0.4324,
"step": 822
},
{
"epoch": 2.247098976109215,
"grad_norm": 0.1611994057893753,
"learning_rate": 3.585666311957817e-05,
"loss": 0.4263,
"step": 823
},
{
"epoch": 2.249829351535836,
"grad_norm": 0.16382110118865967,
"learning_rate": 3.561304416494762e-05,
"loss": 0.4332,
"step": 824
},
{
"epoch": 2.252559726962457,
"grad_norm": 0.15646643936634064,
"learning_rate": 3.53700762139059e-05,
"loss": 0.4132,
"step": 825
},
{
"epoch": 2.2552901023890786,
"grad_norm": 0.16182062029838562,
"learning_rate": 3.512776172305331e-05,
"loss": 0.4199,
"step": 826
},
{
"epoch": 2.2580204778156996,
"grad_norm": 0.15407824516296387,
"learning_rate": 3.4886103142382945e-05,
"loss": 0.4087,
"step": 827
},
{
"epoch": 2.260750853242321,
"grad_norm": 0.1593010425567627,
"learning_rate": 3.46451029152562e-05,
"loss": 0.416,
"step": 828
},
{
"epoch": 2.263481228668942,
"grad_norm": 0.15628038346767426,
"learning_rate": 3.440476347837811e-05,
"loss": 0.4296,
"step": 829
},
{
"epoch": 2.266211604095563,
"grad_norm": 0.1571022868156433,
"learning_rate": 3.41650872617724e-05,
"loss": 0.419,
"step": 830
},
{
"epoch": 2.268941979522184,
"grad_norm": 0.15472716093063354,
"learning_rate": 3.392607668875718e-05,
"loss": 0.4151,
"step": 831
},
{
"epoch": 2.2716723549488056,
"grad_norm": 0.1419110745191574,
"learning_rate": 3.36877341759205e-05,
"loss": 0.3991,
"step": 832
},
{
"epoch": 2.2744027303754266,
"grad_norm": 0.15272633731365204,
"learning_rate": 3.345006213309557e-05,
"loss": 0.3965,
"step": 833
},
{
"epoch": 2.2771331058020476,
"grad_norm": 0.15313783288002014,
"learning_rate": 3.321306296333673e-05,
"loss": 0.4022,
"step": 834
},
{
"epoch": 2.279863481228669,
"grad_norm": 0.14547322690486908,
"learning_rate": 3.29767390628951e-05,
"loss": 0.4127,
"step": 835
},
{
"epoch": 2.28259385665529,
"grad_norm": 0.15116067230701447,
"learning_rate": 3.274109282119413e-05,
"loss": 0.4086,
"step": 836
},
{
"epoch": 2.285324232081911,
"grad_norm": 0.1611739993095398,
"learning_rate": 3.250612662080567e-05,
"loss": 0.4261,
"step": 837
},
{
"epoch": 2.288054607508532,
"grad_norm": 0.1671733260154724,
"learning_rate": 3.227184283742591e-05,
"loss": 0.4244,
"step": 838
},
{
"epoch": 2.2907849829351536,
"grad_norm": 0.16232764720916748,
"learning_rate": 3.2038243839851075e-05,
"loss": 0.4118,
"step": 839
},
{
"epoch": 2.2935153583617747,
"grad_norm": 0.1595815271139145,
"learning_rate": 3.180533198995379e-05,
"loss": 0.4333,
"step": 840
},
{
"epoch": 2.296245733788396,
"grad_norm": 0.14766521751880646,
"learning_rate": 3.1573109642659024e-05,
"loss": 0.4066,
"step": 841
},
{
"epoch": 2.298976109215017,
"grad_norm": 0.16081617772579193,
"learning_rate": 3.134157914592032e-05,
"loss": 0.4032,
"step": 842
},
{
"epoch": 2.301706484641638,
"grad_norm": 0.15681862831115723,
"learning_rate": 3.111074284069606e-05,
"loss": 0.4256,
"step": 843
},
{
"epoch": 2.304436860068259,
"grad_norm": 0.15217512845993042,
"learning_rate": 3.088060306092582e-05,
"loss": 0.4142,
"step": 844
},
{
"epoch": 2.3071672354948807,
"grad_norm": 0.15541419386863708,
"learning_rate": 3.065116213350671e-05,
"loss": 0.4246,
"step": 845
},
{
"epoch": 2.3098976109215017,
"grad_norm": 0.16863110661506653,
"learning_rate": 3.042242237826991e-05,
"loss": 0.4167,
"step": 846
},
{
"epoch": 2.3126279863481227,
"grad_norm": 0.15719062089920044,
"learning_rate": 3.0194386107957173e-05,
"loss": 0.4272,
"step": 847
},
{
"epoch": 2.315358361774744,
"grad_norm": 0.14599116146564484,
"learning_rate": 2.9967055628197472e-05,
"loss": 0.3826,
"step": 848
},
{
"epoch": 2.318088737201365,
"grad_norm": 0.16562673449516296,
"learning_rate": 2.974043323748367e-05,
"loss": 0.4246,
"step": 849
},
{
"epoch": 2.3208191126279862,
"grad_norm": 0.16425776481628418,
"learning_rate": 2.951452122714926e-05,
"loss": 0.4182,
"step": 850
},
{
"epoch": 2.3235494880546073,
"grad_norm": 0.15654055774211884,
"learning_rate": 2.9289321881345254e-05,
"loss": 0.4349,
"step": 851
},
{
"epoch": 2.3262798634812287,
"grad_norm": 0.15405891835689545,
"learning_rate": 2.9064837477017048e-05,
"loss": 0.4045,
"step": 852
},
{
"epoch": 2.3290102389078498,
"grad_norm": 0.16419056057929993,
"learning_rate": 2.88410702838814e-05,
"loss": 0.4174,
"step": 853
},
{
"epoch": 2.331740614334471,
"grad_norm": 0.16045036911964417,
"learning_rate": 2.861802256440348e-05,
"loss": 0.4185,
"step": 854
},
{
"epoch": 2.3344709897610922,
"grad_norm": 0.15399502217769623,
"learning_rate": 2.8395696573774032e-05,
"loss": 0.4228,
"step": 855
},
{
"epoch": 2.3372013651877133,
"grad_norm": 0.1503557562828064,
"learning_rate": 2.8174094559886534e-05,
"loss": 0.4146,
"step": 856
},
{
"epoch": 2.3399317406143343,
"grad_norm": 0.16562052071094513,
"learning_rate": 2.7953218763314458e-05,
"loss": 0.4127,
"step": 857
},
{
"epoch": 2.3426621160409558,
"grad_norm": 0.15860068798065186,
"learning_rate": 2.773307141728867e-05,
"loss": 0.4221,
"step": 858
},
{
"epoch": 2.345392491467577,
"grad_norm": 0.16427016258239746,
"learning_rate": 2.7513654747674788e-05,
"loss": 0.4202,
"step": 859
},
{
"epoch": 2.348122866894198,
"grad_norm": 0.15462426841259003,
"learning_rate": 2.729497097295075e-05,
"loss": 0.3986,
"step": 860
},
{
"epoch": 2.3508532423208193,
"grad_norm": 0.150539368391037,
"learning_rate": 2.7077022304184295e-05,
"loss": 0.4112,
"step": 861
},
{
"epoch": 2.3535836177474403,
"grad_norm": 0.14942197501659393,
"learning_rate": 2.685981094501069e-05,
"loss": 0.4173,
"step": 862
},
{
"epoch": 2.3563139931740613,
"grad_norm": 0.15172016620635986,
"learning_rate": 2.6643339091610377e-05,
"loss": 0.4273,
"step": 863
},
{
"epoch": 2.359044368600683,
"grad_norm": 0.15123523771762848,
"learning_rate": 2.6427608932686843e-05,
"loss": 0.4104,
"step": 864
},
{
"epoch": 2.361774744027304,
"grad_norm": 0.15563499927520752,
"learning_rate": 2.621262264944444e-05,
"loss": 0.4215,
"step": 865
},
{
"epoch": 2.364505119453925,
"grad_norm": 0.15889500081539154,
"learning_rate": 2.599838241556626e-05,
"loss": 0.4058,
"step": 866
},
{
"epoch": 2.3672354948805463,
"grad_norm": 0.1514395922422409,
"learning_rate": 2.5784890397192398e-05,
"loss": 0.4058,
"step": 867
},
{
"epoch": 2.3699658703071673,
"grad_norm": 0.15626998245716095,
"learning_rate": 2.5572148752897795e-05,
"loss": 0.4092,
"step": 868
},
{
"epoch": 2.3726962457337883,
"grad_norm": 0.15478669106960297,
"learning_rate": 2.5360159633670457e-05,
"loss": 0.424,
"step": 869
},
{
"epoch": 2.3754266211604094,
"grad_norm": 0.14791764318943024,
"learning_rate": 2.514892518288988e-05,
"loss": 0.4294,
"step": 870
},
{
"epoch": 2.378156996587031,
"grad_norm": 0.15729525685310364,
"learning_rate": 2.4938447536305243e-05,
"loss": 0.4204,
"step": 871
},
{
"epoch": 2.380887372013652,
"grad_norm": 0.1549883335828781,
"learning_rate": 2.472872882201368e-05,
"loss": 0.4302,
"step": 872
},
{
"epoch": 2.383617747440273,
"grad_norm": 0.15608322620391846,
"learning_rate": 2.451977116043911e-05,
"loss": 0.4208,
"step": 873
},
{
"epoch": 2.3863481228668944,
"grad_norm": 0.15836332738399506,
"learning_rate": 2.431157666431052e-05,
"loss": 0.4141,
"step": 874
},
{
"epoch": 2.3890784982935154,
"grad_norm": 0.15775950253009796,
"learning_rate": 2.410414743864059e-05,
"loss": 0.4142,
"step": 875
},
{
"epoch": 2.3918088737201364,
"grad_norm": 0.1490509808063507,
"learning_rate": 2.3897485580704682e-05,
"loss": 0.4079,
"step": 876
},
{
"epoch": 2.394539249146758,
"grad_norm": 0.16001944243907928,
"learning_rate": 2.3691593180019366e-05,
"loss": 0.4276,
"step": 877
},
{
"epoch": 2.397269624573379,
"grad_norm": 0.14967067539691925,
"learning_rate": 2.3486472318321307e-05,
"loss": 0.4045,
"step": 878
},
{
"epoch": 2.4,
"grad_norm": 0.14551271498203278,
"learning_rate": 2.3282125069546433e-05,
"loss": 0.4169,
"step": 879
},
{
"epoch": 2.4027303754266214,
"grad_norm": 0.15170224010944366,
"learning_rate": 2.3078553499808797e-05,
"loss": 0.4166,
"step": 880
},
{
"epoch": 2.4054607508532424,
"grad_norm": 0.15381450951099396,
"learning_rate": 2.2875759667379614e-05,
"loss": 0.4046,
"step": 881
},
{
"epoch": 2.4081911262798634,
"grad_norm": 0.15133820474147797,
"learning_rate": 2.267374562266662e-05,
"loss": 0.4053,
"step": 882
},
{
"epoch": 2.4109215017064844,
"grad_norm": 0.15135996043682098,
"learning_rate": 2.2472513408193384e-05,
"loss": 0.4145,
"step": 883
},
{
"epoch": 2.413651877133106,
"grad_norm": 0.1551310420036316,
"learning_rate": 2.227206505857834e-05,
"loss": 0.4107,
"step": 884
},
{
"epoch": 2.416382252559727,
"grad_norm": 0.15202271938323975,
"learning_rate": 2.207240260051453e-05,
"loss": 0.409,
"step": 885
},
{
"epoch": 2.419112627986348,
"grad_norm": 0.15912394225597382,
"learning_rate": 2.1873528052749092e-05,
"loss": 0.4293,
"step": 886
},
{
"epoch": 2.4218430034129694,
"grad_norm": 0.164555624127388,
"learning_rate": 2.167544342606256e-05,
"loss": 0.4153,
"step": 887
},
{
"epoch": 2.4245733788395905,
"grad_norm": 0.1536960005760193,
"learning_rate": 2.1478150723248857e-05,
"loss": 0.4161,
"step": 888
},
{
"epoch": 2.4273037542662115,
"grad_norm": 0.1572561115026474,
"learning_rate": 2.1281651939094992e-05,
"loss": 0.3975,
"step": 889
},
{
"epoch": 2.430034129692833,
"grad_norm": 0.14872194826602936,
"learning_rate": 2.1085949060360654e-05,
"loss": 0.4028,
"step": 890
},
{
"epoch": 2.432764505119454,
"grad_norm": 0.15553632378578186,
"learning_rate": 2.089104406575837e-05,
"loss": 0.4113,
"step": 891
},
{
"epoch": 2.435494880546075,
"grad_norm": 0.15172426402568817,
"learning_rate": 2.0696938925933506e-05,
"loss": 0.407,
"step": 892
},
{
"epoch": 2.4382252559726965,
"grad_norm": 0.1533356010913849,
"learning_rate": 2.0503635603444094e-05,
"loss": 0.4225,
"step": 893
},
{
"epoch": 2.4409556313993175,
"grad_norm": 0.15870912373065948,
"learning_rate": 2.0311136052741277e-05,
"loss": 0.4176,
"step": 894
},
{
"epoch": 2.4436860068259385,
"grad_norm": 0.15616737306118011,
"learning_rate": 2.0119442220149353e-05,
"loss": 0.4158,
"step": 895
},
{
"epoch": 2.4464163822525595,
"grad_norm": 0.15833789110183716,
"learning_rate": 1.9928556043846214e-05,
"loss": 0.403,
"step": 896
},
{
"epoch": 2.449146757679181,
"grad_norm": 0.1554342359304428,
"learning_rate": 1.9738479453843682e-05,
"loss": 0.396,
"step": 897
},
{
"epoch": 2.451877133105802,
"grad_norm": 0.17515774071216583,
"learning_rate": 1.9549214371968004e-05,
"loss": 0.4096,
"step": 898
},
{
"epoch": 2.454607508532423,
"grad_norm": 0.15816594660282135,
"learning_rate": 1.936076271184044e-05,
"loss": 0.4232,
"step": 899
},
{
"epoch": 2.4573378839590445,
"grad_norm": 0.1528582125902176,
"learning_rate": 1.9173126378857907e-05,
"loss": 0.4145,
"step": 900
},
{
"epoch": 2.4600682593856655,
"grad_norm": 0.16006483137607574,
"learning_rate": 1.898630727017371e-05,
"loss": 0.4201,
"step": 901
},
{
"epoch": 2.4627986348122866,
"grad_norm": 0.15996922552585602,
"learning_rate": 1.8800307274678364e-05,
"loss": 0.4056,
"step": 902
},
{
"epoch": 2.465529010238908,
"grad_norm": 0.1555391103029251,
"learning_rate": 1.861512827298051e-05,
"loss": 0.4315,
"step": 903
},
{
"epoch": 2.468259385665529,
"grad_norm": 0.15501669049263,
"learning_rate": 1.8430772137387853e-05,
"loss": 0.4159,
"step": 904
},
{
"epoch": 2.47098976109215,
"grad_norm": 0.1523975431919098,
"learning_rate": 1.8247240731888294e-05,
"loss": 0.4004,
"step": 905
},
{
"epoch": 2.4737201365187715,
"grad_norm": 0.1715194135904312,
"learning_rate": 1.806453591213103e-05,
"loss": 0.4124,
"step": 906
},
{
"epoch": 2.4764505119453926,
"grad_norm": 0.15396980941295624,
"learning_rate": 1.788265952540784e-05,
"loss": 0.4094,
"step": 907
},
{
"epoch": 2.4791808873720136,
"grad_norm": 0.1634356677532196,
"learning_rate": 1.7701613410634365e-05,
"loss": 0.4257,
"step": 908
},
{
"epoch": 2.4819112627986346,
"grad_norm": 0.1548430621623993,
"learning_rate": 1.752139939833154e-05,
"loss": 0.3942,
"step": 909
},
{
"epoch": 2.484641638225256,
"grad_norm": 0.15945452451705933,
"learning_rate": 1.734201931060706e-05,
"loss": 0.3979,
"step": 910
},
{
"epoch": 2.487372013651877,
"grad_norm": 0.16635702550411224,
"learning_rate": 1.7163474961137028e-05,
"loss": 0.422,
"step": 911
},
{
"epoch": 2.490102389078498,
"grad_norm": 0.16245630383491516,
"learning_rate": 1.6985768155147496e-05,
"loss": 0.4126,
"step": 912
},
{
"epoch": 2.4928327645051196,
"grad_norm": 0.14662671089172363,
"learning_rate": 1.6808900689396336e-05,
"loss": 0.4062,
"step": 913
},
{
"epoch": 2.4955631399317406,
"grad_norm": 0.1555013507604599,
"learning_rate": 1.663287435215498e-05,
"loss": 0.4101,
"step": 914
},
{
"epoch": 2.4982935153583616,
"grad_norm": 0.14830157160758972,
"learning_rate": 1.645769092319045e-05,
"loss": 0.3843,
"step": 915
},
{
"epoch": 2.5010238907849827,
"grad_norm": 0.16135641932487488,
"learning_rate": 1.6283352173747145e-05,
"loss": 0.4229,
"step": 916
},
{
"epoch": 2.503754266211604,
"grad_norm": 0.15229038894176483,
"learning_rate": 1.6109859866529255e-05,
"loss": 0.4209,
"step": 917
},
{
"epoch": 2.506484641638225,
"grad_norm": 0.15836934745311737,
"learning_rate": 1.5937215755682665e-05,
"loss": 0.422,
"step": 918
},
{
"epoch": 2.5092150170648466,
"grad_norm": 0.1639019101858139,
"learning_rate": 1.5765421586777284e-05,
"loss": 0.4206,
"step": 919
},
{
"epoch": 2.5119453924914676,
"grad_norm": 0.15451960265636444,
"learning_rate": 1.5594479096789537e-05,
"loss": 0.411,
"step": 920
},
{
"epoch": 2.5146757679180887,
"grad_norm": 0.16513267159461975,
"learning_rate": 1.5424390014084644e-05,
"loss": 0.4324,
"step": 921
},
{
"epoch": 2.5174061433447097,
"grad_norm": 0.15432654321193695,
"learning_rate": 1.5255156058399122e-05,
"loss": 0.4074,
"step": 922
},
{
"epoch": 2.520136518771331,
"grad_norm": 0.16064870357513428,
"learning_rate": 1.5086778940823543e-05,
"loss": 0.417,
"step": 923
},
{
"epoch": 2.522866894197952,
"grad_norm": 0.16009055078029633,
"learning_rate": 1.4919260363785215e-05,
"loss": 0.4128,
"step": 924
},
{
"epoch": 2.5255972696245736,
"grad_norm": 0.1598517745733261,
"learning_rate": 1.4752602021030792e-05,
"loss": 0.4191,
"step": 925
},
{
"epoch": 2.5283276450511947,
"grad_norm": 0.15252196788787842,
"learning_rate": 1.4586805597609331e-05,
"loss": 0.4124,
"step": 926
},
{
"epoch": 2.5310580204778157,
"grad_norm": 0.1643335521221161,
"learning_rate": 1.442187276985526e-05,
"loss": 0.4207,
"step": 927
},
{
"epoch": 2.5337883959044367,
"grad_norm": 0.15445098280906677,
"learning_rate": 1.4257805205371234e-05,
"loss": 0.3993,
"step": 928
},
{
"epoch": 2.536518771331058,
"grad_norm": 0.15789660811424255,
"learning_rate": 1.4094604563011472e-05,
"loss": 0.4103,
"step": 929
},
{
"epoch": 2.539249146757679,
"grad_norm": 0.15002034604549408,
"learning_rate": 1.3932272492864984e-05,
"loss": 0.4042,
"step": 930
},
{
"epoch": 2.5419795221843002,
"grad_norm": 0.16790151596069336,
"learning_rate": 1.3770810636238684e-05,
"loss": 0.4245,
"step": 931
},
{
"epoch": 2.5447098976109217,
"grad_norm": 0.1551153063774109,
"learning_rate": 1.3610220625641002e-05,
"loss": 0.4145,
"step": 932
},
{
"epoch": 2.5474402730375427,
"grad_norm": 0.15363937616348267,
"learning_rate": 1.3450504084765381e-05,
"loss": 0.384,
"step": 933
},
{
"epoch": 2.5501706484641637,
"grad_norm": 0.15322524309158325,
"learning_rate": 1.3291662628473633e-05,
"loss": 0.4042,
"step": 934
},
{
"epoch": 2.5529010238907848,
"grad_norm": 0.1649988293647766,
"learning_rate": 1.313369786277987e-05,
"loss": 0.4236,
"step": 935
},
{
"epoch": 2.5556313993174062,
"grad_norm": 0.150667205452919,
"learning_rate": 1.2976611384834148e-05,
"loss": 0.3908,
"step": 936
},
{
"epoch": 2.5583617747440273,
"grad_norm": 0.16129009425640106,
"learning_rate": 1.2820404782906315e-05,
"loss": 0.4167,
"step": 937
},
{
"epoch": 2.5610921501706487,
"grad_norm": 0.16965742409229279,
"learning_rate": 1.2665079636369969e-05,
"loss": 0.4145,
"step": 938
},
{
"epoch": 2.5638225255972698,
"grad_norm": 0.14878158271312714,
"learning_rate": 1.2510637515686496e-05,
"loss": 0.4236,
"step": 939
},
{
"epoch": 2.5665529010238908,
"grad_norm": 0.14919213950634003,
"learning_rate": 1.2357079982389197e-05,
"loss": 0.396,
"step": 940
},
{
"epoch": 2.569283276450512,
"grad_norm": 0.1563798040151596,
"learning_rate": 1.2204408589067462e-05,
"loss": 0.3992,
"step": 941
},
{
"epoch": 2.5720136518771333,
"grad_norm": 0.15581347048282623,
"learning_rate": 1.2052624879351104e-05,
"loss": 0.4261,
"step": 942
},
{
"epoch": 2.5747440273037543,
"grad_norm": 0.15438248217105865,
"learning_rate": 1.190173038789476e-05,
"loss": 0.4013,
"step": 943
},
{
"epoch": 2.5774744027303753,
"grad_norm": 0.15199199318885803,
"learning_rate": 1.1751726640362349e-05,
"loss": 0.4089,
"step": 944
},
{
"epoch": 2.580204778156997,
"grad_norm": 0.1516939401626587,
"learning_rate": 1.1602615153411667e-05,
"loss": 0.4008,
"step": 945
},
{
"epoch": 2.582935153583618,
"grad_norm": 0.15474575757980347,
"learning_rate": 1.1454397434679021e-05,
"loss": 0.4115,
"step": 946
},
{
"epoch": 2.585665529010239,
"grad_norm": 0.15419447422027588,
"learning_rate": 1.1307074982764022e-05,
"loss": 0.4187,
"step": 947
},
{
"epoch": 2.58839590443686,
"grad_norm": 0.15413175523281097,
"learning_rate": 1.116064928721442e-05,
"loss": 0.4191,
"step": 948
},
{
"epoch": 2.5911262798634813,
"grad_norm": 0.1559099704027176,
"learning_rate": 1.1015121828511032e-05,
"loss": 0.4136,
"step": 949
},
{
"epoch": 2.5938566552901023,
"grad_norm": 0.16210560500621796,
"learning_rate": 1.0870494078052796e-05,
"loss": 0.4204,
"step": 950
},
{
"epoch": 2.596587030716724,
"grad_norm": 0.15410131216049194,
"learning_rate": 1.0726767498141877e-05,
"loss": 0.4098,
"step": 951
},
{
"epoch": 2.599317406143345,
"grad_norm": 0.14466793835163116,
"learning_rate": 1.0583943541968856e-05,
"loss": 0.3832,
"step": 952
},
{
"epoch": 2.602047781569966,
"grad_norm": 0.1512717753648758,
"learning_rate": 1.044202365359811e-05,
"loss": 0.4132,
"step": 953
},
{
"epoch": 2.604778156996587,
"grad_norm": 0.1530720293521881,
"learning_rate": 1.0301009267953143e-05,
"loss": 0.4165,
"step": 954
},
{
"epoch": 2.6075085324232083,
"grad_norm": 0.16781674325466156,
"learning_rate": 1.0160901810802115e-05,
"loss": 0.4203,
"step": 955
},
{
"epoch": 2.6102389078498294,
"grad_norm": 0.14876051247119904,
"learning_rate": 1.0021702698743407e-05,
"loss": 0.4168,
"step": 956
},
{
"epoch": 2.6129692832764504,
"grad_norm": 0.15021638572216034,
"learning_rate": 9.883413339191294e-06,
"loss": 0.4173,
"step": 957
},
{
"epoch": 2.615699658703072,
"grad_norm": 0.159826397895813,
"learning_rate": 9.746035130361742e-06,
"loss": 0.4279,
"step": 958
},
{
"epoch": 2.618430034129693,
"grad_norm": 0.156574085354805,
"learning_rate": 9.609569461258262e-06,
"loss": 0.4277,
"step": 959
},
{
"epoch": 2.621160409556314,
"grad_norm": 0.157151460647583,
"learning_rate": 9.474017711657834e-06,
"loss": 0.412,
"step": 960
},
{
"epoch": 2.623890784982935,
"grad_norm": 0.14887213706970215,
"learning_rate": 9.339381252097e-06,
"loss": 0.4012,
"step": 961
},
{
"epoch": 2.6266211604095564,
"grad_norm": 0.15450581908226013,
"learning_rate": 9.205661443857994e-06,
"loss": 0.4077,
"step": 962
},
{
"epoch": 2.6293515358361774,
"grad_norm": 0.14593878388404846,
"learning_rate": 9.072859638954955e-06,
"loss": 0.4064,
"step": 963
},
{
"epoch": 2.632081911262799,
"grad_norm": 0.156602144241333,
"learning_rate": 8.940977180120247e-06,
"loss": 0.4267,
"step": 964
},
{
"epoch": 2.63481228668942,
"grad_norm": 0.16739366948604584,
"learning_rate": 8.810015400790994e-06,
"loss": 0.4176,
"step": 965
},
{
"epoch": 2.637542662116041,
"grad_norm": 0.15772853791713715,
"learning_rate": 8.67997562509546e-06,
"loss": 0.424,
"step": 966
},
{
"epoch": 2.640273037542662,
"grad_norm": 0.15860068798065186,
"learning_rate": 8.550859167839664e-06,
"loss": 0.4167,
"step": 967
},
{
"epoch": 2.6430034129692834,
"grad_norm": 0.15044620633125305,
"learning_rate": 8.422667334494249e-06,
"loss": 0.3916,
"step": 968
},
{
"epoch": 2.6457337883959045,
"grad_norm": 0.150175541639328,
"learning_rate": 8.295401421181125e-06,
"loss": 0.3953,
"step": 969
},
{
"epoch": 2.6484641638225255,
"grad_norm": 0.14696063101291656,
"learning_rate": 8.169062714660346e-06,
"loss": 0.4115,
"step": 970
},
{
"epoch": 2.651194539249147,
"grad_norm": 0.1482568085193634,
"learning_rate": 8.043652492317256e-06,
"loss": 0.4017,
"step": 971
},
{
"epoch": 2.653924914675768,
"grad_norm": 0.15316608548164368,
"learning_rate": 7.919172022149456e-06,
"loss": 0.4176,
"step": 972
},
{
"epoch": 2.656655290102389,
"grad_norm": 0.15325787663459778,
"learning_rate": 7.795622562753957e-06,
"loss": 0.413,
"step": 973
},
{
"epoch": 2.65938566552901,
"grad_norm": 0.16979162395000458,
"learning_rate": 7.673005363314579e-06,
"loss": 0.4244,
"step": 974
},
{
"epoch": 2.6621160409556315,
"grad_norm": 0.16061224043369293,
"learning_rate": 7.551321663589228e-06,
"loss": 0.4082,
"step": 975
},
{
"epoch": 2.6648464163822525,
"grad_norm": 0.1511377990245819,
"learning_rate": 7.430572693897342e-06,
"loss": 0.4047,
"step": 976
},
{
"epoch": 2.667576791808874,
"grad_norm": 0.1549064815044403,
"learning_rate": 7.310759675107515e-06,
"loss": 0.4181,
"step": 977
},
{
"epoch": 2.670307167235495,
"grad_norm": 0.15855662524700165,
"learning_rate": 7.191883818625189e-06,
"loss": 0.4242,
"step": 978
},
{
"epoch": 2.673037542662116,
"grad_norm": 0.16046655178070068,
"learning_rate": 7.073946326380243e-06,
"loss": 0.4077,
"step": 979
},
{
"epoch": 2.675767918088737,
"grad_norm": 0.1561538577079773,
"learning_rate": 6.956948390814977e-06,
"loss": 0.4117,
"step": 980
},
{
"epoch": 2.6784982935153585,
"grad_norm": 0.16078175604343414,
"learning_rate": 6.840891194872112e-06,
"loss": 0.4342,
"step": 981
},
{
"epoch": 2.6812286689419795,
"grad_norm": 0.15352275967597961,
"learning_rate": 6.725775911982601e-06,
"loss": 0.402,
"step": 982
},
{
"epoch": 2.6839590443686006,
"grad_norm": 0.15391647815704346,
"learning_rate": 6.6116037060539704e-06,
"loss": 0.4095,
"step": 983
},
{
"epoch": 2.686689419795222,
"grad_norm": 0.15556836128234863,
"learning_rate": 6.498375731458528e-06,
"loss": 0.4063,
"step": 984
},
{
"epoch": 2.689419795221843,
"grad_norm": 0.15222905576229095,
"learning_rate": 6.386093133021554e-06,
"loss": 0.4139,
"step": 985
},
{
"epoch": 2.692150170648464,
"grad_norm": 0.15189246833324432,
"learning_rate": 6.274757046009871e-06,
"loss": 0.4195,
"step": 986
},
{
"epoch": 2.694880546075085,
"grad_norm": 0.15635477006435394,
"learning_rate": 6.164368596120351e-06,
"loss": 0.4137,
"step": 987
},
{
"epoch": 2.6976109215017066,
"grad_norm": 0.15748678147792816,
"learning_rate": 6.054928899468426e-06,
"loss": 0.396,
"step": 988
},
{
"epoch": 2.7003412969283276,
"grad_norm": 0.1585109978914261,
"learning_rate": 5.946439062576903e-06,
"loss": 0.4111,
"step": 989
},
{
"epoch": 2.703071672354949,
"grad_norm": 0.15509194135665894,
"learning_rate": 5.83890018236476e-06,
"loss": 0.4259,
"step": 990
},
{
"epoch": 2.70580204778157,
"grad_norm": 0.15443935990333557,
"learning_rate": 5.732313346136031e-06,
"loss": 0.4138,
"step": 991
},
{
"epoch": 2.708532423208191,
"grad_norm": 0.1619240939617157,
"learning_rate": 5.626679631568832e-06,
"loss": 0.4091,
"step": 992
},
{
"epoch": 2.711262798634812,
"grad_norm": 0.1597377061843872,
"learning_rate": 5.522000106704439e-06,
"loss": 0.4193,
"step": 993
},
{
"epoch": 2.7139931740614336,
"grad_norm": 0.15020039677619934,
"learning_rate": 5.418275829936537e-06,
"loss": 0.4216,
"step": 994
},
{
"epoch": 2.7167235494880546,
"grad_norm": 0.15186108648777008,
"learning_rate": 5.315507850000456e-06,
"loss": 0.4057,
"step": 995
},
{
"epoch": 2.7194539249146756,
"grad_norm": 0.1630185842514038,
"learning_rate": 5.2136972059626314e-06,
"loss": 0.4141,
"step": 996
},
{
"epoch": 2.722184300341297,
"grad_norm": 0.1610775589942932,
"learning_rate": 5.112844927210048e-06,
"loss": 0.4025,
"step": 997
},
{
"epoch": 2.724914675767918,
"grad_norm": 0.15820352733135223,
"learning_rate": 5.012952033439844e-06,
"loss": 0.4197,
"step": 998
},
{
"epoch": 2.727645051194539,
"grad_norm": 0.1567496657371521,
"learning_rate": 4.914019534649039e-06,
"loss": 0.4215,
"step": 999
},
{
"epoch": 2.73037542662116,
"grad_norm": 0.1540801227092743,
"learning_rate": 4.816048431124265e-06,
"loss": 0.4237,
"step": 1000
},
{
"epoch": 2.7331058020477816,
"grad_norm": 0.15339985489845276,
"learning_rate": 4.719039713431694e-06,
"loss": 0.4127,
"step": 1001
},
{
"epoch": 2.7358361774744027,
"grad_norm": 0.1545177400112152,
"learning_rate": 4.622994362406996e-06,
"loss": 0.424,
"step": 1002
},
{
"epoch": 2.738566552901024,
"grad_norm": 0.15001171827316284,
"learning_rate": 4.527913349145441e-06,
"loss": 0.4199,
"step": 1003
},
{
"epoch": 2.741296928327645,
"grad_norm": 0.15352268517017365,
"learning_rate": 4.433797634992077e-06,
"loss": 0.3991,
"step": 1004
},
{
"epoch": 2.744027303754266,
"grad_norm": 0.1552933305501938,
"learning_rate": 4.340648171531992e-06,
"loss": 0.4173,
"step": 1005
},
{
"epoch": 2.746757679180887,
"grad_norm": 0.1541508436203003,
"learning_rate": 4.248465900580734e-06,
"loss": 0.4231,
"step": 1006
},
{
"epoch": 2.7494880546075087,
"grad_norm": 0.15616647899150848,
"learning_rate": 4.1572517541747294e-06,
"loss": 0.4295,
"step": 1007
},
{
"epoch": 2.7522184300341297,
"grad_norm": 0.14823675155639648,
"learning_rate": 4.0670066545619225e-06,
"loss": 0.403,
"step": 1008
},
{
"epoch": 2.7549488054607507,
"grad_norm": 0.15372464060783386,
"learning_rate": 3.977731514192385e-06,
"loss": 0.4184,
"step": 1009
},
{
"epoch": 2.757679180887372,
"grad_norm": 0.15206997096538544,
"learning_rate": 3.889427235709153e-06,
"loss": 0.4119,
"step": 1010
},
{
"epoch": 2.760409556313993,
"grad_norm": 0.15120883285999298,
"learning_rate": 3.802094711939075e-06,
"loss": 0.413,
"step": 1011
},
{
"epoch": 2.7631399317406142,
"grad_norm": 0.16259510815143585,
"learning_rate": 3.7157348258837652e-06,
"loss": 0.4253,
"step": 1012
},
{
"epoch": 2.7658703071672353,
"grad_norm": 0.15381862223148346,
"learning_rate": 3.6303484507106966e-06,
"loss": 0.4162,
"step": 1013
},
{
"epoch": 2.7686006825938567,
"grad_norm": 0.15234719216823578,
"learning_rate": 3.5459364497443694e-06,
"loss": 0.3893,
"step": 1014
},
{
"epoch": 2.7713310580204777,
"grad_norm": 0.1562722623348236,
"learning_rate": 3.4624996764575977e-06,
"loss": 0.3979,
"step": 1015
},
{
"epoch": 2.774061433447099,
"grad_norm": 0.15057340264320374,
"learning_rate": 3.3800389744628404e-06,
"loss": 0.3861,
"step": 1016
},
{
"epoch": 2.7767918088737202,
"grad_norm": 0.1512756198644638,
"learning_rate": 3.298555177503726e-06,
"loss": 0.4135,
"step": 1017
},
{
"epoch": 2.7795221843003413,
"grad_norm": 0.1467510461807251,
"learning_rate": 3.2180491094465415e-06,
"loss": 0.4133,
"step": 1018
},
{
"epoch": 2.7822525597269623,
"grad_norm": 0.1518513709306717,
"learning_rate": 3.1385215842720027e-06,
"loss": 0.4032,
"step": 1019
},
{
"epoch": 2.7849829351535837,
"grad_norm": 0.15052153170108795,
"learning_rate": 3.059973406066963e-06,
"loss": 0.4157,
"step": 1020
},
{
"epoch": 2.7877133105802048,
"grad_norm": 0.15297654271125793,
"learning_rate": 2.9824053690162723e-06,
"loss": 0.4153,
"step": 1021
},
{
"epoch": 2.790443686006826,
"grad_norm": 0.15268519520759583,
"learning_rate": 2.905818257394799e-06,
"loss": 0.4083,
"step": 1022
},
{
"epoch": 2.7931740614334473,
"grad_norm": 0.1531766653060913,
"learning_rate": 2.8302128455594656e-06,
"loss": 0.3943,
"step": 1023
},
{
"epoch": 2.7959044368600683,
"grad_norm": 0.1546778380870819,
"learning_rate": 2.7555898979413797e-06,
"loss": 0.4091,
"step": 1024
},
{
"epoch": 2.7986348122866893,
"grad_norm": 0.15699811279773712,
"learning_rate": 2.6819501690382277e-06,
"loss": 0.4249,
"step": 1025
},
{
"epoch": 2.8013651877133103,
"grad_norm": 0.15149515867233276,
"learning_rate": 2.609294403406537e-06,
"loss": 0.4064,
"step": 1026
},
{
"epoch": 2.804095563139932,
"grad_norm": 0.15525740385055542,
"learning_rate": 2.537623335654127e-06,
"loss": 0.4086,
"step": 1027
},
{
"epoch": 2.806825938566553,
"grad_norm": 0.15066871047019958,
"learning_rate": 2.4669376904328247e-06,
"loss": 0.4046,
"step": 1028
},
{
"epoch": 2.8095563139931743,
"grad_norm": 0.1503557562828064,
"learning_rate": 2.397238182430994e-06,
"loss": 0.4007,
"step": 1029
},
{
"epoch": 2.8122866894197953,
"grad_norm": 0.15836714208126068,
"learning_rate": 2.3285255163663532e-06,
"loss": 0.4297,
"step": 1030
},
{
"epoch": 2.8150170648464163,
"grad_norm": 0.1532403528690338,
"learning_rate": 2.2608003869788786e-06,
"loss": 0.4096,
"step": 1031
},
{
"epoch": 2.8177474402730374,
"grad_norm": 0.1478443145751953,
"learning_rate": 2.1940634790238e-06,
"loss": 0.3819,
"step": 1032
},
{
"epoch": 2.820477815699659,
"grad_norm": 0.1495964378118515,
"learning_rate": 2.128315467264552e-06,
"loss": 0.4086,
"step": 1033
},
{
"epoch": 2.82320819112628,
"grad_norm": 0.15715493261814117,
"learning_rate": 2.063557016466111e-06,
"loss": 0.403,
"step": 1034
},
{
"epoch": 2.825938566552901,
"grad_norm": 0.14644889533519745,
"learning_rate": 1.999788781388201e-06,
"loss": 0.405,
"step": 1035
},
{
"epoch": 2.8286689419795223,
"grad_norm": 0.15347984433174133,
"learning_rate": 1.9370114067785994e-06,
"loss": 0.4198,
"step": 1036
},
{
"epoch": 2.8313993174061434,
"grad_norm": 0.14934033155441284,
"learning_rate": 1.8752255273667752e-06,
"loss": 0.4078,
"step": 1037
},
{
"epoch": 2.8341296928327644,
"grad_norm": 0.1513030081987381,
"learning_rate": 1.8144317678573497e-06,
"loss": 0.4165,
"step": 1038
},
{
"epoch": 2.8368600682593854,
"grad_norm": 0.15821826457977295,
"learning_rate": 1.754630742923813e-06,
"loss": 0.4213,
"step": 1039
},
{
"epoch": 2.839590443686007,
"grad_norm": 0.1506132036447525,
"learning_rate": 1.6958230572023503e-06,
"loss": 0.4058,
"step": 1040
},
{
"epoch": 2.842320819112628,
"grad_norm": 0.15292277932167053,
"learning_rate": 1.6380093052856483e-06,
"loss": 0.4219,
"step": 1041
},
{
"epoch": 2.8450511945392494,
"grad_norm": 0.15926344692707062,
"learning_rate": 1.5811900717169538e-06,
"loss": 0.4144,
"step": 1042
},
{
"epoch": 2.8477815699658704,
"grad_norm": 0.15735220909118652,
"learning_rate": 1.525365930984146e-06,
"loss": 0.3986,
"step": 1043
},
{
"epoch": 2.8505119453924914,
"grad_norm": 0.15544278919696808,
"learning_rate": 1.4705374475138978e-06,
"loss": 0.4151,
"step": 1044
},
{
"epoch": 2.8532423208191124,
"grad_norm": 0.14867156744003296,
"learning_rate": 1.416705175666e-06,
"loss": 0.4039,
"step": 1045
},
{
"epoch": 2.855972696245734,
"grad_norm": 0.15105663239955902,
"learning_rate": 1.3638696597277679e-06,
"loss": 0.4022,
"step": 1046
},
{
"epoch": 2.858703071672355,
"grad_norm": 0.15061454474925995,
"learning_rate": 1.3120314339084783e-06,
"loss": 0.3928,
"step": 1047
},
{
"epoch": 2.861433447098976,
"grad_norm": 0.15541335940361023,
"learning_rate": 1.2611910223340407e-06,
"loss": 0.4108,
"step": 1048
},
{
"epoch": 2.8641638225255974,
"grad_norm": 0.15430454909801483,
"learning_rate": 1.2113489390416566e-06,
"loss": 0.4142,
"step": 1049
},
{
"epoch": 2.8668941979522184,
"grad_norm": 0.1592140942811966,
"learning_rate": 1.1625056879746133e-06,
"loss": 0.4123,
"step": 1050
},
{
"epoch": 2.8696245733788395,
"grad_norm": 0.15730910003185272,
"learning_rate": 1.1146617629772315e-06,
"loss": 0.4047,
"step": 1051
},
{
"epoch": 2.8723549488054605,
"grad_norm": 0.15714263916015625,
"learning_rate": 1.0678176477898372e-06,
"loss": 0.4258,
"step": 1052
},
{
"epoch": 2.875085324232082,
"grad_norm": 0.15376383066177368,
"learning_rate": 1.0219738160438753e-06,
"loss": 0.4098,
"step": 1053
},
{
"epoch": 2.877815699658703,
"grad_norm": 0.15481893718242645,
"learning_rate": 9.771307312571254e-07,
"loss": 0.408,
"step": 1054
},
{
"epoch": 2.8805460750853245,
"grad_norm": 0.15664780139923096,
"learning_rate": 9.332888468290169e-07,
"loss": 0.3983,
"step": 1055
},
{
"epoch": 2.8832764505119455,
"grad_norm": 0.15640319883823395,
"learning_rate": 8.90448606036054e-07,
"loss": 0.4215,
"step": 1056
},
{
"epoch": 2.8860068259385665,
"grad_norm": 0.15267756581306458,
"learning_rate": 8.486104420272977e-07,
"loss": 0.401,
"step": 1057
},
{
"epoch": 2.8887372013651875,
"grad_norm": 0.15257929265499115,
"learning_rate": 8.077747778200473e-07,
"loss": 0.4145,
"step": 1058
},
{
"epoch": 2.891467576791809,
"grad_norm": 0.1439686268568039,
"learning_rate": 7.679420262954984e-07,
"loss": 0.3806,
"step": 1059
},
{
"epoch": 2.89419795221843,
"grad_norm": 0.15835194289684296,
"learning_rate": 7.291125901946027e-07,
"loss": 0.4156,
"step": 1060
},
{
"epoch": 2.896928327645051,
"grad_norm": 0.15433841943740845,
"learning_rate": 6.912868621140045e-07,
"loss": 0.4198,
"step": 1061
},
{
"epoch": 2.8996587030716725,
"grad_norm": 0.15369294583797455,
"learning_rate": 6.544652245020433e-07,
"loss": 0.4086,
"step": 1062
},
{
"epoch": 2.9023890784982935,
"grad_norm": 0.16046328842639923,
"learning_rate": 6.18648049654913e-07,
"loss": 0.4046,
"step": 1063
},
{
"epoch": 2.9051194539249146,
"grad_norm": 0.15750819444656372,
"learning_rate": 5.838356997128869e-07,
"loss": 0.4245,
"step": 1064
},
{
"epoch": 2.9078498293515356,
"grad_norm": 0.16013328731060028,
"learning_rate": 5.500285266566319e-07,
"loss": 0.4128,
"step": 1065
},
{
"epoch": 2.910580204778157,
"grad_norm": 0.15908414125442505,
"learning_rate": 5.172268723036999e-07,
"loss": 0.4256,
"step": 1066
},
{
"epoch": 2.913310580204778,
"grad_norm": 0.15130603313446045,
"learning_rate": 4.854310683050312e-07,
"loss": 0.4253,
"step": 1067
},
{
"epoch": 2.9160409556313995,
"grad_norm": 0.1521066129207611,
"learning_rate": 4.546414361416229e-07,
"loss": 0.4131,
"step": 1068
},
{
"epoch": 2.9187713310580206,
"grad_norm": 0.15544620156288147,
"learning_rate": 4.2485828712126583e-07,
"loss": 0.4088,
"step": 1069
},
{
"epoch": 2.9215017064846416,
"grad_norm": 0.1541679948568344,
"learning_rate": 3.96081922375402e-07,
"loss": 0.4083,
"step": 1070
},
{
"epoch": 2.9242320819112626,
"grad_norm": 0.15230417251586914,
"learning_rate": 3.6831263285608266e-07,
"loss": 0.4067,
"step": 1071
},
{
"epoch": 2.926962457337884,
"grad_norm": 0.15117131173610687,
"learning_rate": 3.415506993330153e-07,
"loss": 0.4138,
"step": 1072
},
{
"epoch": 2.929692832764505,
"grad_norm": 0.1522316336631775,
"learning_rate": 3.1579639239074365e-07,
"loss": 0.4052,
"step": 1073
},
{
"epoch": 2.932423208191126,
"grad_norm": 0.1582721471786499,
"learning_rate": 2.9104997242590527e-07,
"loss": 0.4056,
"step": 1074
},
{
"epoch": 2.9351535836177476,
"grad_norm": 0.1515754610300064,
"learning_rate": 2.673116896445671e-07,
"loss": 0.4054,
"step": 1075
},
{
"epoch": 2.9378839590443686,
"grad_norm": 0.15097399055957794,
"learning_rate": 2.4458178405974975e-07,
"loss": 0.4036,
"step": 1076
},
{
"epoch": 2.9406143344709896,
"grad_norm": 0.1505846381187439,
"learning_rate": 2.2286048548897376e-07,
"loss": 0.4269,
"step": 1077
},
{
"epoch": 2.9433447098976107,
"grad_norm": 0.14722640812397003,
"learning_rate": 2.0214801355192824e-07,
"loss": 0.3938,
"step": 1078
},
{
"epoch": 2.946075085324232,
"grad_norm": 0.15851718187332153,
"learning_rate": 1.824445776682504e-07,
"loss": 0.4163,
"step": 1079
},
{
"epoch": 2.948805460750853,
"grad_norm": 0.1539052575826645,
"learning_rate": 1.6375037705543826e-07,
"loss": 0.4119,
"step": 1080
},
{
"epoch": 2.9515358361774746,
"grad_norm": 0.15709060430526733,
"learning_rate": 1.4606560072679687e-07,
"loss": 0.4125,
"step": 1081
},
{
"epoch": 2.9542662116040956,
"grad_norm": 0.14519765973091125,
"learning_rate": 1.2939042748955077e-07,
"loss": 0.399,
"step": 1082
},
{
"epoch": 2.9569965870307167,
"grad_norm": 0.1516365110874176,
"learning_rate": 1.1372502594303446e-07,
"loss": 0.4181,
"step": 1083
},
{
"epoch": 2.9597269624573377,
"grad_norm": 0.15326349437236786,
"learning_rate": 9.906955447697153e-08,
"loss": 0.4174,
"step": 1084
},
{
"epoch": 2.962457337883959,
"grad_norm": 0.15291821956634521,
"learning_rate": 8.542416126989805e-08,
"loss": 0.4166,
"step": 1085
},
{
"epoch": 2.96518771331058,
"grad_norm": 0.14680640399456024,
"learning_rate": 7.27889842876417e-08,
"loss": 0.4111,
"step": 1086
},
{
"epoch": 2.967918088737201,
"grad_norm": 0.15194863080978394,
"learning_rate": 6.116415128194497e-08,
"loss": 0.4054,
"step": 1087
},
{
"epoch": 2.9706484641638227,
"grad_norm": 0.160339817404747,
"learning_rate": 5.054977978916631e-08,
"loss": 0.4133,
"step": 1088
},
{
"epoch": 2.9733788395904437,
"grad_norm": 0.15164430439472198,
"learning_rate": 4.094597712908099e-08,
"loss": 0.4134,
"step": 1089
},
{
"epoch": 2.9761092150170647,
"grad_norm": 0.14412933588027954,
"learning_rate": 3.2352840403804264e-08,
"loss": 0.3914,
"step": 1090
},
{
"epoch": 2.9788395904436857,
"grad_norm": 0.15993493795394897,
"learning_rate": 2.477045649681431e-08,
"loss": 0.4335,
"step": 1091
},
{
"epoch": 2.981569965870307,
"grad_norm": 0.1526769995689392,
"learning_rate": 1.81989020720974e-08,
"loss": 0.4129,
"step": 1092
},
{
"epoch": 2.9843003412969282,
"grad_norm": 0.14546047151088715,
"learning_rate": 1.2638243573293018e-08,
"loss": 0.3924,
"step": 1093
},
{
"epoch": 2.9870307167235497,
"grad_norm": 0.150767520070076,
"learning_rate": 8.088537223116532e-09,
"loss": 0.3982,
"step": 1094
},
{
"epoch": 2.9897610921501707,
"grad_norm": 0.15718306601047516,
"learning_rate": 4.549829022748586e-09,
"loss": 0.418,
"step": 1095
},
{
"epoch": 2.9924914675767917,
"grad_norm": 0.14539772272109985,
"learning_rate": 2.0221547513243897e-09,
"loss": 0.4034,
"step": 1096
},
{
"epoch": 2.9952218430034128,
"grad_norm": 0.15158307552337646,
"learning_rate": 5.055399656894721e-10,
"loss": 0.4038,
"step": 1097
},
{
"epoch": 2.9979522184300342,
"grad_norm": 0.148399218916893,
"learning_rate": 0.0,
"loss": 0.4121,
"step": 1098
}
],
"logging_steps": 1,
"max_steps": 1098,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.529684799263867e+17,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}