{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.236421725239617, "eval_steps": 500, "global_step": 1400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001597444089456869, "grad_norm": 128.6180419921875, "learning_rate": 0.0, "loss": 1.9747, "step": 1 }, { "epoch": 0.003194888178913738, "grad_norm": 127.44181060791016, "learning_rate": 7.628243530057963e-08, "loss": 1.9176, "step": 2 }, { "epoch": 0.004792332268370607, "grad_norm": 130.29476928710938, "learning_rate": 1.2090479941510652e-07, "loss": 1.9397, "step": 3 }, { "epoch": 0.006389776357827476, "grad_norm": 128.3740234375, "learning_rate": 1.5256487060115926e-07, "loss": 1.9046, "step": 4 }, { "epoch": 0.007987220447284345, "grad_norm": 136.06153869628906, "learning_rate": 1.7712232967084336e-07, "loss": 1.947, "step": 5 }, { "epoch": 0.009584664536741214, "grad_norm": 134.7310333251953, "learning_rate": 1.9718723471568615e-07, "loss": 1.9161, "step": 6 }, { "epoch": 0.011182108626198083, "grad_norm": 136.66058349609375, "learning_rate": 2.14151870207623e-07, "loss": 1.779, "step": 7 }, { "epoch": 0.012779552715654952, "grad_norm": 142.6785888671875, "learning_rate": 2.288473059017389e-07, "loss": 1.7793, "step": 8 }, { "epoch": 0.01437699680511182, "grad_norm": 139.76708984375, "learning_rate": 2.4180959883021303e-07, "loss": 1.7003, "step": 9 }, { "epoch": 0.01597444089456869, "grad_norm": 147.7816162109375, "learning_rate": 2.53404764971423e-07, "loss": 1.742, "step": 10 }, { "epoch": 0.01757188498402556, "grad_norm": 159.0968780517578, "learning_rate": 2.638938686254791e-07, "loss": 1.6079, "step": 11 }, { "epoch": 0.019169329073482427, "grad_norm": 138.03765869140625, "learning_rate": 2.7346967001626583e-07, "loss": 1.1859, "step": 12 }, { "epoch": 0.020766773162939296, "grad_norm": 130.79318237304688, "learning_rate": 2.8227855338279304e-07, "loss": 1.1544, "step": 13 }, { "epoch": 0.022364217252396165, "grad_norm": 128.29359436035156, "learning_rate": 2.904343055082026e-07, "loss": 1.0924, "step": 14 }, { "epoch": 0.023961661341853034, "grad_norm": 118.21534729003906, "learning_rate": 2.9802712908594985e-07, "loss": 1.0143, "step": 15 }, { "epoch": 0.025559105431309903, "grad_norm": 110.80689239501953, "learning_rate": 3.051297412023185e-07, "loss": 0.941, "step": 16 }, { "epoch": 0.027156549520766772, "grad_norm": 112.3552474975586, "learning_rate": 3.1180161973120243e-07, "loss": 0.9826, "step": 17 }, { "epoch": 0.02875399361022364, "grad_norm": 59.84565353393555, "learning_rate": 3.1809203413079266e-07, "loss": 0.6646, "step": 18 }, { "epoch": 0.03035143769968051, "grad_norm": 61.63838195800781, "learning_rate": 3.2404225570581245e-07, "loss": 0.535, "step": 19 }, { "epoch": 0.03194888178913738, "grad_norm": 57.68476486206055, "learning_rate": 3.2968720027200265e-07, "loss": 0.451, "step": 20 }, { "epoch": 0.03354632587859425, "grad_norm": 58.70195388793945, "learning_rate": 3.350566696227295e-07, "loss": 0.4398, "step": 21 }, { "epoch": 0.03514376996805112, "grad_norm": 60.67943572998047, "learning_rate": 3.401763039260588e-07, "loss": 0.3542, "step": 22 }, { "epoch": 0.036741214057507986, "grad_norm": 54.30119705200195, "learning_rate": 3.450683222410826e-07, "loss": 0.3362, "step": 23 }, { "epoch": 0.038338658146964855, "grad_norm": 27.211286544799805, "learning_rate": 3.4975210531684546e-07, "loss": 0.3206, "step": 24 }, { "epoch": 0.039936102236421724, "grad_norm": 20.16046905517578, "learning_rate": 3.542446593416867e-07, "loss": 0.2747, "step": 25 }, { "epoch": 0.04153354632587859, "grad_norm": 18.191621780395508, "learning_rate": 3.585609886833727e-07, "loss": 0.2485, "step": 26 }, { "epoch": 0.04313099041533546, "grad_norm": 17.691118240356445, "learning_rate": 3.627143982453195e-07, "loss": 0.2113, "step": 27 }, { "epoch": 0.04472843450479233, "grad_norm": 16.775856018066406, "learning_rate": 3.6671674080878223e-07, "loss": 0.2395, "step": 28 }, { "epoch": 0.0463258785942492, "grad_norm": 11.414810180664062, "learning_rate": 3.7057862095226456e-07, "loss": 0.1754, "step": 29 }, { "epoch": 0.04792332268370607, "grad_norm": 11.763614654541016, "learning_rate": 3.7430956438652954e-07, "loss": 0.171, "step": 30 }, { "epoch": 0.04952076677316294, "grad_norm": 11.317436218261719, "learning_rate": 3.7791815951345716e-07, "loss": 0.1304, "step": 31 }, { "epoch": 0.051118210862619806, "grad_norm": 11.866049766540527, "learning_rate": 3.814121765028982e-07, "loss": 0.1429, "step": 32 }, { "epoch": 0.052715654952076675, "grad_norm": 7.949328899383545, "learning_rate": 3.8479866804058563e-07, "loss": 0.1008, "step": 33 }, { "epoch": 0.054313099041533544, "grad_norm": 6.994161605834961, "learning_rate": 3.880840550317821e-07, "loss": 0.0818, "step": 34 }, { "epoch": 0.05591054313099041, "grad_norm": 8.409013748168945, "learning_rate": 3.912741998784663e-07, "loss": 0.1117, "step": 35 }, { "epoch": 0.05750798722044728, "grad_norm": 5.039752006530762, "learning_rate": 3.943744694313723e-07, "loss": 0.0769, "step": 36 }, { "epoch": 0.05910543130990415, "grad_norm": 5.305444717407227, "learning_rate": 3.973897893149772e-07, "loss": 0.0547, "step": 37 }, { "epoch": 0.06070287539936102, "grad_norm": 6.901062488555908, "learning_rate": 4.003246910063921e-07, "loss": 0.0888, "step": 38 }, { "epoch": 0.06230031948881789, "grad_norm": 5.630770683288574, "learning_rate": 4.031833527978995e-07, "loss": 0.0664, "step": 39 }, { "epoch": 0.06389776357827476, "grad_norm": 5.812962532043457, "learning_rate": 4.059696355725823e-07, "loss": 0.0643, "step": 40 }, { "epoch": 0.06549520766773163, "grad_norm": 4.684011936187744, "learning_rate": 4.0868711416176974e-07, "loss": 0.0563, "step": 41 }, { "epoch": 0.0670926517571885, "grad_norm": 5.2407050132751465, "learning_rate": 4.113391049233091e-07, "loss": 0.0546, "step": 42 }, { "epoch": 0.06869009584664537, "grad_norm": 5.062054634094238, "learning_rate": 4.139286900743784e-07, "loss": 0.0488, "step": 43 }, { "epoch": 0.07028753993610223, "grad_norm": 4.678827285766602, "learning_rate": 4.1645873922663837e-07, "loss": 0.039, "step": 44 }, { "epoch": 0.07188498402555911, "grad_norm": 4.474348068237305, "learning_rate": 4.1893192850105637e-07, "loss": 0.0327, "step": 45 }, { "epoch": 0.07348242811501597, "grad_norm": 5.580406665802002, "learning_rate": 4.213507575416622e-07, "loss": 0.0265, "step": 46 }, { "epoch": 0.07507987220447285, "grad_norm": 4.145895957946777, "learning_rate": 4.2371756469942035e-07, "loss": 0.034, "step": 47 }, { "epoch": 0.07667731629392971, "grad_norm": 2.864013671875, "learning_rate": 4.260345406174251e-07, "loss": 0.0195, "step": 48 }, { "epoch": 0.07827476038338659, "grad_norm": 3.931077718734741, "learning_rate": 4.28303740415246e-07, "loss": 0.028, "step": 49 }, { "epoch": 0.07987220447284345, "grad_norm": 3.2536861896514893, "learning_rate": 4.305270946422664e-07, "loss": 0.0167, "step": 50 }, { "epoch": 0.08146964856230032, "grad_norm": 2.361112356185913, "learning_rate": 4.3270641914630895e-07, "loss": 0.0137, "step": 51 }, { "epoch": 0.08306709265175719, "grad_norm": 3.177391529083252, "learning_rate": 4.3484342398395235e-07, "loss": 0.0168, "step": 52 }, { "epoch": 0.08466453674121406, "grad_norm": 3.647589921951294, "learning_rate": 4.3693972148208397e-07, "loss": 0.0171, "step": 53 }, { "epoch": 0.08626198083067092, "grad_norm": 3.1123926639556885, "learning_rate": 4.389968335458992e-07, "loss": 0.0096, "step": 54 }, { "epoch": 0.0878594249201278, "grad_norm": 3.218924045562744, "learning_rate": 4.4101619829632255e-07, "loss": 0.0128, "step": 55 }, { "epoch": 0.08945686900958466, "grad_norm": 3.422558307647705, "learning_rate": 4.429991761093619e-07, "loss": 0.0075, "step": 56 }, { "epoch": 0.09105431309904154, "grad_norm": 4.4138288497924805, "learning_rate": 4.44947055120919e-07, "loss": 0.0106, "step": 57 }, { "epoch": 0.0926517571884984, "grad_norm": 4.1317219734191895, "learning_rate": 4.4686105625284414e-07, "loss": 0.0147, "step": 58 }, { "epoch": 0.09424920127795527, "grad_norm": 1.8419156074523926, "learning_rate": 4.4874233780934923e-07, "loss": 0.002, "step": 59 }, { "epoch": 0.09584664536741214, "grad_norm": 5.771890640258789, "learning_rate": 4.5059199968710917e-07, "loss": 0.0132, "step": 60 }, { "epoch": 0.09744408945686901, "grad_norm": 3.2066822052001953, "learning_rate": 4.5241108723737287e-07, "loss": 0.0121, "step": 61 }, { "epoch": 0.09904153354632587, "grad_norm": 2.5449273586273193, "learning_rate": 4.5420059481403685e-07, "loss": 0.0112, "step": 62 }, { "epoch": 0.10063897763578275, "grad_norm": 1.833320140838623, "learning_rate": 4.55961469037836e-07, "loss": 0.0057, "step": 63 }, { "epoch": 0.10223642172523961, "grad_norm": 0.9052779674530029, "learning_rate": 4.576946118034778e-07, "loss": 0.0029, "step": 64 }, { "epoch": 0.10383386581469649, "grad_norm": 1.7721703052520752, "learning_rate": 4.594008830536363e-07, "loss": 0.0073, "step": 65 }, { "epoch": 0.10543130990415335, "grad_norm": 1.3743584156036377, "learning_rate": 4.6108110334116526e-07, "loss": 0.0042, "step": 66 }, { "epoch": 0.10702875399361023, "grad_norm": 5.354936599731445, "learning_rate": 4.627360561986405e-07, "loss": 0.0067, "step": 67 }, { "epoch": 0.10862619808306709, "grad_norm": 0.8433341979980469, "learning_rate": 4.6436649033236175e-07, "loss": 0.0036, "step": 68 }, { "epoch": 0.11022364217252396, "grad_norm": 1.8066205978393555, "learning_rate": 4.6597312165618915e-07, "loss": 0.0065, "step": 69 }, { "epoch": 0.11182108626198083, "grad_norm": 2.4936585426330566, "learning_rate": 4.6755663517904604e-07, "loss": 0.013, "step": 70 }, { "epoch": 0.1134185303514377, "grad_norm": 1.3142043352127075, "learning_rate": 4.691176867585419e-07, "loss": 0.0106, "step": 71 }, { "epoch": 0.11501597444089456, "grad_norm": 1.3923885822296143, "learning_rate": 4.70656904731952e-07, "loss": 0.0033, "step": 72 }, { "epoch": 0.11661341853035144, "grad_norm": 1.4490530490875244, "learning_rate": 4.721748914347038e-07, "loss": 0.0079, "step": 73 }, { "epoch": 0.1182108626198083, "grad_norm": 0.8172028660774231, "learning_rate": 4.736722246155569e-07, "loss": 0.0039, "step": 74 }, { "epoch": 0.11980830670926518, "grad_norm": 1.627395510673523, "learning_rate": 4.751494587567932e-07, "loss": 0.0038, "step": 75 }, { "epoch": 0.12140575079872204, "grad_norm": 1.3568391799926758, "learning_rate": 4.7660712630697176e-07, "loss": 0.0067, "step": 76 }, { "epoch": 0.12300319488817892, "grad_norm": 1.5196893215179443, "learning_rate": 4.780457388331021e-07, "loss": 0.0064, "step": 77 }, { "epoch": 0.12460063897763578, "grad_norm": 1.7342876195907593, "learning_rate": 4.794657880984791e-07, "loss": 0.0112, "step": 78 }, { "epoch": 0.12619808306709265, "grad_norm": 1.1238106489181519, "learning_rate": 4.808677470718594e-07, "loss": 0.008, "step": 79 }, { "epoch": 0.12779552715654952, "grad_norm": 0.6927244663238525, "learning_rate": 4.822520708731619e-07, "loss": 0.0028, "step": 80 }, { "epoch": 0.12939297124600638, "grad_norm": 0.710464596748352, "learning_rate": 4.836191976604261e-07, "loss": 0.0046, "step": 81 }, { "epoch": 0.13099041533546327, "grad_norm": 1.29885995388031, "learning_rate": 4.849695494623494e-07, "loss": 0.0028, "step": 82 }, { "epoch": 0.13258785942492013, "grad_norm": 0.7128888964653015, "learning_rate": 4.863035329603659e-07, "loss": 0.0017, "step": 83 }, { "epoch": 0.134185303514377, "grad_norm": 1.3970448970794678, "learning_rate": 4.876215402238887e-07, "loss": 0.007, "step": 84 }, { "epoch": 0.13578274760383385, "grad_norm": 1.8840289115905762, "learning_rate": 4.889239494020458e-07, "loss": 0.0073, "step": 85 }, { "epoch": 0.13738019169329074, "grad_norm": 0.9983999729156494, "learning_rate": 4.902111253749581e-07, "loss": 0.0008, "step": 86 }, { "epoch": 0.1389776357827476, "grad_norm": 1.001712441444397, "learning_rate": 4.91483420367371e-07, "loss": 0.0045, "step": 87 }, { "epoch": 0.14057507987220447, "grad_norm": 0.8022336959838867, "learning_rate": 4.927411745272181e-07, "loss": 0.005, "step": 88 }, { "epoch": 0.14217252396166133, "grad_norm": 1.9497357606887817, "learning_rate": 4.939847164714948e-07, "loss": 0.0132, "step": 89 }, { "epoch": 0.14376996805111822, "grad_norm": 0.4873954951763153, "learning_rate": 4.952143638016361e-07, "loss": 0.0036, "step": 90 }, { "epoch": 0.14536741214057508, "grad_norm": 1.6238665580749512, "learning_rate": 4.96430423590416e-07, "loss": 0.0079, "step": 91 }, { "epoch": 0.14696485623003194, "grad_norm": 2.2722995281219482, "learning_rate": 4.976331928422418e-07, "loss": 0.0112, "step": 92 }, { "epoch": 0.1485623003194888, "grad_norm": 1.2474795579910278, "learning_rate": 4.988229589285637e-07, "loss": 0.0063, "step": 93 }, { "epoch": 0.1501597444089457, "grad_norm": 1.0009785890579224, "learning_rate": 5e-07, "loss": 0.0085, "step": 94 }, { "epoch": 0.15175718849840256, "grad_norm": 0.6343518495559692, "learning_rate": 5e-07, "loss": 0.0032, "step": 95 }, { "epoch": 0.15335463258785942, "grad_norm": 0.4515174329280853, "learning_rate": 5e-07, "loss": 0.003, "step": 96 }, { "epoch": 0.15495207667731628, "grad_norm": 0.7897706627845764, "learning_rate": 5e-07, "loss": 0.0057, "step": 97 }, { "epoch": 0.15654952076677317, "grad_norm": 1.2340161800384521, "learning_rate": 5e-07, "loss": 0.006, "step": 98 }, { "epoch": 0.15814696485623003, "grad_norm": 0.5955281257629395, "learning_rate": 5e-07, "loss": 0.0043, "step": 99 }, { "epoch": 0.1597444089456869, "grad_norm": 1.140373706817627, "learning_rate": 5e-07, "loss": 0.0094, "step": 100 }, { "epoch": 0.16134185303514376, "grad_norm": 0.5656710267066956, "learning_rate": 5e-07, "loss": 0.0049, "step": 101 }, { "epoch": 0.16293929712460065, "grad_norm": 1.3329702615737915, "learning_rate": 5e-07, "loss": 0.0102, "step": 102 }, { "epoch": 0.1645367412140575, "grad_norm": 0.42556625604629517, "learning_rate": 5e-07, "loss": 0.0027, "step": 103 }, { "epoch": 0.16613418530351437, "grad_norm": 0.6244504451751709, "learning_rate": 5e-07, "loss": 0.0046, "step": 104 }, { "epoch": 0.16773162939297126, "grad_norm": 0.6920517086982727, "learning_rate": 5e-07, "loss": 0.0041, "step": 105 }, { "epoch": 0.16932907348242812, "grad_norm": 0.660217821598053, "learning_rate": 5e-07, "loss": 0.0038, "step": 106 }, { "epoch": 0.17092651757188498, "grad_norm": 1.0536694526672363, "learning_rate": 5e-07, "loss": 0.0039, "step": 107 }, { "epoch": 0.17252396166134185, "grad_norm": 0.862185537815094, "learning_rate": 5e-07, "loss": 0.005, "step": 108 }, { "epoch": 0.17412140575079874, "grad_norm": 0.33221954107284546, "learning_rate": 5e-07, "loss": 0.0027, "step": 109 }, { "epoch": 0.1757188498402556, "grad_norm": 3.744462013244629, "learning_rate": 5e-07, "loss": 0.0047, "step": 110 }, { "epoch": 0.17731629392971246, "grad_norm": 1.6666382551193237, "learning_rate": 5e-07, "loss": 0.0067, "step": 111 }, { "epoch": 0.17891373801916932, "grad_norm": 0.3559662103652954, "learning_rate": 5e-07, "loss": 0.002, "step": 112 }, { "epoch": 0.1805111821086262, "grad_norm": 0.3662293255329132, "learning_rate": 5e-07, "loss": 0.0014, "step": 113 }, { "epoch": 0.18210862619808307, "grad_norm": 0.49407243728637695, "learning_rate": 5e-07, "loss": 0.0021, "step": 114 }, { "epoch": 0.18370607028753994, "grad_norm": 1.0715949535369873, "learning_rate": 5e-07, "loss": 0.0054, "step": 115 }, { "epoch": 0.1853035143769968, "grad_norm": 0.618974506855011, "learning_rate": 5e-07, "loss": 0.0033, "step": 116 }, { "epoch": 0.1869009584664537, "grad_norm": 1.6139345169067383, "learning_rate": 5e-07, "loss": 0.0089, "step": 117 }, { "epoch": 0.18849840255591055, "grad_norm": 0.9241402745246887, "learning_rate": 5e-07, "loss": 0.0041, "step": 118 }, { "epoch": 0.1900958466453674, "grad_norm": 0.7571372389793396, "learning_rate": 5e-07, "loss": 0.0031, "step": 119 }, { "epoch": 0.19169329073482427, "grad_norm": 0.5674700736999512, "learning_rate": 5e-07, "loss": 0.0022, "step": 120 }, { "epoch": 0.19329073482428116, "grad_norm": 2.3434317111968994, "learning_rate": 5e-07, "loss": 0.0113, "step": 121 }, { "epoch": 0.19488817891373802, "grad_norm": 0.5285120010375977, "learning_rate": 5e-07, "loss": 0.0027, "step": 122 }, { "epoch": 0.1964856230031949, "grad_norm": 0.3353559076786041, "learning_rate": 5e-07, "loss": 0.0024, "step": 123 }, { "epoch": 0.19808306709265175, "grad_norm": 0.6090297698974609, "learning_rate": 5e-07, "loss": 0.0052, "step": 124 }, { "epoch": 0.19968051118210864, "grad_norm": 0.6662172079086304, "learning_rate": 5e-07, "loss": 0.0039, "step": 125 }, { "epoch": 0.2012779552715655, "grad_norm": 0.6109351515769958, "learning_rate": 5e-07, "loss": 0.0035, "step": 126 }, { "epoch": 0.20287539936102236, "grad_norm": 1.2161667346954346, "learning_rate": 5e-07, "loss": 0.0077, "step": 127 }, { "epoch": 0.20447284345047922, "grad_norm": 0.5375782251358032, "learning_rate": 5e-07, "loss": 0.003, "step": 128 }, { "epoch": 0.20607028753993611, "grad_norm": 0.8255560398101807, "learning_rate": 5e-07, "loss": 0.0046, "step": 129 }, { "epoch": 0.20766773162939298, "grad_norm": 1.2899589538574219, "learning_rate": 5e-07, "loss": 0.0073, "step": 130 }, { "epoch": 0.20926517571884984, "grad_norm": 0.4597526490688324, "learning_rate": 5e-07, "loss": 0.0019, "step": 131 }, { "epoch": 0.2108626198083067, "grad_norm": 0.2904457151889801, "learning_rate": 5e-07, "loss": 0.0014, "step": 132 }, { "epoch": 0.2124600638977636, "grad_norm": 0.2778950035572052, "learning_rate": 5e-07, "loss": 0.0022, "step": 133 }, { "epoch": 0.21405750798722045, "grad_norm": 0.45779383182525635, "learning_rate": 5e-07, "loss": 0.002, "step": 134 }, { "epoch": 0.21565495207667731, "grad_norm": 1.7134373188018799, "learning_rate": 5e-07, "loss": 0.0073, "step": 135 }, { "epoch": 0.21725239616613418, "grad_norm": 0.8295688033103943, "learning_rate": 5e-07, "loss": 0.005, "step": 136 }, { "epoch": 0.21884984025559107, "grad_norm": 1.5231693983078003, "learning_rate": 5e-07, "loss": 0.0087, "step": 137 }, { "epoch": 0.22044728434504793, "grad_norm": 0.6921976804733276, "learning_rate": 5e-07, "loss": 0.0028, "step": 138 }, { "epoch": 0.2220447284345048, "grad_norm": 0.613254964351654, "learning_rate": 5e-07, "loss": 0.002, "step": 139 }, { "epoch": 0.22364217252396165, "grad_norm": 0.9046704769134521, "learning_rate": 5e-07, "loss": 0.0035, "step": 140 }, { "epoch": 0.22523961661341854, "grad_norm": 0.827433705329895, "learning_rate": 5e-07, "loss": 0.0035, "step": 141 }, { "epoch": 0.2268370607028754, "grad_norm": 0.42438846826553345, "learning_rate": 5e-07, "loss": 0.0027, "step": 142 }, { "epoch": 0.22843450479233227, "grad_norm": 0.6912664771080017, "learning_rate": 5e-07, "loss": 0.0019, "step": 143 }, { "epoch": 0.23003194888178913, "grad_norm": 0.6359020471572876, "learning_rate": 5e-07, "loss": 0.0031, "step": 144 }, { "epoch": 0.23162939297124602, "grad_norm": 0.45392975211143494, "learning_rate": 5e-07, "loss": 0.0022, "step": 145 }, { "epoch": 0.23322683706070288, "grad_norm": 0.2925805151462555, "learning_rate": 5e-07, "loss": 0.0023, "step": 146 }, { "epoch": 0.23482428115015974, "grad_norm": 0.6216511130332947, "learning_rate": 5e-07, "loss": 0.0028, "step": 147 }, { "epoch": 0.2364217252396166, "grad_norm": 1.4257566928863525, "learning_rate": 5e-07, "loss": 0.0049, "step": 148 }, { "epoch": 0.2380191693290735, "grad_norm": 0.5238153338432312, "learning_rate": 5e-07, "loss": 0.0027, "step": 149 }, { "epoch": 0.23961661341853036, "grad_norm": 0.4455612301826477, "learning_rate": 5e-07, "loss": 0.0027, "step": 150 }, { "epoch": 0.24121405750798722, "grad_norm": 0.23266546428203583, "learning_rate": 5e-07, "loss": 0.0015, "step": 151 }, { "epoch": 0.24281150159744408, "grad_norm": 0.3799627125263214, "learning_rate": 5e-07, "loss": 0.0025, "step": 152 }, { "epoch": 0.24440894568690097, "grad_norm": 0.48497480154037476, "learning_rate": 5e-07, "loss": 0.0028, "step": 153 }, { "epoch": 0.24600638977635783, "grad_norm": 0.6147518157958984, "learning_rate": 5e-07, "loss": 0.0039, "step": 154 }, { "epoch": 0.2476038338658147, "grad_norm": 0.6456011533737183, "learning_rate": 5e-07, "loss": 0.0027, "step": 155 }, { "epoch": 0.24920127795527156, "grad_norm": 0.5384286642074585, "learning_rate": 5e-07, "loss": 0.0023, "step": 156 }, { "epoch": 0.2507987220447284, "grad_norm": 0.8868528008460999, "learning_rate": 5e-07, "loss": 0.0027, "step": 157 }, { "epoch": 0.2523961661341853, "grad_norm": 0.25091347098350525, "learning_rate": 5e-07, "loss": 0.0023, "step": 158 }, { "epoch": 0.2539936102236422, "grad_norm": 0.5789719820022583, "learning_rate": 5e-07, "loss": 0.0021, "step": 159 }, { "epoch": 0.25559105431309903, "grad_norm": 1.5586203336715698, "learning_rate": 5e-07, "loss": 0.0013, "step": 160 }, { "epoch": 0.2571884984025559, "grad_norm": 0.705531895160675, "learning_rate": 5e-07, "loss": 0.0031, "step": 161 }, { "epoch": 0.25878594249201275, "grad_norm": 0.7628393173217773, "learning_rate": 5e-07, "loss": 0.0024, "step": 162 }, { "epoch": 0.26038338658146964, "grad_norm": 0.8703659176826477, "learning_rate": 5e-07, "loss": 0.004, "step": 163 }, { "epoch": 0.26198083067092653, "grad_norm": 0.3005010783672333, "learning_rate": 5e-07, "loss": 0.0015, "step": 164 }, { "epoch": 0.26357827476038337, "grad_norm": 0.19402694702148438, "learning_rate": 5e-07, "loss": 0.0013, "step": 165 }, { "epoch": 0.26517571884984026, "grad_norm": 2.271894931793213, "learning_rate": 5e-07, "loss": 0.006, "step": 166 }, { "epoch": 0.26677316293929715, "grad_norm": 0.3555186092853546, "learning_rate": 5e-07, "loss": 0.0016, "step": 167 }, { "epoch": 0.268370607028754, "grad_norm": 1.9556201696395874, "learning_rate": 5e-07, "loss": 0.0045, "step": 168 }, { "epoch": 0.26996805111821087, "grad_norm": 1.7738531827926636, "learning_rate": 5e-07, "loss": 0.0072, "step": 169 }, { "epoch": 0.2715654952076677, "grad_norm": 2.2895989418029785, "learning_rate": 5e-07, "loss": 0.0049, "step": 170 }, { "epoch": 0.2731629392971246, "grad_norm": 1.7725859880447388, "learning_rate": 5e-07, "loss": 0.009, "step": 171 }, { "epoch": 0.2747603833865815, "grad_norm": 0.5722166895866394, "learning_rate": 5e-07, "loss": 0.0032, "step": 172 }, { "epoch": 0.2763578274760383, "grad_norm": 1.063871145248413, "learning_rate": 5e-07, "loss": 0.0016, "step": 173 }, { "epoch": 0.2779552715654952, "grad_norm": 0.3389774560928345, "learning_rate": 5e-07, "loss": 0.0026, "step": 174 }, { "epoch": 0.2795527156549521, "grad_norm": 0.45647579431533813, "learning_rate": 5e-07, "loss": 0.0022, "step": 175 }, { "epoch": 0.28115015974440893, "grad_norm": 0.7398928999900818, "learning_rate": 5e-07, "loss": 0.0031, "step": 176 }, { "epoch": 0.2827476038338658, "grad_norm": 0.7850615382194519, "learning_rate": 5e-07, "loss": 0.003, "step": 177 }, { "epoch": 0.28434504792332266, "grad_norm": 3.63692045211792, "learning_rate": 5e-07, "loss": 0.006, "step": 178 }, { "epoch": 0.28594249201277955, "grad_norm": 0.44707363843917847, "learning_rate": 5e-07, "loss": 0.0025, "step": 179 }, { "epoch": 0.28753993610223644, "grad_norm": 0.9111068248748779, "learning_rate": 5e-07, "loss": 0.0041, "step": 180 }, { "epoch": 0.28913738019169327, "grad_norm": 0.6590073704719543, "learning_rate": 5e-07, "loss": 0.0018, "step": 181 }, { "epoch": 0.29073482428115016, "grad_norm": 0.5553420782089233, "learning_rate": 5e-07, "loss": 0.0022, "step": 182 }, { "epoch": 0.29233226837060705, "grad_norm": 0.47674331068992615, "learning_rate": 5e-07, "loss": 0.0031, "step": 183 }, { "epoch": 0.2939297124600639, "grad_norm": 0.29876044392585754, "learning_rate": 5e-07, "loss": 0.0015, "step": 184 }, { "epoch": 0.2955271565495208, "grad_norm": 0.6842983365058899, "learning_rate": 5e-07, "loss": 0.0031, "step": 185 }, { "epoch": 0.2971246006389776, "grad_norm": 0.4935712516307831, "learning_rate": 5e-07, "loss": 0.0033, "step": 186 }, { "epoch": 0.2987220447284345, "grad_norm": 1.2657763957977295, "learning_rate": 5e-07, "loss": 0.0076, "step": 187 }, { "epoch": 0.3003194888178914, "grad_norm": 0.48839613795280457, "learning_rate": 5e-07, "loss": 0.0016, "step": 188 }, { "epoch": 0.3019169329073482, "grad_norm": 1.063694715499878, "learning_rate": 5e-07, "loss": 0.0037, "step": 189 }, { "epoch": 0.3035143769968051, "grad_norm": 0.7980586290359497, "learning_rate": 5e-07, "loss": 0.0023, "step": 190 }, { "epoch": 0.305111821086262, "grad_norm": 0.45484256744384766, "learning_rate": 5e-07, "loss": 0.0024, "step": 191 }, { "epoch": 0.30670926517571884, "grad_norm": 1.688433051109314, "learning_rate": 5e-07, "loss": 0.0056, "step": 192 }, { "epoch": 0.3083067092651757, "grad_norm": 0.7312079071998596, "learning_rate": 5e-07, "loss": 0.0037, "step": 193 }, { "epoch": 0.30990415335463256, "grad_norm": 0.4801974594593048, "learning_rate": 5e-07, "loss": 0.0024, "step": 194 }, { "epoch": 0.31150159744408945, "grad_norm": 0.4754345118999481, "learning_rate": 5e-07, "loss": 0.0058, "step": 195 }, { "epoch": 0.31309904153354634, "grad_norm": 2.1157238483428955, "learning_rate": 5e-07, "loss": 0.0091, "step": 196 }, { "epoch": 0.3146964856230032, "grad_norm": 0.6343661546707153, "learning_rate": 5e-07, "loss": 0.0037, "step": 197 }, { "epoch": 0.31629392971246006, "grad_norm": 0.5099918842315674, "learning_rate": 5e-07, "loss": 0.0029, "step": 198 }, { "epoch": 0.31789137380191695, "grad_norm": 0.333951860666275, "learning_rate": 5e-07, "loss": 0.0032, "step": 199 }, { "epoch": 0.3194888178913738, "grad_norm": 0.7018607258796692, "learning_rate": 5e-07, "loss": 0.0044, "step": 200 }, { "epoch": 0.3210862619808307, "grad_norm": 0.6244291067123413, "learning_rate": 5e-07, "loss": 0.0043, "step": 201 }, { "epoch": 0.3226837060702875, "grad_norm": 0.8400171399116516, "learning_rate": 5e-07, "loss": 0.0039, "step": 202 }, { "epoch": 0.3242811501597444, "grad_norm": 1.1659806966781616, "learning_rate": 5e-07, "loss": 0.0047, "step": 203 }, { "epoch": 0.3258785942492013, "grad_norm": 0.9981520175933838, "learning_rate": 5e-07, "loss": 0.0056, "step": 204 }, { "epoch": 0.3274760383386581, "grad_norm": 0.32153695821762085, "learning_rate": 5e-07, "loss": 0.0012, "step": 205 }, { "epoch": 0.329073482428115, "grad_norm": 3.8695359230041504, "learning_rate": 5e-07, "loss": 0.0028, "step": 206 }, { "epoch": 0.3306709265175719, "grad_norm": 0.6529371738433838, "learning_rate": 5e-07, "loss": 0.0033, "step": 207 }, { "epoch": 0.33226837060702874, "grad_norm": 0.7385216951370239, "learning_rate": 5e-07, "loss": 0.0048, "step": 208 }, { "epoch": 0.33386581469648563, "grad_norm": 0.2696060836315155, "learning_rate": 5e-07, "loss": 0.0016, "step": 209 }, { "epoch": 0.3354632587859425, "grad_norm": 2.1431872844696045, "learning_rate": 5e-07, "loss": 0.0104, "step": 210 }, { "epoch": 0.33706070287539935, "grad_norm": 0.45158419013023376, "learning_rate": 5e-07, "loss": 0.0029, "step": 211 }, { "epoch": 0.33865814696485624, "grad_norm": 0.28861358761787415, "learning_rate": 5e-07, "loss": 0.002, "step": 212 }, { "epoch": 0.3402555910543131, "grad_norm": 0.8536441326141357, "learning_rate": 5e-07, "loss": 0.0025, "step": 213 }, { "epoch": 0.34185303514376997, "grad_norm": 0.3299103081226349, "learning_rate": 5e-07, "loss": 0.0023, "step": 214 }, { "epoch": 0.34345047923322686, "grad_norm": 0.6404473185539246, "learning_rate": 5e-07, "loss": 0.002, "step": 215 }, { "epoch": 0.3450479233226837, "grad_norm": 0.2772029638290405, "learning_rate": 5e-07, "loss": 0.0018, "step": 216 }, { "epoch": 0.3466453674121406, "grad_norm": 3.0106074810028076, "learning_rate": 5e-07, "loss": 0.0051, "step": 217 }, { "epoch": 0.34824281150159747, "grad_norm": 1.7244184017181396, "learning_rate": 5e-07, "loss": 0.0066, "step": 218 }, { "epoch": 0.3498402555910543, "grad_norm": 0.25184711813926697, "learning_rate": 5e-07, "loss": 0.0007, "step": 219 }, { "epoch": 0.3514376996805112, "grad_norm": 0.6831691861152649, "learning_rate": 5e-07, "loss": 0.0027, "step": 220 }, { "epoch": 0.35303514376996803, "grad_norm": 0.309230774641037, "learning_rate": 5e-07, "loss": 0.0021, "step": 221 }, { "epoch": 0.3546325878594249, "grad_norm": 0.5623671412467957, "learning_rate": 5e-07, "loss": 0.0038, "step": 222 }, { "epoch": 0.3562300319488818, "grad_norm": 0.31381672620773315, "learning_rate": 5e-07, "loss": 0.0019, "step": 223 }, { "epoch": 0.35782747603833864, "grad_norm": 0.6246241927146912, "learning_rate": 5e-07, "loss": 0.0026, "step": 224 }, { "epoch": 0.35942492012779553, "grad_norm": 0.9443689584732056, "learning_rate": 5e-07, "loss": 0.0056, "step": 225 }, { "epoch": 0.3610223642172524, "grad_norm": 0.8880484104156494, "learning_rate": 5e-07, "loss": 0.0038, "step": 226 }, { "epoch": 0.36261980830670926, "grad_norm": 0.9324328899383545, "learning_rate": 5e-07, "loss": 0.0035, "step": 227 }, { "epoch": 0.36421725239616615, "grad_norm": 0.7231195569038391, "learning_rate": 5e-07, "loss": 0.0038, "step": 228 }, { "epoch": 0.365814696485623, "grad_norm": 0.5298541188240051, "learning_rate": 5e-07, "loss": 0.0047, "step": 229 }, { "epoch": 0.36741214057507987, "grad_norm": 1.5357270240783691, "learning_rate": 5e-07, "loss": 0.0054, "step": 230 }, { "epoch": 0.36900958466453676, "grad_norm": 0.4374171495437622, "learning_rate": 5e-07, "loss": 0.0026, "step": 231 }, { "epoch": 0.3706070287539936, "grad_norm": 0.2250860631465912, "learning_rate": 5e-07, "loss": 0.0025, "step": 232 }, { "epoch": 0.3722044728434505, "grad_norm": 1.8229856491088867, "learning_rate": 5e-07, "loss": 0.0026, "step": 233 }, { "epoch": 0.3738019169329074, "grad_norm": 1.0502337217330933, "learning_rate": 5e-07, "loss": 0.005, "step": 234 }, { "epoch": 0.3753993610223642, "grad_norm": 1.1811085939407349, "learning_rate": 5e-07, "loss": 0.0025, "step": 235 }, { "epoch": 0.3769968051118211, "grad_norm": 0.5202884674072266, "learning_rate": 5e-07, "loss": 0.0021, "step": 236 }, { "epoch": 0.37859424920127793, "grad_norm": 0.29556524753570557, "learning_rate": 5e-07, "loss": 0.0033, "step": 237 }, { "epoch": 0.3801916932907348, "grad_norm": 0.7253068685531616, "learning_rate": 5e-07, "loss": 0.0027, "step": 238 }, { "epoch": 0.3817891373801917, "grad_norm": 0.9315080642700195, "learning_rate": 5e-07, "loss": 0.0044, "step": 239 }, { "epoch": 0.38338658146964855, "grad_norm": 0.5284143090248108, "learning_rate": 5e-07, "loss": 0.0043, "step": 240 }, { "epoch": 0.38498402555910544, "grad_norm": 0.15872539579868317, "learning_rate": 5e-07, "loss": 0.0014, "step": 241 }, { "epoch": 0.3865814696485623, "grad_norm": 0.3570381999015808, "learning_rate": 5e-07, "loss": 0.0029, "step": 242 }, { "epoch": 0.38817891373801916, "grad_norm": 1.1983716487884521, "learning_rate": 5e-07, "loss": 0.0026, "step": 243 }, { "epoch": 0.38977635782747605, "grad_norm": 0.4511491656303406, "learning_rate": 5e-07, "loss": 0.0027, "step": 244 }, { "epoch": 0.3913738019169329, "grad_norm": 0.21462838351726532, "learning_rate": 5e-07, "loss": 0.0015, "step": 245 }, { "epoch": 0.3929712460063898, "grad_norm": 3.4422619342803955, "learning_rate": 5e-07, "loss": 0.0036, "step": 246 }, { "epoch": 0.39456869009584666, "grad_norm": 2.029860496520996, "learning_rate": 5e-07, "loss": 0.0025, "step": 247 }, { "epoch": 0.3961661341853035, "grad_norm": 0.4887297749519348, "learning_rate": 5e-07, "loss": 0.0029, "step": 248 }, { "epoch": 0.3977635782747604, "grad_norm": 3.5348706245422363, "learning_rate": 5e-07, "loss": 0.0038, "step": 249 }, { "epoch": 0.3993610223642173, "grad_norm": 0.12261940538883209, "learning_rate": 5e-07, "loss": 0.0006, "step": 250 }, { "epoch": 0.4009584664536741, "grad_norm": 1.2506657838821411, "learning_rate": 5e-07, "loss": 0.0044, "step": 251 }, { "epoch": 0.402555910543131, "grad_norm": 1.385932445526123, "learning_rate": 5e-07, "loss": 0.0071, "step": 252 }, { "epoch": 0.40415335463258784, "grad_norm": 0.6065950393676758, "learning_rate": 5e-07, "loss": 0.0029, "step": 253 }, { "epoch": 0.4057507987220447, "grad_norm": 0.3198873698711395, "learning_rate": 5e-07, "loss": 0.002, "step": 254 }, { "epoch": 0.4073482428115016, "grad_norm": 0.21003501117229462, "learning_rate": 5e-07, "loss": 0.0013, "step": 255 }, { "epoch": 0.40894568690095845, "grad_norm": 0.70599764585495, "learning_rate": 5e-07, "loss": 0.0029, "step": 256 }, { "epoch": 0.41054313099041534, "grad_norm": 0.27071261405944824, "learning_rate": 5e-07, "loss": 0.0021, "step": 257 }, { "epoch": 0.41214057507987223, "grad_norm": 0.4544181525707245, "learning_rate": 5e-07, "loss": 0.002, "step": 258 }, { "epoch": 0.41373801916932906, "grad_norm": 0.3489130437374115, "learning_rate": 5e-07, "loss": 0.0015, "step": 259 }, { "epoch": 0.41533546325878595, "grad_norm": 2.232909679412842, "learning_rate": 5e-07, "loss": 0.0035, "step": 260 }, { "epoch": 0.4169329073482428, "grad_norm": 0.24653400480747223, "learning_rate": 5e-07, "loss": 0.0013, "step": 261 }, { "epoch": 0.4185303514376997, "grad_norm": 0.70169597864151, "learning_rate": 5e-07, "loss": 0.0019, "step": 262 }, { "epoch": 0.42012779552715657, "grad_norm": 0.26590874791145325, "learning_rate": 5e-07, "loss": 0.0022, "step": 263 }, { "epoch": 0.4217252396166134, "grad_norm": 1.3357137441635132, "learning_rate": 5e-07, "loss": 0.0031, "step": 264 }, { "epoch": 0.4233226837060703, "grad_norm": 0.17348721623420715, "learning_rate": 5e-07, "loss": 0.0021, "step": 265 }, { "epoch": 0.4249201277955272, "grad_norm": 0.21971674263477325, "learning_rate": 5e-07, "loss": 0.0017, "step": 266 }, { "epoch": 0.426517571884984, "grad_norm": 0.3568759262561798, "learning_rate": 5e-07, "loss": 0.0026, "step": 267 }, { "epoch": 0.4281150159744409, "grad_norm": 0.22301608324050903, "learning_rate": 5e-07, "loss": 0.002, "step": 268 }, { "epoch": 0.42971246006389774, "grad_norm": 0.2780276834964752, "learning_rate": 5e-07, "loss": 0.0022, "step": 269 }, { "epoch": 0.43130990415335463, "grad_norm": 0.399155855178833, "learning_rate": 5e-07, "loss": 0.0022, "step": 270 }, { "epoch": 0.4329073482428115, "grad_norm": 0.35904622077941895, "learning_rate": 5e-07, "loss": 0.0022, "step": 271 }, { "epoch": 0.43450479233226835, "grad_norm": 0.14762169122695923, "learning_rate": 5e-07, "loss": 0.0015, "step": 272 }, { "epoch": 0.43610223642172524, "grad_norm": 0.26977628469467163, "learning_rate": 5e-07, "loss": 0.0016, "step": 273 }, { "epoch": 0.43769968051118213, "grad_norm": 0.9291149377822876, "learning_rate": 5e-07, "loss": 0.004, "step": 274 }, { "epoch": 0.43929712460063897, "grad_norm": 0.6625785827636719, "learning_rate": 5e-07, "loss": 0.0027, "step": 275 }, { "epoch": 0.44089456869009586, "grad_norm": 0.8231006860733032, "learning_rate": 5e-07, "loss": 0.0019, "step": 276 }, { "epoch": 0.4424920127795527, "grad_norm": 0.44783130288124084, "learning_rate": 5e-07, "loss": 0.002, "step": 277 }, { "epoch": 0.4440894568690096, "grad_norm": 0.8134899139404297, "learning_rate": 5e-07, "loss": 0.0036, "step": 278 }, { "epoch": 0.44568690095846647, "grad_norm": 0.18782544136047363, "learning_rate": 5e-07, "loss": 0.001, "step": 279 }, { "epoch": 0.4472843450479233, "grad_norm": 2.3220651149749756, "learning_rate": 5e-07, "loss": 0.0057, "step": 280 }, { "epoch": 0.4488817891373802, "grad_norm": 3.4821548461914062, "learning_rate": 5e-07, "loss": 0.0034, "step": 281 }, { "epoch": 0.4504792332268371, "grad_norm": 1.2853020429611206, "learning_rate": 5e-07, "loss": 0.0029, "step": 282 }, { "epoch": 0.4520766773162939, "grad_norm": 2.967637777328491, "learning_rate": 5e-07, "loss": 0.0045, "step": 283 }, { "epoch": 0.4536741214057508, "grad_norm": 1.1847063302993774, "learning_rate": 5e-07, "loss": 0.0066, "step": 284 }, { "epoch": 0.45527156549520764, "grad_norm": 0.6564436554908752, "learning_rate": 5e-07, "loss": 0.0025, "step": 285 }, { "epoch": 0.45686900958466453, "grad_norm": 1.7592343091964722, "learning_rate": 5e-07, "loss": 0.0043, "step": 286 }, { "epoch": 0.4584664536741214, "grad_norm": 0.21900290250778198, "learning_rate": 5e-07, "loss": 0.0025, "step": 287 }, { "epoch": 0.46006389776357826, "grad_norm": 0.31253165006637573, "learning_rate": 5e-07, "loss": 0.0025, "step": 288 }, { "epoch": 0.46166134185303515, "grad_norm": 0.7273770570755005, "learning_rate": 5e-07, "loss": 0.004, "step": 289 }, { "epoch": 0.46325878594249204, "grad_norm": 0.9572885036468506, "learning_rate": 5e-07, "loss": 0.005, "step": 290 }, { "epoch": 0.46485623003194887, "grad_norm": 4.609229564666748, "learning_rate": 5e-07, "loss": 0.0105, "step": 291 }, { "epoch": 0.46645367412140576, "grad_norm": 0.7994480729103088, "learning_rate": 5e-07, "loss": 0.0026, "step": 292 }, { "epoch": 0.4680511182108626, "grad_norm": 0.512992262840271, "learning_rate": 5e-07, "loss": 0.0025, "step": 293 }, { "epoch": 0.4696485623003195, "grad_norm": 0.7132886052131653, "learning_rate": 5e-07, "loss": 0.0029, "step": 294 }, { "epoch": 0.4712460063897764, "grad_norm": 0.4453645348548889, "learning_rate": 5e-07, "loss": 0.0025, "step": 295 }, { "epoch": 0.4728434504792332, "grad_norm": 0.32878679037094116, "learning_rate": 5e-07, "loss": 0.0028, "step": 296 }, { "epoch": 0.4744408945686901, "grad_norm": 0.4557294249534607, "learning_rate": 5e-07, "loss": 0.0029, "step": 297 }, { "epoch": 0.476038338658147, "grad_norm": 0.7930683493614197, "learning_rate": 5e-07, "loss": 0.0028, "step": 298 }, { "epoch": 0.4776357827476038, "grad_norm": 0.16142797470092773, "learning_rate": 5e-07, "loss": 0.0024, "step": 299 }, { "epoch": 0.4792332268370607, "grad_norm": 0.16838717460632324, "learning_rate": 5e-07, "loss": 0.0017, "step": 300 }, { "epoch": 0.48083067092651754, "grad_norm": 1.0744272470474243, "learning_rate": 5e-07, "loss": 0.0032, "step": 301 }, { "epoch": 0.48242811501597443, "grad_norm": 1.177935004234314, "learning_rate": 5e-07, "loss": 0.0043, "step": 302 }, { "epoch": 0.4840255591054313, "grad_norm": 0.33150434494018555, "learning_rate": 5e-07, "loss": 0.0018, "step": 303 }, { "epoch": 0.48562300319488816, "grad_norm": 0.14007702469825745, "learning_rate": 5e-07, "loss": 0.0023, "step": 304 }, { "epoch": 0.48722044728434505, "grad_norm": 0.23931963741779327, "learning_rate": 5e-07, "loss": 0.0031, "step": 305 }, { "epoch": 0.48881789137380194, "grad_norm": 1.721514344215393, "learning_rate": 5e-07, "loss": 0.0025, "step": 306 }, { "epoch": 0.4904153354632588, "grad_norm": 1.3192071914672852, "learning_rate": 5e-07, "loss": 0.0095, "step": 307 }, { "epoch": 0.49201277955271566, "grad_norm": 6.429077625274658, "learning_rate": 5e-07, "loss": 0.011, "step": 308 }, { "epoch": 0.4936102236421725, "grad_norm": 0.919262707233429, "learning_rate": 5e-07, "loss": 0.0066, "step": 309 }, { "epoch": 0.4952076677316294, "grad_norm": 0.6547032594680786, "learning_rate": 5e-07, "loss": 0.0017, "step": 310 }, { "epoch": 0.4968051118210863, "grad_norm": 0.487663209438324, "learning_rate": 5e-07, "loss": 0.0033, "step": 311 }, { "epoch": 0.4984025559105431, "grad_norm": 0.33844229578971863, "learning_rate": 5e-07, "loss": 0.0039, "step": 312 }, { "epoch": 0.5, "grad_norm": 0.686505138874054, "learning_rate": 5e-07, "loss": 0.0035, "step": 313 }, { "epoch": 0.5015974440894568, "grad_norm": 0.2250794619321823, "learning_rate": 5e-07, "loss": 0.0021, "step": 314 }, { "epoch": 0.5031948881789138, "grad_norm": 0.40359318256378174, "learning_rate": 5e-07, "loss": 0.0019, "step": 315 }, { "epoch": 0.5047923322683706, "grad_norm": 0.3062373697757721, "learning_rate": 5e-07, "loss": 0.0027, "step": 316 }, { "epoch": 0.5063897763578274, "grad_norm": 0.3083527386188507, "learning_rate": 5e-07, "loss": 0.0021, "step": 317 }, { "epoch": 0.5079872204472844, "grad_norm": 0.4790252447128296, "learning_rate": 5e-07, "loss": 0.0045, "step": 318 }, { "epoch": 0.5095846645367412, "grad_norm": 0.4106828570365906, "learning_rate": 5e-07, "loss": 0.0045, "step": 319 }, { "epoch": 0.5111821086261981, "grad_norm": 0.5916593670845032, "learning_rate": 5e-07, "loss": 0.0021, "step": 320 }, { "epoch": 0.512779552715655, "grad_norm": 0.29794517159461975, "learning_rate": 5e-07, "loss": 0.0011, "step": 321 }, { "epoch": 0.5143769968051118, "grad_norm": 0.3094945251941681, "learning_rate": 5e-07, "loss": 0.004, "step": 322 }, { "epoch": 0.5159744408945687, "grad_norm": 0.8037744760513306, "learning_rate": 5e-07, "loss": 0.0035, "step": 323 }, { "epoch": 0.5175718849840255, "grad_norm": 1.0918713808059692, "learning_rate": 5e-07, "loss": 0.0035, "step": 324 }, { "epoch": 0.5191693290734825, "grad_norm": 0.1457972228527069, "learning_rate": 5e-07, "loss": 0.001, "step": 325 }, { "epoch": 0.5207667731629393, "grad_norm": 0.9666080474853516, "learning_rate": 5e-07, "loss": 0.003, "step": 326 }, { "epoch": 0.5223642172523961, "grad_norm": 0.5684059262275696, "learning_rate": 5e-07, "loss": 0.0023, "step": 327 }, { "epoch": 0.5239616613418531, "grad_norm": 0.5098831057548523, "learning_rate": 5e-07, "loss": 0.0034, "step": 328 }, { "epoch": 0.5255591054313099, "grad_norm": 0.23925842344760895, "learning_rate": 5e-07, "loss": 0.001, "step": 329 }, { "epoch": 0.5271565495207667, "grad_norm": 0.17939390242099762, "learning_rate": 5e-07, "loss": 0.0021, "step": 330 }, { "epoch": 0.5287539936102237, "grad_norm": 0.16178075969219208, "learning_rate": 5e-07, "loss": 0.0017, "step": 331 }, { "epoch": 0.5303514376996805, "grad_norm": 0.671220064163208, "learning_rate": 5e-07, "loss": 0.0046, "step": 332 }, { "epoch": 0.5319488817891374, "grad_norm": 0.816387414932251, "learning_rate": 5e-07, "loss": 0.0021, "step": 333 }, { "epoch": 0.5335463258785943, "grad_norm": 0.24421027302742004, "learning_rate": 5e-07, "loss": 0.0019, "step": 334 }, { "epoch": 0.5351437699680511, "grad_norm": 0.557515025138855, "learning_rate": 5e-07, "loss": 0.0027, "step": 335 }, { "epoch": 0.536741214057508, "grad_norm": 1.0557501316070557, "learning_rate": 5e-07, "loss": 0.002, "step": 336 }, { "epoch": 0.5383386581469649, "grad_norm": 0.959538459777832, "learning_rate": 5e-07, "loss": 0.0033, "step": 337 }, { "epoch": 0.5399361022364217, "grad_norm": 0.3274182677268982, "learning_rate": 5e-07, "loss": 0.0027, "step": 338 }, { "epoch": 0.5415335463258786, "grad_norm": 0.6471118927001953, "learning_rate": 5e-07, "loss": 0.0011, "step": 339 }, { "epoch": 0.5431309904153354, "grad_norm": 0.38304996490478516, "learning_rate": 5e-07, "loss": 0.0028, "step": 340 }, { "epoch": 0.5447284345047924, "grad_norm": 1.0141305923461914, "learning_rate": 5e-07, "loss": 0.0021, "step": 341 }, { "epoch": 0.5463258785942492, "grad_norm": 0.9558109641075134, "learning_rate": 5e-07, "loss": 0.0017, "step": 342 }, { "epoch": 0.547923322683706, "grad_norm": 0.20122268795967102, "learning_rate": 5e-07, "loss": 0.0013, "step": 343 }, { "epoch": 0.549520766773163, "grad_norm": 0.4421844780445099, "learning_rate": 5e-07, "loss": 0.0037, "step": 344 }, { "epoch": 0.5511182108626198, "grad_norm": 0.4365698993206024, "learning_rate": 5e-07, "loss": 0.0032, "step": 345 }, { "epoch": 0.5527156549520766, "grad_norm": 1.578316330909729, "learning_rate": 5e-07, "loss": 0.0053, "step": 346 }, { "epoch": 0.5543130990415336, "grad_norm": 0.21290868520736694, "learning_rate": 5e-07, "loss": 0.0017, "step": 347 }, { "epoch": 0.5559105431309904, "grad_norm": 0.6181979179382324, "learning_rate": 5e-07, "loss": 0.0022, "step": 348 }, { "epoch": 0.5575079872204473, "grad_norm": 0.20121750235557556, "learning_rate": 5e-07, "loss": 0.0007, "step": 349 }, { "epoch": 0.5591054313099042, "grad_norm": 0.5325523018836975, "learning_rate": 5e-07, "loss": 0.0036, "step": 350 }, { "epoch": 0.560702875399361, "grad_norm": 0.6079647541046143, "learning_rate": 5e-07, "loss": 0.0052, "step": 351 }, { "epoch": 0.5623003194888179, "grad_norm": 0.2826935946941376, "learning_rate": 5e-07, "loss": 0.0016, "step": 352 }, { "epoch": 0.5638977635782748, "grad_norm": 0.15892595052719116, "learning_rate": 5e-07, "loss": 0.0011, "step": 353 }, { "epoch": 0.5654952076677316, "grad_norm": 0.3881969749927521, "learning_rate": 5e-07, "loss": 0.0021, "step": 354 }, { "epoch": 0.5670926517571885, "grad_norm": 0.3390671908855438, "learning_rate": 5e-07, "loss": 0.0032, "step": 355 }, { "epoch": 0.5686900958466453, "grad_norm": 0.2904258668422699, "learning_rate": 5e-07, "loss": 0.0018, "step": 356 }, { "epoch": 0.5702875399361023, "grad_norm": 0.46645137667655945, "learning_rate": 5e-07, "loss": 0.0035, "step": 357 }, { "epoch": 0.5718849840255591, "grad_norm": 0.6814522743225098, "learning_rate": 5e-07, "loss": 0.0034, "step": 358 }, { "epoch": 0.5734824281150159, "grad_norm": 0.9146270751953125, "learning_rate": 5e-07, "loss": 0.0044, "step": 359 }, { "epoch": 0.5750798722044729, "grad_norm": 0.21922031044960022, "learning_rate": 5e-07, "loss": 0.0009, "step": 360 }, { "epoch": 0.5766773162939297, "grad_norm": 0.939836859703064, "learning_rate": 5e-07, "loss": 0.0039, "step": 361 }, { "epoch": 0.5782747603833865, "grad_norm": 0.6049763560295105, "learning_rate": 5e-07, "loss": 0.003, "step": 362 }, { "epoch": 0.5798722044728435, "grad_norm": 0.25285854935646057, "learning_rate": 5e-07, "loss": 0.0029, "step": 363 }, { "epoch": 0.5814696485623003, "grad_norm": 0.3022539019584656, "learning_rate": 5e-07, "loss": 0.0029, "step": 364 }, { "epoch": 0.5830670926517572, "grad_norm": 0.25078529119491577, "learning_rate": 5e-07, "loss": 0.0028, "step": 365 }, { "epoch": 0.5846645367412141, "grad_norm": 0.3507131040096283, "learning_rate": 5e-07, "loss": 0.0043, "step": 366 }, { "epoch": 0.5862619808306709, "grad_norm": 0.6294477581977844, "learning_rate": 5e-07, "loss": 0.0058, "step": 367 }, { "epoch": 0.5878594249201278, "grad_norm": 0.48274868726730347, "learning_rate": 5e-07, "loss": 0.0029, "step": 368 }, { "epoch": 0.5894568690095847, "grad_norm": 0.39043542742729187, "learning_rate": 5e-07, "loss": 0.0033, "step": 369 }, { "epoch": 0.5910543130990416, "grad_norm": 0.19649553298950195, "learning_rate": 5e-07, "loss": 0.0018, "step": 370 }, { "epoch": 0.5926517571884984, "grad_norm": 0.12885907292366028, "learning_rate": 5e-07, "loss": 0.0023, "step": 371 }, { "epoch": 0.5942492012779552, "grad_norm": 0.1488872468471527, "learning_rate": 5e-07, "loss": 0.0026, "step": 372 }, { "epoch": 0.5958466453674122, "grad_norm": 0.3566787540912628, "learning_rate": 5e-07, "loss": 0.0022, "step": 373 }, { "epoch": 0.597444089456869, "grad_norm": 0.986865758895874, "learning_rate": 5e-07, "loss": 0.0047, "step": 374 }, { "epoch": 0.5990415335463258, "grad_norm": 0.1579206883907318, "learning_rate": 5e-07, "loss": 0.0011, "step": 375 }, { "epoch": 0.6006389776357828, "grad_norm": 0.13746462762355804, "learning_rate": 5e-07, "loss": 0.0021, "step": 376 }, { "epoch": 0.6022364217252396, "grad_norm": 0.25995907187461853, "learning_rate": 5e-07, "loss": 0.0032, "step": 377 }, { "epoch": 0.6038338658146964, "grad_norm": 0.38791167736053467, "learning_rate": 5e-07, "loss": 0.0015, "step": 378 }, { "epoch": 0.6054313099041534, "grad_norm": 1.6591804027557373, "learning_rate": 5e-07, "loss": 0.0034, "step": 379 }, { "epoch": 0.6070287539936102, "grad_norm": 0.11244110763072968, "learning_rate": 5e-07, "loss": 0.0015, "step": 380 }, { "epoch": 0.6086261980830671, "grad_norm": 0.28090763092041016, "learning_rate": 5e-07, "loss": 0.0012, "step": 381 }, { "epoch": 0.610223642172524, "grad_norm": 0.8654282689094543, "learning_rate": 5e-07, "loss": 0.0032, "step": 382 }, { "epoch": 0.6118210862619808, "grad_norm": 0.3197280764579773, "learning_rate": 5e-07, "loss": 0.0025, "step": 383 }, { "epoch": 0.6134185303514377, "grad_norm": 0.13376931846141815, "learning_rate": 5e-07, "loss": 0.0018, "step": 384 }, { "epoch": 0.6150159744408946, "grad_norm": 0.3771628439426422, "learning_rate": 5e-07, "loss": 0.0025, "step": 385 }, { "epoch": 0.6166134185303515, "grad_norm": 3.0465352535247803, "learning_rate": 5e-07, "loss": 0.0057, "step": 386 }, { "epoch": 0.6182108626198083, "grad_norm": 0.6134064793586731, "learning_rate": 5e-07, "loss": 0.0045, "step": 387 }, { "epoch": 0.6198083067092651, "grad_norm": 1.984861969947815, "learning_rate": 5e-07, "loss": 0.0021, "step": 388 }, { "epoch": 0.6214057507987221, "grad_norm": 0.7715848684310913, "learning_rate": 5e-07, "loss": 0.0034, "step": 389 }, { "epoch": 0.6230031948881789, "grad_norm": 1.4465832710266113, "learning_rate": 5e-07, "loss": 0.0011, "step": 390 }, { "epoch": 0.6246006389776357, "grad_norm": 0.4437814950942993, "learning_rate": 5e-07, "loss": 0.0021, "step": 391 }, { "epoch": 0.6261980830670927, "grad_norm": 0.7803134322166443, "learning_rate": 5e-07, "loss": 0.003, "step": 392 }, { "epoch": 0.6277955271565495, "grad_norm": 0.4495057761669159, "learning_rate": 5e-07, "loss": 0.0014, "step": 393 }, { "epoch": 0.6293929712460063, "grad_norm": 0.9235338568687439, "learning_rate": 5e-07, "loss": 0.0062, "step": 394 }, { "epoch": 0.6309904153354633, "grad_norm": 1.1879374980926514, "learning_rate": 5e-07, "loss": 0.0046, "step": 395 }, { "epoch": 0.6325878594249201, "grad_norm": 0.22155530750751495, "learning_rate": 5e-07, "loss": 0.0015, "step": 396 }, { "epoch": 0.634185303514377, "grad_norm": 0.6639124155044556, "learning_rate": 5e-07, "loss": 0.0031, "step": 397 }, { "epoch": 0.6357827476038339, "grad_norm": 0.24812071025371552, "learning_rate": 5e-07, "loss": 0.0014, "step": 398 }, { "epoch": 0.6373801916932907, "grad_norm": 0.24182091653347015, "learning_rate": 5e-07, "loss": 0.0028, "step": 399 }, { "epoch": 0.6389776357827476, "grad_norm": 0.25194039940834045, "learning_rate": 5e-07, "loss": 0.0012, "step": 400 }, { "epoch": 0.6405750798722045, "grad_norm": 1.6056197881698608, "learning_rate": 5e-07, "loss": 0.006, "step": 401 }, { "epoch": 0.6421725239616614, "grad_norm": 3.634866952896118, "learning_rate": 5e-07, "loss": 0.0048, "step": 402 }, { "epoch": 0.6437699680511182, "grad_norm": 1.0231901407241821, "learning_rate": 5e-07, "loss": 0.0038, "step": 403 }, { "epoch": 0.645367412140575, "grad_norm": 0.27629023790359497, "learning_rate": 5e-07, "loss": 0.0026, "step": 404 }, { "epoch": 0.646964856230032, "grad_norm": 0.1949998289346695, "learning_rate": 5e-07, "loss": 0.0019, "step": 405 }, { "epoch": 0.6485623003194888, "grad_norm": 0.34071701765060425, "learning_rate": 5e-07, "loss": 0.002, "step": 406 }, { "epoch": 0.6501597444089456, "grad_norm": 0.21671850979328156, "learning_rate": 5e-07, "loss": 0.0011, "step": 407 }, { "epoch": 0.6517571884984026, "grad_norm": 0.27686911821365356, "learning_rate": 5e-07, "loss": 0.0014, "step": 408 }, { "epoch": 0.6533546325878594, "grad_norm": 0.3393467366695404, "learning_rate": 5e-07, "loss": 0.0022, "step": 409 }, { "epoch": 0.6549520766773163, "grad_norm": 0.5887823700904846, "learning_rate": 5e-07, "loss": 0.004, "step": 410 }, { "epoch": 0.6565495207667732, "grad_norm": 0.22454427182674408, "learning_rate": 5e-07, "loss": 0.0023, "step": 411 }, { "epoch": 0.65814696485623, "grad_norm": 0.13433396816253662, "learning_rate": 5e-07, "loss": 0.0024, "step": 412 }, { "epoch": 0.6597444089456869, "grad_norm": 0.14173242449760437, "learning_rate": 5e-07, "loss": 0.0022, "step": 413 }, { "epoch": 0.6613418530351438, "grad_norm": 0.30982744693756104, "learning_rate": 5e-07, "loss": 0.0027, "step": 414 }, { "epoch": 0.6629392971246006, "grad_norm": 0.47411805391311646, "learning_rate": 5e-07, "loss": 0.004, "step": 415 }, { "epoch": 0.6645367412140575, "grad_norm": 0.19577734172344208, "learning_rate": 5e-07, "loss": 0.0029, "step": 416 }, { "epoch": 0.6661341853035144, "grad_norm": 0.6344949007034302, "learning_rate": 5e-07, "loss": 0.0032, "step": 417 }, { "epoch": 0.6677316293929713, "grad_norm": 0.45841068029403687, "learning_rate": 5e-07, "loss": 0.0033, "step": 418 }, { "epoch": 0.6693290734824281, "grad_norm": 0.4367104768753052, "learning_rate": 5e-07, "loss": 0.0022, "step": 419 }, { "epoch": 0.670926517571885, "grad_norm": 0.363915354013443, "learning_rate": 5e-07, "loss": 0.0031, "step": 420 }, { "epoch": 0.6725239616613419, "grad_norm": 0.18759390711784363, "learning_rate": 5e-07, "loss": 0.0018, "step": 421 }, { "epoch": 0.6741214057507987, "grad_norm": 0.15988245606422424, "learning_rate": 5e-07, "loss": 0.0029, "step": 422 }, { "epoch": 0.6757188498402555, "grad_norm": 0.1806800663471222, "learning_rate": 5e-07, "loss": 0.0032, "step": 423 }, { "epoch": 0.6773162939297125, "grad_norm": 0.21876202523708344, "learning_rate": 5e-07, "loss": 0.0024, "step": 424 }, { "epoch": 0.6789137380191693, "grad_norm": 0.13135991990566254, "learning_rate": 5e-07, "loss": 0.0005, "step": 425 }, { "epoch": 0.6805111821086262, "grad_norm": 1.3671129941940308, "learning_rate": 5e-07, "loss": 0.0039, "step": 426 }, { "epoch": 0.6821086261980831, "grad_norm": 0.1489122062921524, "learning_rate": 5e-07, "loss": 0.0019, "step": 427 }, { "epoch": 0.6837060702875399, "grad_norm": 0.17866690456867218, "learning_rate": 5e-07, "loss": 0.0014, "step": 428 }, { "epoch": 0.6853035143769968, "grad_norm": 0.655277669429779, "learning_rate": 5e-07, "loss": 0.0045, "step": 429 }, { "epoch": 0.6869009584664537, "grad_norm": 0.17315036058425903, "learning_rate": 5e-07, "loss": 0.0017, "step": 430 }, { "epoch": 0.6884984025559105, "grad_norm": 0.20632481575012207, "learning_rate": 5e-07, "loss": 0.0019, "step": 431 }, { "epoch": 0.6900958466453674, "grad_norm": 0.2345188409090042, "learning_rate": 5e-07, "loss": 0.0033, "step": 432 }, { "epoch": 0.6916932907348243, "grad_norm": 3.234980583190918, "learning_rate": 5e-07, "loss": 0.0033, "step": 433 }, { "epoch": 0.6932907348242812, "grad_norm": 0.5852277874946594, "learning_rate": 5e-07, "loss": 0.0028, "step": 434 }, { "epoch": 0.694888178913738, "grad_norm": 2.7778074741363525, "learning_rate": 5e-07, "loss": 0.0047, "step": 435 }, { "epoch": 0.6964856230031949, "grad_norm": 0.13274647295475006, "learning_rate": 5e-07, "loss": 0.0013, "step": 436 }, { "epoch": 0.6980830670926518, "grad_norm": 0.3051118850708008, "learning_rate": 5e-07, "loss": 0.0011, "step": 437 }, { "epoch": 0.6996805111821086, "grad_norm": 0.1823217123746872, "learning_rate": 5e-07, "loss": 0.0034, "step": 438 }, { "epoch": 0.7012779552715654, "grad_norm": 0.24809595942497253, "learning_rate": 5e-07, "loss": 0.0017, "step": 439 }, { "epoch": 0.7028753993610224, "grad_norm": 0.191731795668602, "learning_rate": 5e-07, "loss": 0.0019, "step": 440 }, { "epoch": 0.7044728434504792, "grad_norm": 0.32459452748298645, "learning_rate": 5e-07, "loss": 0.0025, "step": 441 }, { "epoch": 0.7060702875399361, "grad_norm": 0.46944916248321533, "learning_rate": 5e-07, "loss": 0.0037, "step": 442 }, { "epoch": 0.707667731629393, "grad_norm": 0.28335535526275635, "learning_rate": 5e-07, "loss": 0.0021, "step": 443 }, { "epoch": 0.7092651757188498, "grad_norm": 0.4613920748233795, "learning_rate": 5e-07, "loss": 0.0022, "step": 444 }, { "epoch": 0.7108626198083067, "grad_norm": 0.25062379240989685, "learning_rate": 5e-07, "loss": 0.0037, "step": 445 }, { "epoch": 0.7124600638977636, "grad_norm": 0.18467797338962555, "learning_rate": 5e-07, "loss": 0.0011, "step": 446 }, { "epoch": 0.7140575079872205, "grad_norm": 0.11972685158252716, "learning_rate": 5e-07, "loss": 0.0007, "step": 447 }, { "epoch": 0.7156549520766773, "grad_norm": 0.42692673206329346, "learning_rate": 5e-07, "loss": 0.0019, "step": 448 }, { "epoch": 0.7172523961661342, "grad_norm": 0.24261529743671417, "learning_rate": 5e-07, "loss": 0.0036, "step": 449 }, { "epoch": 0.7188498402555911, "grad_norm": 0.5654745101928711, "learning_rate": 5e-07, "loss": 0.0026, "step": 450 }, { "epoch": 0.7204472843450479, "grad_norm": 1.1742912530899048, "learning_rate": 5e-07, "loss": 0.005, "step": 451 }, { "epoch": 0.7220447284345048, "grad_norm": 0.4380282163619995, "learning_rate": 5e-07, "loss": 0.0032, "step": 452 }, { "epoch": 0.7236421725239617, "grad_norm": 0.10869846493005753, "learning_rate": 5e-07, "loss": 0.0019, "step": 453 }, { "epoch": 0.7252396166134185, "grad_norm": 0.24093355238437653, "learning_rate": 5e-07, "loss": 0.0018, "step": 454 }, { "epoch": 0.7268370607028753, "grad_norm": 0.10973268747329712, "learning_rate": 5e-07, "loss": 0.0018, "step": 455 }, { "epoch": 0.7284345047923323, "grad_norm": 0.3785342276096344, "learning_rate": 5e-07, "loss": 0.0018, "step": 456 }, { "epoch": 0.7300319488817891, "grad_norm": 0.9914066791534424, "learning_rate": 5e-07, "loss": 0.0018, "step": 457 }, { "epoch": 0.731629392971246, "grad_norm": 0.4453500211238861, "learning_rate": 5e-07, "loss": 0.0019, "step": 458 }, { "epoch": 0.7332268370607029, "grad_norm": 0.2674676477909088, "learning_rate": 5e-07, "loss": 0.0027, "step": 459 }, { "epoch": 0.7348242811501597, "grad_norm": 0.9725813269615173, "learning_rate": 5e-07, "loss": 0.0039, "step": 460 }, { "epoch": 0.7364217252396166, "grad_norm": 0.14569059014320374, "learning_rate": 5e-07, "loss": 0.0022, "step": 461 }, { "epoch": 0.7380191693290735, "grad_norm": 0.25622859597206116, "learning_rate": 5e-07, "loss": 0.0016, "step": 462 }, { "epoch": 0.7396166134185304, "grad_norm": 0.33373990654945374, "learning_rate": 5e-07, "loss": 0.0012, "step": 463 }, { "epoch": 0.7412140575079872, "grad_norm": 0.28780269622802734, "learning_rate": 5e-07, "loss": 0.0022, "step": 464 }, { "epoch": 0.7428115015974441, "grad_norm": 0.3068907856941223, "learning_rate": 5e-07, "loss": 0.0028, "step": 465 }, { "epoch": 0.744408945686901, "grad_norm": 0.1721941977739334, "learning_rate": 5e-07, "loss": 0.0025, "step": 466 }, { "epoch": 0.7460063897763578, "grad_norm": 0.08544081449508667, "learning_rate": 5e-07, "loss": 0.0014, "step": 467 }, { "epoch": 0.7476038338658147, "grad_norm": 0.5271486043930054, "learning_rate": 5e-07, "loss": 0.0021, "step": 468 }, { "epoch": 0.7492012779552716, "grad_norm": 1.1334803104400635, "learning_rate": 5e-07, "loss": 0.0045, "step": 469 }, { "epoch": 0.7507987220447284, "grad_norm": 0.7498302459716797, "learning_rate": 5e-07, "loss": 0.0023, "step": 470 }, { "epoch": 0.7523961661341853, "grad_norm": 0.2573578357696533, "learning_rate": 5e-07, "loss": 0.0019, "step": 471 }, { "epoch": 0.7539936102236422, "grad_norm": 0.658050537109375, "learning_rate": 5e-07, "loss": 0.0023, "step": 472 }, { "epoch": 0.755591054313099, "grad_norm": 0.9222865104675293, "learning_rate": 5e-07, "loss": 0.0021, "step": 473 }, { "epoch": 0.7571884984025559, "grad_norm": 0.15267683565616608, "learning_rate": 5e-07, "loss": 0.0008, "step": 474 }, { "epoch": 0.7587859424920128, "grad_norm": 1.6186271905899048, "learning_rate": 5e-07, "loss": 0.0042, "step": 475 }, { "epoch": 0.7603833865814696, "grad_norm": 0.27657464146614075, "learning_rate": 5e-07, "loss": 0.0012, "step": 476 }, { "epoch": 0.7619808306709265, "grad_norm": 1.7262461185455322, "learning_rate": 5e-07, "loss": 0.0028, "step": 477 }, { "epoch": 0.7635782747603834, "grad_norm": 0.2438318431377411, "learning_rate": 5e-07, "loss": 0.0025, "step": 478 }, { "epoch": 0.7651757188498403, "grad_norm": 1.9847776889801025, "learning_rate": 5e-07, "loss": 0.0029, "step": 479 }, { "epoch": 0.7667731629392971, "grad_norm": 1.081231713294983, "learning_rate": 5e-07, "loss": 0.0019, "step": 480 }, { "epoch": 0.768370607028754, "grad_norm": 0.23922041058540344, "learning_rate": 5e-07, "loss": 0.0022, "step": 481 }, { "epoch": 0.7699680511182109, "grad_norm": 0.9737673401832581, "learning_rate": 5e-07, "loss": 0.0065, "step": 482 }, { "epoch": 0.7715654952076677, "grad_norm": 0.14736409485340118, "learning_rate": 5e-07, "loss": 0.0025, "step": 483 }, { "epoch": 0.7731629392971247, "grad_norm": 0.2849075198173523, "learning_rate": 5e-07, "loss": 0.0034, "step": 484 }, { "epoch": 0.7747603833865815, "grad_norm": 0.7391708493232727, "learning_rate": 5e-07, "loss": 0.0033, "step": 485 }, { "epoch": 0.7763578274760383, "grad_norm": 0.24760426580905914, "learning_rate": 5e-07, "loss": 0.0033, "step": 486 }, { "epoch": 0.7779552715654952, "grad_norm": 0.1888248175382614, "learning_rate": 5e-07, "loss": 0.0026, "step": 487 }, { "epoch": 0.7795527156549521, "grad_norm": 0.27008160948753357, "learning_rate": 5e-07, "loss": 0.0032, "step": 488 }, { "epoch": 0.7811501597444089, "grad_norm": 0.17157332599163055, "learning_rate": 5e-07, "loss": 0.0013, "step": 489 }, { "epoch": 0.7827476038338658, "grad_norm": 0.2907833755016327, "learning_rate": 5e-07, "loss": 0.0014, "step": 490 }, { "epoch": 0.7843450479233227, "grad_norm": 0.4900013506412506, "learning_rate": 5e-07, "loss": 0.002, "step": 491 }, { "epoch": 0.7859424920127795, "grad_norm": 0.1433720886707306, "learning_rate": 5e-07, "loss": 0.0019, "step": 492 }, { "epoch": 0.7875399361022364, "grad_norm": 0.09725961834192276, "learning_rate": 5e-07, "loss": 0.0008, "step": 493 }, { "epoch": 0.7891373801916933, "grad_norm": 0.8984055519104004, "learning_rate": 5e-07, "loss": 0.0048, "step": 494 }, { "epoch": 0.7907348242811502, "grad_norm": 0.22446835041046143, "learning_rate": 5e-07, "loss": 0.0019, "step": 495 }, { "epoch": 0.792332268370607, "grad_norm": 0.2916003167629242, "learning_rate": 5e-07, "loss": 0.0025, "step": 496 }, { "epoch": 0.7939297124600639, "grad_norm": 0.42406606674194336, "learning_rate": 5e-07, "loss": 0.0022, "step": 497 }, { "epoch": 0.7955271565495208, "grad_norm": 0.8635863661766052, "learning_rate": 5e-07, "loss": 0.0029, "step": 498 }, { "epoch": 0.7971246006389776, "grad_norm": 0.4940405488014221, "learning_rate": 5e-07, "loss": 0.0029, "step": 499 }, { "epoch": 0.7987220447284346, "grad_norm": 0.6001198291778564, "learning_rate": 5e-07, "loss": 0.0028, "step": 500 }, { "epoch": 0.8003194888178914, "grad_norm": 0.4194888174533844, "learning_rate": 5e-07, "loss": 0.0022, "step": 501 }, { "epoch": 0.8019169329073482, "grad_norm": 0.1558961719274521, "learning_rate": 5e-07, "loss": 0.0018, "step": 502 }, { "epoch": 0.8035143769968051, "grad_norm": 0.1700737476348877, "learning_rate": 5e-07, "loss": 0.0021, "step": 503 }, { "epoch": 0.805111821086262, "grad_norm": 0.15276680886745453, "learning_rate": 5e-07, "loss": 0.0017, "step": 504 }, { "epoch": 0.8067092651757188, "grad_norm": 0.21490108966827393, "learning_rate": 5e-07, "loss": 0.0027, "step": 505 }, { "epoch": 0.8083067092651757, "grad_norm": 0.48414719104766846, "learning_rate": 5e-07, "loss": 0.0031, "step": 506 }, { "epoch": 0.8099041533546326, "grad_norm": 0.14232124388217926, "learning_rate": 5e-07, "loss": 0.0009, "step": 507 }, { "epoch": 0.8115015974440895, "grad_norm": 0.13643090426921844, "learning_rate": 5e-07, "loss": 0.0015, "step": 508 }, { "epoch": 0.8130990415335463, "grad_norm": 0.4378473162651062, "learning_rate": 5e-07, "loss": 0.0016, "step": 509 }, { "epoch": 0.8146964856230032, "grad_norm": 0.5300367474555969, "learning_rate": 5e-07, "loss": 0.0036, "step": 510 }, { "epoch": 0.8162939297124601, "grad_norm": 0.3394503593444824, "learning_rate": 5e-07, "loss": 0.0031, "step": 511 }, { "epoch": 0.8178913738019169, "grad_norm": 0.15926502645015717, "learning_rate": 5e-07, "loss": 0.0007, "step": 512 }, { "epoch": 0.8194888178913738, "grad_norm": 0.14627420902252197, "learning_rate": 5e-07, "loss": 0.0011, "step": 513 }, { "epoch": 0.8210862619808307, "grad_norm": 0.2294456660747528, "learning_rate": 5e-07, "loss": 0.0022, "step": 514 }, { "epoch": 0.8226837060702875, "grad_norm": 0.2830236554145813, "learning_rate": 5e-07, "loss": 0.0023, "step": 515 }, { "epoch": 0.8242811501597445, "grad_norm": 0.49337977170944214, "learning_rate": 5e-07, "loss": 0.0026, "step": 516 }, { "epoch": 0.8258785942492013, "grad_norm": 0.12382300198078156, "learning_rate": 5e-07, "loss": 0.0007, "step": 517 }, { "epoch": 0.8274760383386581, "grad_norm": 0.5072153210639954, "learning_rate": 5e-07, "loss": 0.0026, "step": 518 }, { "epoch": 0.829073482428115, "grad_norm": 0.6549058556556702, "learning_rate": 5e-07, "loss": 0.0044, "step": 519 }, { "epoch": 0.8306709265175719, "grad_norm": 0.21236664056777954, "learning_rate": 5e-07, "loss": 0.0008, "step": 520 }, { "epoch": 0.8322683706070287, "grad_norm": 0.6704789996147156, "learning_rate": 5e-07, "loss": 0.0038, "step": 521 }, { "epoch": 0.8338658146964856, "grad_norm": 0.1254369020462036, "learning_rate": 5e-07, "loss": 0.0007, "step": 522 }, { "epoch": 0.8354632587859425, "grad_norm": 2.3178699016571045, "learning_rate": 5e-07, "loss": 0.0033, "step": 523 }, { "epoch": 0.8370607028753994, "grad_norm": 3.5792298316955566, "learning_rate": 5e-07, "loss": 0.0038, "step": 524 }, { "epoch": 0.8386581469648562, "grad_norm": 0.7073982357978821, "learning_rate": 5e-07, "loss": 0.0027, "step": 525 }, { "epoch": 0.8402555910543131, "grad_norm": 0.5759369134902954, "learning_rate": 5e-07, "loss": 0.0035, "step": 526 }, { "epoch": 0.84185303514377, "grad_norm": 1.6669633388519287, "learning_rate": 5e-07, "loss": 0.0043, "step": 527 }, { "epoch": 0.8434504792332268, "grad_norm": 0.6380763649940491, "learning_rate": 5e-07, "loss": 0.0025, "step": 528 }, { "epoch": 0.8450479233226837, "grad_norm": 0.19805264472961426, "learning_rate": 5e-07, "loss": 0.002, "step": 529 }, { "epoch": 0.8466453674121406, "grad_norm": 0.6069849133491516, "learning_rate": 5e-07, "loss": 0.0019, "step": 530 }, { "epoch": 0.8482428115015974, "grad_norm": 0.1687164306640625, "learning_rate": 5e-07, "loss": 0.002, "step": 531 }, { "epoch": 0.8498402555910544, "grad_norm": 8.362351417541504, "learning_rate": 5e-07, "loss": 0.0054, "step": 532 }, { "epoch": 0.8514376996805112, "grad_norm": 0.2093094438314438, "learning_rate": 5e-07, "loss": 0.0016, "step": 533 }, { "epoch": 0.853035143769968, "grad_norm": 0.34365883469581604, "learning_rate": 5e-07, "loss": 0.0016, "step": 534 }, { "epoch": 0.854632587859425, "grad_norm": 0.12474914640188217, "learning_rate": 5e-07, "loss": 0.0019, "step": 535 }, { "epoch": 0.8562300319488818, "grad_norm": 0.510906994342804, "learning_rate": 5e-07, "loss": 0.0028, "step": 536 }, { "epoch": 0.8578274760383386, "grad_norm": 0.587397038936615, "learning_rate": 5e-07, "loss": 0.0036, "step": 537 }, { "epoch": 0.8594249201277955, "grad_norm": 0.844524621963501, "learning_rate": 5e-07, "loss": 0.0052, "step": 538 }, { "epoch": 0.8610223642172524, "grad_norm": 0.13176505267620087, "learning_rate": 5e-07, "loss": 0.0026, "step": 539 }, { "epoch": 0.8626198083067093, "grad_norm": 0.16751867532730103, "learning_rate": 5e-07, "loss": 0.0021, "step": 540 }, { "epoch": 0.8642172523961661, "grad_norm": 0.41158822178840637, "learning_rate": 5e-07, "loss": 0.0017, "step": 541 }, { "epoch": 0.865814696485623, "grad_norm": 0.12536601722240448, "learning_rate": 5e-07, "loss": 0.0018, "step": 542 }, { "epoch": 0.8674121405750799, "grad_norm": 0.08342156559228897, "learning_rate": 5e-07, "loss": 0.0008, "step": 543 }, { "epoch": 0.8690095846645367, "grad_norm": 1.5413963794708252, "learning_rate": 5e-07, "loss": 0.0033, "step": 544 }, { "epoch": 0.8706070287539937, "grad_norm": 0.24465155601501465, "learning_rate": 5e-07, "loss": 0.0016, "step": 545 }, { "epoch": 0.8722044728434505, "grad_norm": 0.1531795710325241, "learning_rate": 5e-07, "loss": 0.0019, "step": 546 }, { "epoch": 0.8738019169329073, "grad_norm": 0.1711510270833969, "learning_rate": 5e-07, "loss": 0.0023, "step": 547 }, { "epoch": 0.8753993610223643, "grad_norm": 0.10270589590072632, "learning_rate": 5e-07, "loss": 0.0018, "step": 548 }, { "epoch": 0.8769968051118211, "grad_norm": 0.6564199328422546, "learning_rate": 5e-07, "loss": 0.004, "step": 549 }, { "epoch": 0.8785942492012779, "grad_norm": 0.15586939454078674, "learning_rate": 5e-07, "loss": 0.0015, "step": 550 }, { "epoch": 0.8801916932907349, "grad_norm": 0.11600361764431, "learning_rate": 5e-07, "loss": 0.0022, "step": 551 }, { "epoch": 0.8817891373801917, "grad_norm": 0.10974587500095367, "learning_rate": 5e-07, "loss": 0.0011, "step": 552 }, { "epoch": 0.8833865814696485, "grad_norm": 0.2567409574985504, "learning_rate": 5e-07, "loss": 0.002, "step": 553 }, { "epoch": 0.8849840255591054, "grad_norm": 0.27306219935417175, "learning_rate": 5e-07, "loss": 0.0043, "step": 554 }, { "epoch": 0.8865814696485623, "grad_norm": 0.3767613470554352, "learning_rate": 5e-07, "loss": 0.0017, "step": 555 }, { "epoch": 0.8881789137380192, "grad_norm": 1.1817584037780762, "learning_rate": 5e-07, "loss": 0.0057, "step": 556 }, { "epoch": 0.889776357827476, "grad_norm": 0.3905037045478821, "learning_rate": 5e-07, "loss": 0.0031, "step": 557 }, { "epoch": 0.8913738019169329, "grad_norm": 0.23763906955718994, "learning_rate": 5e-07, "loss": 0.0006, "step": 558 }, { "epoch": 0.8929712460063898, "grad_norm": 0.3837161362171173, "learning_rate": 5e-07, "loss": 0.0023, "step": 559 }, { "epoch": 0.8945686900958466, "grad_norm": 0.6008383631706238, "learning_rate": 5e-07, "loss": 0.0041, "step": 560 }, { "epoch": 0.8961661341853036, "grad_norm": 0.1975802332162857, "learning_rate": 5e-07, "loss": 0.0017, "step": 561 }, { "epoch": 0.8977635782747604, "grad_norm": 0.15433986485004425, "learning_rate": 5e-07, "loss": 0.0024, "step": 562 }, { "epoch": 0.8993610223642172, "grad_norm": 0.4065248966217041, "learning_rate": 5e-07, "loss": 0.0021, "step": 563 }, { "epoch": 0.9009584664536742, "grad_norm": 0.25533992052078247, "learning_rate": 5e-07, "loss": 0.0023, "step": 564 }, { "epoch": 0.902555910543131, "grad_norm": 0.46665945649147034, "learning_rate": 5e-07, "loss": 0.0028, "step": 565 }, { "epoch": 0.9041533546325878, "grad_norm": 0.16390737891197205, "learning_rate": 5e-07, "loss": 0.0012, "step": 566 }, { "epoch": 0.9057507987220448, "grad_norm": 0.06191791966557503, "learning_rate": 5e-07, "loss": 0.0005, "step": 567 }, { "epoch": 0.9073482428115016, "grad_norm": 1.692974328994751, "learning_rate": 5e-07, "loss": 0.0011, "step": 568 }, { "epoch": 0.9089456869009584, "grad_norm": 0.21359291672706604, "learning_rate": 5e-07, "loss": 0.0016, "step": 569 }, { "epoch": 0.9105431309904153, "grad_norm": 0.09417448937892914, "learning_rate": 5e-07, "loss": 0.0019, "step": 570 }, { "epoch": 0.9121405750798722, "grad_norm": 0.22416774928569794, "learning_rate": 5e-07, "loss": 0.0025, "step": 571 }, { "epoch": 0.9137380191693291, "grad_norm": 0.15862461924552917, "learning_rate": 5e-07, "loss": 0.0018, "step": 572 }, { "epoch": 0.9153354632587859, "grad_norm": 0.325225293636322, "learning_rate": 5e-07, "loss": 0.0021, "step": 573 }, { "epoch": 0.9169329073482428, "grad_norm": 0.4235207140445709, "learning_rate": 5e-07, "loss": 0.0017, "step": 574 }, { "epoch": 0.9185303514376997, "grad_norm": 0.17048902809619904, "learning_rate": 5e-07, "loss": 0.0024, "step": 575 }, { "epoch": 0.9201277955271565, "grad_norm": 0.28779876232147217, "learning_rate": 5e-07, "loss": 0.0011, "step": 576 }, { "epoch": 0.9217252396166135, "grad_norm": 0.16058605909347534, "learning_rate": 5e-07, "loss": 0.0019, "step": 577 }, { "epoch": 0.9233226837060703, "grad_norm": 0.08717779070138931, "learning_rate": 5e-07, "loss": 0.0011, "step": 578 }, { "epoch": 0.9249201277955271, "grad_norm": 0.852260172367096, "learning_rate": 5e-07, "loss": 0.0033, "step": 579 }, { "epoch": 0.9265175718849841, "grad_norm": 0.34097403287887573, "learning_rate": 5e-07, "loss": 0.0022, "step": 580 }, { "epoch": 0.9281150159744409, "grad_norm": 0.3408906161785126, "learning_rate": 5e-07, "loss": 0.0032, "step": 581 }, { "epoch": 0.9297124600638977, "grad_norm": 0.4731082618236542, "learning_rate": 5e-07, "loss": 0.0037, "step": 582 }, { "epoch": 0.9313099041533547, "grad_norm": 0.49347537755966187, "learning_rate": 5e-07, "loss": 0.0023, "step": 583 }, { "epoch": 0.9329073482428115, "grad_norm": 0.17723320424556732, "learning_rate": 5e-07, "loss": 0.0018, "step": 584 }, { "epoch": 0.9345047923322684, "grad_norm": 0.16061758995056152, "learning_rate": 5e-07, "loss": 0.0022, "step": 585 }, { "epoch": 0.9361022364217252, "grad_norm": 0.11341333389282227, "learning_rate": 5e-07, "loss": 0.0017, "step": 586 }, { "epoch": 0.9376996805111821, "grad_norm": 0.12614861130714417, "learning_rate": 5e-07, "loss": 0.0015, "step": 587 }, { "epoch": 0.939297124600639, "grad_norm": 0.1919727921485901, "learning_rate": 5e-07, "loss": 0.0033, "step": 588 }, { "epoch": 0.9408945686900958, "grad_norm": 0.5518249273300171, "learning_rate": 5e-07, "loss": 0.0034, "step": 589 }, { "epoch": 0.9424920127795527, "grad_norm": 0.11890781670808792, "learning_rate": 5e-07, "loss": 0.0028, "step": 590 }, { "epoch": 0.9440894568690096, "grad_norm": 0.5765843391418457, "learning_rate": 5e-07, "loss": 0.003, "step": 591 }, { "epoch": 0.9456869009584664, "grad_norm": 0.17196467518806458, "learning_rate": 5e-07, "loss": 0.0019, "step": 592 }, { "epoch": 0.9472843450479234, "grad_norm": 0.2513171136379242, "learning_rate": 5e-07, "loss": 0.0026, "step": 593 }, { "epoch": 0.9488817891373802, "grad_norm": 1.7397063970565796, "learning_rate": 5e-07, "loss": 0.0108, "step": 594 }, { "epoch": 0.950479233226837, "grad_norm": 0.2406977415084839, "learning_rate": 5e-07, "loss": 0.002, "step": 595 }, { "epoch": 0.952076677316294, "grad_norm": 0.23791703581809998, "learning_rate": 5e-07, "loss": 0.0022, "step": 596 }, { "epoch": 0.9536741214057508, "grad_norm": 0.22479400038719177, "learning_rate": 5e-07, "loss": 0.0025, "step": 597 }, { "epoch": 0.9552715654952076, "grad_norm": 0.23164299130439758, "learning_rate": 5e-07, "loss": 0.0015, "step": 598 }, { "epoch": 0.9568690095846646, "grad_norm": 0.1600014865398407, "learning_rate": 5e-07, "loss": 0.0019, "step": 599 }, { "epoch": 0.9584664536741214, "grad_norm": 0.2001635879278183, "learning_rate": 5e-07, "loss": 0.002, "step": 600 }, { "epoch": 0.9600638977635783, "grad_norm": 0.2946987450122833, "learning_rate": 5e-07, "loss": 0.0011, "step": 601 }, { "epoch": 0.9616613418530351, "grad_norm": 0.2080981582403183, "learning_rate": 5e-07, "loss": 0.0021, "step": 602 }, { "epoch": 0.963258785942492, "grad_norm": 0.6526065468788147, "learning_rate": 5e-07, "loss": 0.0035, "step": 603 }, { "epoch": 0.9648562300319489, "grad_norm": 0.3620392382144928, "learning_rate": 5e-07, "loss": 0.0024, "step": 604 }, { "epoch": 0.9664536741214057, "grad_norm": 0.21057778596878052, "learning_rate": 5e-07, "loss": 0.0026, "step": 605 }, { "epoch": 0.9680511182108626, "grad_norm": 0.11293834447860718, "learning_rate": 5e-07, "loss": 0.0008, "step": 606 }, { "epoch": 0.9696485623003195, "grad_norm": 1.2960716485977173, "learning_rate": 5e-07, "loss": 0.0027, "step": 607 }, { "epoch": 0.9712460063897763, "grad_norm": 0.1437765210866928, "learning_rate": 5e-07, "loss": 0.0014, "step": 608 }, { "epoch": 0.9728434504792333, "grad_norm": 0.6358578205108643, "learning_rate": 5e-07, "loss": 0.0049, "step": 609 }, { "epoch": 0.9744408945686901, "grad_norm": 0.14403465390205383, "learning_rate": 5e-07, "loss": 0.0027, "step": 610 }, { "epoch": 0.9760383386581469, "grad_norm": 0.4725677967071533, "learning_rate": 5e-07, "loss": 0.0045, "step": 611 }, { "epoch": 0.9776357827476039, "grad_norm": 0.8543768525123596, "learning_rate": 5e-07, "loss": 0.0027, "step": 612 }, { "epoch": 0.9792332268370607, "grad_norm": 0.27371206879615784, "learning_rate": 5e-07, "loss": 0.0016, "step": 613 }, { "epoch": 0.9808306709265175, "grad_norm": 0.23279932141304016, "learning_rate": 5e-07, "loss": 0.0009, "step": 614 }, { "epoch": 0.9824281150159745, "grad_norm": 0.8952144384384155, "learning_rate": 5e-07, "loss": 0.0037, "step": 615 }, { "epoch": 0.9840255591054313, "grad_norm": 0.11720918118953705, "learning_rate": 5e-07, "loss": 0.0017, "step": 616 }, { "epoch": 0.9856230031948882, "grad_norm": 0.9963851571083069, "learning_rate": 5e-07, "loss": 0.0028, "step": 617 }, { "epoch": 0.987220447284345, "grad_norm": 0.08971522003412247, "learning_rate": 5e-07, "loss": 0.0014, "step": 618 }, { "epoch": 0.9888178913738019, "grad_norm": 0.1214229092001915, "learning_rate": 5e-07, "loss": 0.0016, "step": 619 }, { "epoch": 0.9904153354632588, "grad_norm": 0.2924940288066864, "learning_rate": 5e-07, "loss": 0.002, "step": 620 }, { "epoch": 0.9920127795527156, "grad_norm": 1.189788579940796, "learning_rate": 5e-07, "loss": 0.003, "step": 621 }, { "epoch": 0.9936102236421726, "grad_norm": 0.5802843570709229, "learning_rate": 5e-07, "loss": 0.0044, "step": 622 }, { "epoch": 0.9952076677316294, "grad_norm": 0.11823207885026932, "learning_rate": 5e-07, "loss": 0.0021, "step": 623 }, { "epoch": 0.9968051118210862, "grad_norm": 0.16288286447525024, "learning_rate": 5e-07, "loss": 0.0009, "step": 624 }, { "epoch": 0.9984025559105432, "grad_norm": 0.13317349553108215, "learning_rate": 5e-07, "loss": 0.0023, "step": 625 }, { "epoch": 1.0, "grad_norm": 0.1848553717136383, "learning_rate": 5e-07, "loss": 0.0028, "step": 626 }, { "epoch": 1.0015974440894568, "grad_norm": 0.5851624608039856, "learning_rate": 5e-07, "loss": 0.0025, "step": 627 }, { "epoch": 1.0031948881789137, "grad_norm": 0.09685046225786209, "learning_rate": 5e-07, "loss": 0.0009, "step": 628 }, { "epoch": 1.0047923322683705, "grad_norm": 0.11342930048704147, "learning_rate": 5e-07, "loss": 0.0012, "step": 629 }, { "epoch": 1.0063897763578276, "grad_norm": 0.46440932154655457, "learning_rate": 5e-07, "loss": 0.0026, "step": 630 }, { "epoch": 1.0079872204472844, "grad_norm": 0.19794267416000366, "learning_rate": 5e-07, "loss": 0.0014, "step": 631 }, { "epoch": 1.0095846645367412, "grad_norm": 0.3643769919872284, "learning_rate": 5e-07, "loss": 0.0041, "step": 632 }, { "epoch": 1.011182108626198, "grad_norm": 0.08710943907499313, "learning_rate": 5e-07, "loss": 0.0017, "step": 633 }, { "epoch": 1.012779552715655, "grad_norm": 0.4915333688259125, "learning_rate": 5e-07, "loss": 0.0031, "step": 634 }, { "epoch": 1.0143769968051117, "grad_norm": 0.23040416836738586, "learning_rate": 5e-07, "loss": 0.0026, "step": 635 }, { "epoch": 1.0159744408945688, "grad_norm": 0.24800904095172882, "learning_rate": 5e-07, "loss": 0.0025, "step": 636 }, { "epoch": 1.0175718849840256, "grad_norm": 0.29919004440307617, "learning_rate": 5e-07, "loss": 0.0024, "step": 637 }, { "epoch": 1.0191693290734825, "grad_norm": 0.11212212592363358, "learning_rate": 5e-07, "loss": 0.0012, "step": 638 }, { "epoch": 1.0207667731629393, "grad_norm": 0.16430723667144775, "learning_rate": 5e-07, "loss": 0.0019, "step": 639 }, { "epoch": 1.0223642172523961, "grad_norm": 0.32202014327049255, "learning_rate": 5e-07, "loss": 0.0046, "step": 640 }, { "epoch": 1.023961661341853, "grad_norm": 0.3237399458885193, "learning_rate": 5e-07, "loss": 0.002, "step": 641 }, { "epoch": 1.0255591054313098, "grad_norm": 0.1953740268945694, "learning_rate": 5e-07, "loss": 0.003, "step": 642 }, { "epoch": 1.0271565495207668, "grad_norm": 0.12679028511047363, "learning_rate": 5e-07, "loss": 0.0008, "step": 643 }, { "epoch": 1.0287539936102237, "grad_norm": 0.10133963078260422, "learning_rate": 5e-07, "loss": 0.0012, "step": 644 }, { "epoch": 1.0303514376996805, "grad_norm": 0.0973275899887085, "learning_rate": 5e-07, "loss": 0.0012, "step": 645 }, { "epoch": 1.0319488817891374, "grad_norm": 0.5217294096946716, "learning_rate": 5e-07, "loss": 0.0028, "step": 646 }, { "epoch": 1.0335463258785942, "grad_norm": 0.15706409513950348, "learning_rate": 5e-07, "loss": 0.0016, "step": 647 }, { "epoch": 1.035143769968051, "grad_norm": 0.10915224999189377, "learning_rate": 5e-07, "loss": 0.0022, "step": 648 }, { "epoch": 1.036741214057508, "grad_norm": 0.4759727716445923, "learning_rate": 5e-07, "loss": 0.0018, "step": 649 }, { "epoch": 1.038338658146965, "grad_norm": 0.11390835046768188, "learning_rate": 5e-07, "loss": 0.0016, "step": 650 }, { "epoch": 1.0399361022364217, "grad_norm": 1.093390703201294, "learning_rate": 5e-07, "loss": 0.0046, "step": 651 }, { "epoch": 1.0415335463258786, "grad_norm": 0.07437872141599655, "learning_rate": 5e-07, "loss": 0.0016, "step": 652 }, { "epoch": 1.0431309904153354, "grad_norm": 0.21833601593971252, "learning_rate": 5e-07, "loss": 0.003, "step": 653 }, { "epoch": 1.0447284345047922, "grad_norm": 0.21499225497245789, "learning_rate": 5e-07, "loss": 0.0021, "step": 654 }, { "epoch": 1.0463258785942493, "grad_norm": 0.12159962207078934, "learning_rate": 5e-07, "loss": 0.0013, "step": 655 }, { "epoch": 1.0479233226837061, "grad_norm": 0.3396832346916199, "learning_rate": 5e-07, "loss": 0.0028, "step": 656 }, { "epoch": 1.049520766773163, "grad_norm": 0.21546514332294464, "learning_rate": 5e-07, "loss": 0.0027, "step": 657 }, { "epoch": 1.0511182108626198, "grad_norm": 0.19969788193702698, "learning_rate": 5e-07, "loss": 0.0014, "step": 658 }, { "epoch": 1.0527156549520766, "grad_norm": 0.08988461643457413, "learning_rate": 5e-07, "loss": 0.0017, "step": 659 }, { "epoch": 1.0543130990415335, "grad_norm": 0.26418858766555786, "learning_rate": 5e-07, "loss": 0.0027, "step": 660 }, { "epoch": 1.0559105431309903, "grad_norm": 0.2093944102525711, "learning_rate": 5e-07, "loss": 0.0025, "step": 661 }, { "epoch": 1.0575079872204474, "grad_norm": 0.36498188972473145, "learning_rate": 5e-07, "loss": 0.0014, "step": 662 }, { "epoch": 1.0591054313099042, "grad_norm": 0.07945135235786438, "learning_rate": 5e-07, "loss": 0.0017, "step": 663 }, { "epoch": 1.060702875399361, "grad_norm": 0.12991008162498474, "learning_rate": 5e-07, "loss": 0.0017, "step": 664 }, { "epoch": 1.0623003194888179, "grad_norm": 0.2589464783668518, "learning_rate": 5e-07, "loss": 0.0023, "step": 665 }, { "epoch": 1.0638977635782747, "grad_norm": 0.10802029818296432, "learning_rate": 5e-07, "loss": 0.0012, "step": 666 }, { "epoch": 1.0654952076677315, "grad_norm": 0.4823663532733917, "learning_rate": 5e-07, "loss": 0.0021, "step": 667 }, { "epoch": 1.0670926517571886, "grad_norm": 0.44493567943573, "learning_rate": 5e-07, "loss": 0.0011, "step": 668 }, { "epoch": 1.0686900958466454, "grad_norm": 0.23464570939540863, "learning_rate": 5e-07, "loss": 0.0031, "step": 669 }, { "epoch": 1.0702875399361023, "grad_norm": 0.28489264845848083, "learning_rate": 5e-07, "loss": 0.0017, "step": 670 }, { "epoch": 1.071884984025559, "grad_norm": 0.2952616214752197, "learning_rate": 5e-07, "loss": 0.0028, "step": 671 }, { "epoch": 1.073482428115016, "grad_norm": 0.20882096886634827, "learning_rate": 5e-07, "loss": 0.0011, "step": 672 }, { "epoch": 1.0750798722044728, "grad_norm": 0.12549445033073425, "learning_rate": 5e-07, "loss": 0.0018, "step": 673 }, { "epoch": 1.0766773162939298, "grad_norm": 0.22115099430084229, "learning_rate": 5e-07, "loss": 0.0023, "step": 674 }, { "epoch": 1.0782747603833867, "grad_norm": 0.15710076689720154, "learning_rate": 5e-07, "loss": 0.0009, "step": 675 }, { "epoch": 1.0798722044728435, "grad_norm": 0.15250109136104584, "learning_rate": 5e-07, "loss": 0.002, "step": 676 }, { "epoch": 1.0814696485623003, "grad_norm": 0.1936073750257492, "learning_rate": 5e-07, "loss": 0.0018, "step": 677 }, { "epoch": 1.0830670926517572, "grad_norm": 0.08779247850179672, "learning_rate": 5e-07, "loss": 0.0006, "step": 678 }, { "epoch": 1.084664536741214, "grad_norm": 0.1437022089958191, "learning_rate": 5e-07, "loss": 0.0011, "step": 679 }, { "epoch": 1.0862619808306708, "grad_norm": 0.2850930392742157, "learning_rate": 5e-07, "loss": 0.0029, "step": 680 }, { "epoch": 1.0878594249201279, "grad_norm": 0.3779478967189789, "learning_rate": 5e-07, "loss": 0.0051, "step": 681 }, { "epoch": 1.0894568690095847, "grad_norm": 0.20855233073234558, "learning_rate": 5e-07, "loss": 0.0025, "step": 682 }, { "epoch": 1.0910543130990416, "grad_norm": 0.11478544771671295, "learning_rate": 5e-07, "loss": 0.0021, "step": 683 }, { "epoch": 1.0926517571884984, "grad_norm": 0.3283560276031494, "learning_rate": 5e-07, "loss": 0.0014, "step": 684 }, { "epoch": 1.0942492012779552, "grad_norm": 0.35279855132102966, "learning_rate": 5e-07, "loss": 0.0025, "step": 685 }, { "epoch": 1.095846645367412, "grad_norm": 0.166551411151886, "learning_rate": 5e-07, "loss": 0.0023, "step": 686 }, { "epoch": 1.097444089456869, "grad_norm": 0.10785111039876938, "learning_rate": 5e-07, "loss": 0.0025, "step": 687 }, { "epoch": 1.099041533546326, "grad_norm": 0.1150643453001976, "learning_rate": 5e-07, "loss": 0.0012, "step": 688 }, { "epoch": 1.1006389776357828, "grad_norm": 0.10543517768383026, "learning_rate": 5e-07, "loss": 0.0019, "step": 689 }, { "epoch": 1.1022364217252396, "grad_norm": 0.12300563603639603, "learning_rate": 5e-07, "loss": 0.0021, "step": 690 }, { "epoch": 1.1038338658146964, "grad_norm": 0.15775950253009796, "learning_rate": 5e-07, "loss": 0.0018, "step": 691 }, { "epoch": 1.1054313099041533, "grad_norm": 0.052068062126636505, "learning_rate": 5e-07, "loss": 0.0008, "step": 692 }, { "epoch": 1.1070287539936103, "grad_norm": 0.1371004432439804, "learning_rate": 5e-07, "loss": 0.0029, "step": 693 }, { "epoch": 1.1086261980830672, "grad_norm": 0.11871246248483658, "learning_rate": 5e-07, "loss": 0.0014, "step": 694 }, { "epoch": 1.110223642172524, "grad_norm": 0.106499083340168, "learning_rate": 5e-07, "loss": 0.0017, "step": 695 }, { "epoch": 1.1118210862619808, "grad_norm": 0.05005495250225067, "learning_rate": 5e-07, "loss": 0.001, "step": 696 }, { "epoch": 1.1134185303514377, "grad_norm": 0.6656812429428101, "learning_rate": 5e-07, "loss": 0.0036, "step": 697 }, { "epoch": 1.1150159744408945, "grad_norm": 0.14445379376411438, "learning_rate": 5e-07, "loss": 0.0023, "step": 698 }, { "epoch": 1.1166134185303513, "grad_norm": 0.268867164850235, "learning_rate": 5e-07, "loss": 0.0013, "step": 699 }, { "epoch": 1.1182108626198084, "grad_norm": 0.16358579695224762, "learning_rate": 5e-07, "loss": 0.0018, "step": 700 }, { "epoch": 1.1198083067092652, "grad_norm": 0.11881609261035919, "learning_rate": 5e-07, "loss": 0.0011, "step": 701 }, { "epoch": 1.121405750798722, "grad_norm": 0.10697460919618607, "learning_rate": 5e-07, "loss": 0.0014, "step": 702 }, { "epoch": 1.123003194888179, "grad_norm": 0.1314931958913803, "learning_rate": 5e-07, "loss": 0.0022, "step": 703 }, { "epoch": 1.1246006389776357, "grad_norm": 0.13953609764575958, "learning_rate": 5e-07, "loss": 0.0027, "step": 704 }, { "epoch": 1.1261980830670926, "grad_norm": 0.30124762654304504, "learning_rate": 5e-07, "loss": 0.0029, "step": 705 }, { "epoch": 1.1277955271565494, "grad_norm": 0.2047244906425476, "learning_rate": 5e-07, "loss": 0.0029, "step": 706 }, { "epoch": 1.1293929712460065, "grad_norm": 0.28825926780700684, "learning_rate": 5e-07, "loss": 0.0026, "step": 707 }, { "epoch": 1.1309904153354633, "grad_norm": 0.1030370369553566, "learning_rate": 5e-07, "loss": 0.0016, "step": 708 }, { "epoch": 1.1325878594249201, "grad_norm": 0.08416575193405151, "learning_rate": 5e-07, "loss": 0.0011, "step": 709 }, { "epoch": 1.134185303514377, "grad_norm": 0.6813434958457947, "learning_rate": 5e-07, "loss": 0.0046, "step": 710 }, { "epoch": 1.1357827476038338, "grad_norm": 0.7300125360488892, "learning_rate": 5e-07, "loss": 0.0022, "step": 711 }, { "epoch": 1.1373801916932909, "grad_norm": 0.21185138821601868, "learning_rate": 5e-07, "loss": 0.0015, "step": 712 }, { "epoch": 1.1389776357827477, "grad_norm": 0.3095090389251709, "learning_rate": 5e-07, "loss": 0.003, "step": 713 }, { "epoch": 1.1405750798722045, "grad_norm": 0.13913819193840027, "learning_rate": 5e-07, "loss": 0.0016, "step": 714 }, { "epoch": 1.1421725239616614, "grad_norm": 0.2016637921333313, "learning_rate": 5e-07, "loss": 0.0025, "step": 715 }, { "epoch": 1.1437699680511182, "grad_norm": 0.12468226253986359, "learning_rate": 5e-07, "loss": 0.0011, "step": 716 }, { "epoch": 1.145367412140575, "grad_norm": 0.1510785073041916, "learning_rate": 5e-07, "loss": 0.002, "step": 717 }, { "epoch": 1.1469648562300319, "grad_norm": 0.11004221439361572, "learning_rate": 5e-07, "loss": 0.0015, "step": 718 }, { "epoch": 1.148562300319489, "grad_norm": 0.12669166922569275, "learning_rate": 5e-07, "loss": 0.003, "step": 719 }, { "epoch": 1.1501597444089458, "grad_norm": 0.11249173432588577, "learning_rate": 5e-07, "loss": 0.0012, "step": 720 }, { "epoch": 1.1517571884984026, "grad_norm": 0.21114954352378845, "learning_rate": 5e-07, "loss": 0.0025, "step": 721 }, { "epoch": 1.1533546325878594, "grad_norm": 0.3067387342453003, "learning_rate": 5e-07, "loss": 0.0037, "step": 722 }, { "epoch": 1.1549520766773163, "grad_norm": 0.11942733824253082, "learning_rate": 5e-07, "loss": 0.0021, "step": 723 }, { "epoch": 1.156549520766773, "grad_norm": 0.2498818337917328, "learning_rate": 5e-07, "loss": 0.0027, "step": 724 }, { "epoch": 1.15814696485623, "grad_norm": 0.589302122592926, "learning_rate": 5e-07, "loss": 0.0014, "step": 725 }, { "epoch": 1.159744408945687, "grad_norm": 0.11040360480546951, "learning_rate": 5e-07, "loss": 0.0015, "step": 726 }, { "epoch": 1.1613418530351438, "grad_norm": 0.1865539252758026, "learning_rate": 5e-07, "loss": 0.0013, "step": 727 }, { "epoch": 1.1629392971246006, "grad_norm": 0.09886899590492249, "learning_rate": 5e-07, "loss": 0.0025, "step": 728 }, { "epoch": 1.1645367412140575, "grad_norm": 0.10603392124176025, "learning_rate": 5e-07, "loss": 0.0012, "step": 729 }, { "epoch": 1.1661341853035143, "grad_norm": 0.17114412784576416, "learning_rate": 5e-07, "loss": 0.0034, "step": 730 }, { "epoch": 1.1677316293929714, "grad_norm": 0.11507996916770935, "learning_rate": 5e-07, "loss": 0.0017, "step": 731 }, { "epoch": 1.1693290734824282, "grad_norm": 0.10117539763450623, "learning_rate": 5e-07, "loss": 0.001, "step": 732 }, { "epoch": 1.170926517571885, "grad_norm": 0.18606479465961456, "learning_rate": 5e-07, "loss": 0.0021, "step": 733 }, { "epoch": 1.1725239616613419, "grad_norm": 0.1562090963125229, "learning_rate": 5e-07, "loss": 0.0013, "step": 734 }, { "epoch": 1.1741214057507987, "grad_norm": 0.2511482834815979, "learning_rate": 5e-07, "loss": 0.0028, "step": 735 }, { "epoch": 1.1757188498402555, "grad_norm": 0.131776362657547, "learning_rate": 5e-07, "loss": 0.0023, "step": 736 }, { "epoch": 1.1773162939297124, "grad_norm": 0.2350272536277771, "learning_rate": 5e-07, "loss": 0.0026, "step": 737 }, { "epoch": 1.1789137380191694, "grad_norm": 0.5341865420341492, "learning_rate": 5e-07, "loss": 0.0045, "step": 738 }, { "epoch": 1.1805111821086263, "grad_norm": 0.18602542579174042, "learning_rate": 5e-07, "loss": 0.0028, "step": 739 }, { "epoch": 1.182108626198083, "grad_norm": 0.08645334094762802, "learning_rate": 5e-07, "loss": 0.0009, "step": 740 }, { "epoch": 1.18370607028754, "grad_norm": 0.1708468645811081, "learning_rate": 5e-07, "loss": 0.0018, "step": 741 }, { "epoch": 1.1853035143769968, "grad_norm": 0.11233728379011154, "learning_rate": 5e-07, "loss": 0.0026, "step": 742 }, { "epoch": 1.1869009584664536, "grad_norm": 0.20240649580955505, "learning_rate": 5e-07, "loss": 0.0013, "step": 743 }, { "epoch": 1.1884984025559104, "grad_norm": 0.14982427656650543, "learning_rate": 5e-07, "loss": 0.0018, "step": 744 }, { "epoch": 1.1900958466453675, "grad_norm": 0.13709725439548492, "learning_rate": 5e-07, "loss": 0.0018, "step": 745 }, { "epoch": 1.1916932907348243, "grad_norm": 0.18215787410736084, "learning_rate": 5e-07, "loss": 0.0017, "step": 746 }, { "epoch": 1.1932907348242812, "grad_norm": 0.12986968457698822, "learning_rate": 5e-07, "loss": 0.0026, "step": 747 }, { "epoch": 1.194888178913738, "grad_norm": 0.10665430873632431, "learning_rate": 5e-07, "loss": 0.0017, "step": 748 }, { "epoch": 1.1964856230031948, "grad_norm": 0.11768273264169693, "learning_rate": 5e-07, "loss": 0.0018, "step": 749 }, { "epoch": 1.1980830670926517, "grad_norm": 0.6945746541023254, "learning_rate": 5e-07, "loss": 0.0008, "step": 750 }, { "epoch": 1.1996805111821087, "grad_norm": 0.25220727920532227, "learning_rate": 5e-07, "loss": 0.0024, "step": 751 }, { "epoch": 1.2012779552715656, "grad_norm": 0.07079087197780609, "learning_rate": 5e-07, "loss": 0.0005, "step": 752 }, { "epoch": 1.2028753993610224, "grad_norm": 0.16933268308639526, "learning_rate": 5e-07, "loss": 0.0021, "step": 753 }, { "epoch": 1.2044728434504792, "grad_norm": 0.15669192373752594, "learning_rate": 5e-07, "loss": 0.001, "step": 754 }, { "epoch": 1.206070287539936, "grad_norm": 6.026957035064697, "learning_rate": 5e-07, "loss": 0.0035, "step": 755 }, { "epoch": 1.207667731629393, "grad_norm": 0.5998969674110413, "learning_rate": 5e-07, "loss": 0.0039, "step": 756 }, { "epoch": 1.20926517571885, "grad_norm": 0.06143517419695854, "learning_rate": 5e-07, "loss": 0.001, "step": 757 }, { "epoch": 1.2108626198083068, "grad_norm": 0.2319423258304596, "learning_rate": 5e-07, "loss": 0.0017, "step": 758 }, { "epoch": 1.2124600638977636, "grad_norm": 0.701130211353302, "learning_rate": 5e-07, "loss": 0.0056, "step": 759 }, { "epoch": 1.2140575079872205, "grad_norm": 0.3482913076877594, "learning_rate": 5e-07, "loss": 0.0029, "step": 760 }, { "epoch": 1.2156549520766773, "grad_norm": 0.21112686395645142, "learning_rate": 5e-07, "loss": 0.0013, "step": 761 }, { "epoch": 1.2172523961661341, "grad_norm": 5.14704704284668, "learning_rate": 5e-07, "loss": 0.0024, "step": 762 }, { "epoch": 1.218849840255591, "grad_norm": 0.41259634494781494, "learning_rate": 5e-07, "loss": 0.0018, "step": 763 }, { "epoch": 1.220447284345048, "grad_norm": 0.2882460951805115, "learning_rate": 5e-07, "loss": 0.0027, "step": 764 }, { "epoch": 1.2220447284345048, "grad_norm": 0.22736166417598724, "learning_rate": 5e-07, "loss": 0.0026, "step": 765 }, { "epoch": 1.2236421725239617, "grad_norm": 0.2859097719192505, "learning_rate": 5e-07, "loss": 0.0023, "step": 766 }, { "epoch": 1.2252396166134185, "grad_norm": 1.8469651937484741, "learning_rate": 5e-07, "loss": 0.0032, "step": 767 }, { "epoch": 1.2268370607028753, "grad_norm": 0.15771770477294922, "learning_rate": 5e-07, "loss": 0.0007, "step": 768 }, { "epoch": 1.2284345047923322, "grad_norm": 0.10112886875867844, "learning_rate": 5e-07, "loss": 0.001, "step": 769 }, { "epoch": 1.230031948881789, "grad_norm": 0.7125353813171387, "learning_rate": 5e-07, "loss": 0.003, "step": 770 }, { "epoch": 1.231629392971246, "grad_norm": 0.10188297927379608, "learning_rate": 5e-07, "loss": 0.0009, "step": 771 }, { "epoch": 1.233226837060703, "grad_norm": 0.9515169262886047, "learning_rate": 5e-07, "loss": 0.0027, "step": 772 }, { "epoch": 1.2348242811501597, "grad_norm": 0.06836796551942825, "learning_rate": 5e-07, "loss": 0.0011, "step": 773 }, { "epoch": 1.2364217252396166, "grad_norm": 0.13632221519947052, "learning_rate": 5e-07, "loss": 0.0024, "step": 774 }, { "epoch": 1.2380191693290734, "grad_norm": 0.572127103805542, "learning_rate": 5e-07, "loss": 0.0027, "step": 775 }, { "epoch": 1.2396166134185305, "grad_norm": 0.11243536323308945, "learning_rate": 5e-07, "loss": 0.0024, "step": 776 }, { "epoch": 1.2412140575079873, "grad_norm": 0.47509658336639404, "learning_rate": 5e-07, "loss": 0.0013, "step": 777 }, { "epoch": 1.2428115015974441, "grad_norm": 3.768028974533081, "learning_rate": 5e-07, "loss": 0.0039, "step": 778 }, { "epoch": 1.244408945686901, "grad_norm": 0.346608966588974, "learning_rate": 5e-07, "loss": 0.0017, "step": 779 }, { "epoch": 1.2460063897763578, "grad_norm": 0.23147884011268616, "learning_rate": 5e-07, "loss": 0.0018, "step": 780 }, { "epoch": 1.2476038338658146, "grad_norm": 0.14679104089736938, "learning_rate": 5e-07, "loss": 0.0021, "step": 781 }, { "epoch": 1.2492012779552715, "grad_norm": 0.23181918263435364, "learning_rate": 5e-07, "loss": 0.0014, "step": 782 }, { "epoch": 1.2507987220447285, "grad_norm": 0.16475149989128113, "learning_rate": 5e-07, "loss": 0.0018, "step": 783 }, { "epoch": 1.2523961661341854, "grad_norm": 0.08807860314846039, "learning_rate": 5e-07, "loss": 0.0017, "step": 784 }, { "epoch": 1.2539936102236422, "grad_norm": 0.16376349329948425, "learning_rate": 5e-07, "loss": 0.0018, "step": 785 }, { "epoch": 1.255591054313099, "grad_norm": 0.19524359703063965, "learning_rate": 5e-07, "loss": 0.0016, "step": 786 }, { "epoch": 1.2571884984025559, "grad_norm": 0.06623630225658417, "learning_rate": 5e-07, "loss": 0.0013, "step": 787 }, { "epoch": 1.2587859424920127, "grad_norm": 0.03929225727915764, "learning_rate": 5e-07, "loss": 0.0003, "step": 788 }, { "epoch": 1.2603833865814695, "grad_norm": 0.11669357866048813, "learning_rate": 5e-07, "loss": 0.0011, "step": 789 }, { "epoch": 1.2619808306709266, "grad_norm": 0.00831072498112917, "learning_rate": 5e-07, "loss": 0.0, "step": 790 }, { "epoch": 1.2635782747603834, "grad_norm": 0.13926006853580475, "learning_rate": 5e-07, "loss": 0.0025, "step": 791 }, { "epoch": 1.2651757188498403, "grad_norm": 4.567328929901123, "learning_rate": 5e-07, "loss": 0.0034, "step": 792 }, { "epoch": 1.266773162939297, "grad_norm": 1.018916368484497, "learning_rate": 5e-07, "loss": 0.0067, "step": 793 }, { "epoch": 1.268370607028754, "grad_norm": 0.2978529632091522, "learning_rate": 5e-07, "loss": 0.0034, "step": 794 }, { "epoch": 1.269968051118211, "grad_norm": 0.5247751474380493, "learning_rate": 5e-07, "loss": 0.0008, "step": 795 }, { "epoch": 1.2715654952076676, "grad_norm": 0.06592018157243729, "learning_rate": 5e-07, "loss": 0.0006, "step": 796 }, { "epoch": 1.2731629392971247, "grad_norm": 0.1426411271095276, "learning_rate": 5e-07, "loss": 0.0007, "step": 797 }, { "epoch": 1.2747603833865815, "grad_norm": 0.14098455011844635, "learning_rate": 5e-07, "loss": 0.0011, "step": 798 }, { "epoch": 1.2763578274760383, "grad_norm": 0.11731888353824615, "learning_rate": 5e-07, "loss": 0.0031, "step": 799 }, { "epoch": 1.2779552715654952, "grad_norm": 0.8000310063362122, "learning_rate": 5e-07, "loss": 0.0044, "step": 800 }, { "epoch": 1.279552715654952, "grad_norm": 0.8283488154411316, "learning_rate": 5e-07, "loss": 0.003, "step": 801 }, { "epoch": 1.281150159744409, "grad_norm": 0.6855292916297913, "learning_rate": 5e-07, "loss": 0.0027, "step": 802 }, { "epoch": 1.2827476038338659, "grad_norm": 0.12922891974449158, "learning_rate": 5e-07, "loss": 0.002, "step": 803 }, { "epoch": 1.2843450479233227, "grad_norm": 0.8001663088798523, "learning_rate": 5e-07, "loss": 0.0055, "step": 804 }, { "epoch": 1.2859424920127795, "grad_norm": 0.09436812996864319, "learning_rate": 5e-07, "loss": 0.0019, "step": 805 }, { "epoch": 1.2875399361022364, "grad_norm": 0.3059661388397217, "learning_rate": 5e-07, "loss": 0.002, "step": 806 }, { "epoch": 1.2891373801916932, "grad_norm": 0.22149628400802612, "learning_rate": 5e-07, "loss": 0.0025, "step": 807 }, { "epoch": 1.29073482428115, "grad_norm": 0.1476268321275711, "learning_rate": 5e-07, "loss": 0.0014, "step": 808 }, { "epoch": 1.292332268370607, "grad_norm": 0.11144911497831345, "learning_rate": 5e-07, "loss": 0.002, "step": 809 }, { "epoch": 1.293929712460064, "grad_norm": 0.2204524427652359, "learning_rate": 5e-07, "loss": 0.0031, "step": 810 }, { "epoch": 1.2955271565495208, "grad_norm": 0.07029737532138824, "learning_rate": 5e-07, "loss": 0.0009, "step": 811 }, { "epoch": 1.2971246006389776, "grad_norm": 0.09939467906951904, "learning_rate": 5e-07, "loss": 0.0032, "step": 812 }, { "epoch": 1.2987220447284344, "grad_norm": 0.1769058108329773, "learning_rate": 5e-07, "loss": 0.0015, "step": 813 }, { "epoch": 1.3003194888178915, "grad_norm": 0.3592926561832428, "learning_rate": 5e-07, "loss": 0.0015, "step": 814 }, { "epoch": 1.3019169329073481, "grad_norm": 0.2275543361902237, "learning_rate": 5e-07, "loss": 0.0019, "step": 815 }, { "epoch": 1.3035143769968052, "grad_norm": 0.07945606857538223, "learning_rate": 5e-07, "loss": 0.0012, "step": 816 }, { "epoch": 1.305111821086262, "grad_norm": 0.08711861073970795, "learning_rate": 5e-07, "loss": 0.0019, "step": 817 }, { "epoch": 1.3067092651757188, "grad_norm": 0.28836479783058167, "learning_rate": 5e-07, "loss": 0.0022, "step": 818 }, { "epoch": 1.3083067092651757, "grad_norm": 0.08159520477056503, "learning_rate": 5e-07, "loss": 0.0013, "step": 819 }, { "epoch": 1.3099041533546325, "grad_norm": 0.07466080784797668, "learning_rate": 5e-07, "loss": 0.0021, "step": 820 }, { "epoch": 1.3115015974440896, "grad_norm": 0.5451674461364746, "learning_rate": 5e-07, "loss": 0.0023, "step": 821 }, { "epoch": 1.3130990415335464, "grad_norm": 0.14322105050086975, "learning_rate": 5e-07, "loss": 0.0018, "step": 822 }, { "epoch": 1.3146964856230032, "grad_norm": 0.12718692421913147, "learning_rate": 5e-07, "loss": 0.0008, "step": 823 }, { "epoch": 1.31629392971246, "grad_norm": 0.13329675793647766, "learning_rate": 5e-07, "loss": 0.0025, "step": 824 }, { "epoch": 1.317891373801917, "grad_norm": 0.22774100303649902, "learning_rate": 5e-07, "loss": 0.0035, "step": 825 }, { "epoch": 1.3194888178913737, "grad_norm": 0.8411527276039124, "learning_rate": 5e-07, "loss": 0.0038, "step": 826 }, { "epoch": 1.3210862619808306, "grad_norm": 0.18316972255706787, "learning_rate": 5e-07, "loss": 0.0022, "step": 827 }, { "epoch": 1.3226837060702876, "grad_norm": 0.07081547379493713, "learning_rate": 5e-07, "loss": 0.0013, "step": 828 }, { "epoch": 1.3242811501597445, "grad_norm": 0.3037576973438263, "learning_rate": 5e-07, "loss": 0.0036, "step": 829 }, { "epoch": 1.3258785942492013, "grad_norm": 0.10193005204200745, "learning_rate": 5e-07, "loss": 0.001, "step": 830 }, { "epoch": 1.3274760383386581, "grad_norm": 0.10602962970733643, "learning_rate": 5e-07, "loss": 0.0012, "step": 831 }, { "epoch": 1.329073482428115, "grad_norm": 0.2908097505569458, "learning_rate": 5e-07, "loss": 0.0025, "step": 832 }, { "epoch": 1.330670926517572, "grad_norm": 0.3724125027656555, "learning_rate": 5e-07, "loss": 0.0018, "step": 833 }, { "epoch": 1.3322683706070286, "grad_norm": 0.11064999550580978, "learning_rate": 5e-07, "loss": 0.0023, "step": 834 }, { "epoch": 1.3338658146964857, "grad_norm": 0.228192538022995, "learning_rate": 5e-07, "loss": 0.0022, "step": 835 }, { "epoch": 1.3354632587859425, "grad_norm": 0.7090662717819214, "learning_rate": 5e-07, "loss": 0.0032, "step": 836 }, { "epoch": 1.3370607028753994, "grad_norm": 0.14330582320690155, "learning_rate": 5e-07, "loss": 0.003, "step": 837 }, { "epoch": 1.3386581469648562, "grad_norm": 0.22641916573047638, "learning_rate": 5e-07, "loss": 0.0033, "step": 838 }, { "epoch": 1.340255591054313, "grad_norm": 0.21619798243045807, "learning_rate": 5e-07, "loss": 0.0026, "step": 839 }, { "epoch": 1.34185303514377, "grad_norm": 0.10568831115961075, "learning_rate": 5e-07, "loss": 0.0017, "step": 840 }, { "epoch": 1.343450479233227, "grad_norm": 0.1081472709774971, "learning_rate": 5e-07, "loss": 0.0024, "step": 841 }, { "epoch": 1.3450479233226837, "grad_norm": 0.10540477186441422, "learning_rate": 5e-07, "loss": 0.0021, "step": 842 }, { "epoch": 1.3466453674121406, "grad_norm": 0.11737114936113358, "learning_rate": 5e-07, "loss": 0.0032, "step": 843 }, { "epoch": 1.3482428115015974, "grad_norm": 0.17713451385498047, "learning_rate": 5e-07, "loss": 0.002, "step": 844 }, { "epoch": 1.3498402555910542, "grad_norm": 0.2664765417575836, "learning_rate": 5e-07, "loss": 0.0019, "step": 845 }, { "epoch": 1.351437699680511, "grad_norm": 0.11203097552061081, "learning_rate": 5e-07, "loss": 0.0016, "step": 846 }, { "epoch": 1.3530351437699681, "grad_norm": 0.24115419387817383, "learning_rate": 5e-07, "loss": 0.0023, "step": 847 }, { "epoch": 1.354632587859425, "grad_norm": 0.13034223020076752, "learning_rate": 5e-07, "loss": 0.0021, "step": 848 }, { "epoch": 1.3562300319488818, "grad_norm": 1.2168869972229004, "learning_rate": 5e-07, "loss": 0.0027, "step": 849 }, { "epoch": 1.3578274760383386, "grad_norm": 0.04777536913752556, "learning_rate": 5e-07, "loss": 0.0005, "step": 850 }, { "epoch": 1.3594249201277955, "grad_norm": 0.14492247998714447, "learning_rate": 5e-07, "loss": 0.0015, "step": 851 }, { "epoch": 1.3610223642172525, "grad_norm": 0.11650210618972778, "learning_rate": 5e-07, "loss": 0.0026, "step": 852 }, { "epoch": 1.3626198083067091, "grad_norm": 0.13418716192245483, "learning_rate": 5e-07, "loss": 0.0009, "step": 853 }, { "epoch": 1.3642172523961662, "grad_norm": 1.2964794635772705, "learning_rate": 5e-07, "loss": 0.0038, "step": 854 }, { "epoch": 1.365814696485623, "grad_norm": 0.17427192628383636, "learning_rate": 5e-07, "loss": 0.0035, "step": 855 }, { "epoch": 1.3674121405750799, "grad_norm": 0.8252497911453247, "learning_rate": 5e-07, "loss": 0.0049, "step": 856 }, { "epoch": 1.3690095846645367, "grad_norm": 0.3598852753639221, "learning_rate": 5e-07, "loss": 0.0017, "step": 857 }, { "epoch": 1.3706070287539935, "grad_norm": 0.23161448538303375, "learning_rate": 5e-07, "loss": 0.0027, "step": 858 }, { "epoch": 1.3722044728434506, "grad_norm": 0.17750093340873718, "learning_rate": 5e-07, "loss": 0.0016, "step": 859 }, { "epoch": 1.3738019169329074, "grad_norm": 0.07043975591659546, "learning_rate": 5e-07, "loss": 0.0011, "step": 860 }, { "epoch": 1.3753993610223643, "grad_norm": 0.06552740186452866, "learning_rate": 5e-07, "loss": 0.0013, "step": 861 }, { "epoch": 1.376996805111821, "grad_norm": 0.167598158121109, "learning_rate": 5e-07, "loss": 0.0016, "step": 862 }, { "epoch": 1.378594249201278, "grad_norm": 0.8627744913101196, "learning_rate": 5e-07, "loss": 0.0009, "step": 863 }, { "epoch": 1.3801916932907348, "grad_norm": 0.23403562605381012, "learning_rate": 5e-07, "loss": 0.0017, "step": 864 }, { "epoch": 1.3817891373801916, "grad_norm": 1.3149168491363525, "learning_rate": 5e-07, "loss": 0.0043, "step": 865 }, { "epoch": 1.3833865814696487, "grad_norm": 0.16277176141738892, "learning_rate": 5e-07, "loss": 0.0024, "step": 866 }, { "epoch": 1.3849840255591055, "grad_norm": 0.1406593918800354, "learning_rate": 5e-07, "loss": 0.0015, "step": 867 }, { "epoch": 1.3865814696485623, "grad_norm": 0.24820701777935028, "learning_rate": 5e-07, "loss": 0.002, "step": 868 }, { "epoch": 1.3881789137380192, "grad_norm": 0.10332539677619934, "learning_rate": 5e-07, "loss": 0.0023, "step": 869 }, { "epoch": 1.389776357827476, "grad_norm": 0.16070209443569183, "learning_rate": 5e-07, "loss": 0.0007, "step": 870 }, { "epoch": 1.3913738019169328, "grad_norm": 0.16537906229496002, "learning_rate": 5e-07, "loss": 0.0017, "step": 871 }, { "epoch": 1.3929712460063897, "grad_norm": 0.2067955881357193, "learning_rate": 5e-07, "loss": 0.0011, "step": 872 }, { "epoch": 1.3945686900958467, "grad_norm": 0.1595505326986313, "learning_rate": 5e-07, "loss": 0.0008, "step": 873 }, { "epoch": 1.3961661341853036, "grad_norm": 0.7071959972381592, "learning_rate": 5e-07, "loss": 0.0031, "step": 874 }, { "epoch": 1.3977635782747604, "grad_norm": 0.08548534661531448, "learning_rate": 5e-07, "loss": 0.0005, "step": 875 }, { "epoch": 1.3993610223642172, "grad_norm": 1.5511912107467651, "learning_rate": 5e-07, "loss": 0.005, "step": 876 }, { "epoch": 1.400958466453674, "grad_norm": 0.622124969959259, "learning_rate": 5e-07, "loss": 0.0026, "step": 877 }, { "epoch": 1.4025559105431311, "grad_norm": 3.853699207305908, "learning_rate": 5e-07, "loss": 0.0054, "step": 878 }, { "epoch": 1.4041533546325877, "grad_norm": 0.07736141234636307, "learning_rate": 5e-07, "loss": 0.0011, "step": 879 }, { "epoch": 1.4057507987220448, "grad_norm": 0.2597537338733673, "learning_rate": 5e-07, "loss": 0.0024, "step": 880 }, { "epoch": 1.4073482428115016, "grad_norm": 0.09090615063905716, "learning_rate": 5e-07, "loss": 0.001, "step": 881 }, { "epoch": 1.4089456869009584, "grad_norm": 0.4754495620727539, "learning_rate": 5e-07, "loss": 0.0015, "step": 882 }, { "epoch": 1.4105431309904153, "grad_norm": 0.4713440537452698, "learning_rate": 5e-07, "loss": 0.0045, "step": 883 }, { "epoch": 1.4121405750798721, "grad_norm": 0.3018096387386322, "learning_rate": 5e-07, "loss": 0.0027, "step": 884 }, { "epoch": 1.4137380191693292, "grad_norm": 0.21582446992397308, "learning_rate": 5e-07, "loss": 0.0021, "step": 885 }, { "epoch": 1.415335463258786, "grad_norm": 0.28057220578193665, "learning_rate": 5e-07, "loss": 0.0012, "step": 886 }, { "epoch": 1.4169329073482428, "grad_norm": 0.25230610370635986, "learning_rate": 5e-07, "loss": 0.0026, "step": 887 }, { "epoch": 1.4185303514376997, "grad_norm": 0.3251979947090149, "learning_rate": 5e-07, "loss": 0.0016, "step": 888 }, { "epoch": 1.4201277955271565, "grad_norm": 1.2839736938476562, "learning_rate": 5e-07, "loss": 0.0045, "step": 889 }, { "epoch": 1.4217252396166133, "grad_norm": 0.6521425247192383, "learning_rate": 5e-07, "loss": 0.0023, "step": 890 }, { "epoch": 1.4233226837060702, "grad_norm": 0.19354148209095, "learning_rate": 5e-07, "loss": 0.0017, "step": 891 }, { "epoch": 1.4249201277955272, "grad_norm": 0.26791778206825256, "learning_rate": 5e-07, "loss": 0.0028, "step": 892 }, { "epoch": 1.426517571884984, "grad_norm": 0.9978561997413635, "learning_rate": 5e-07, "loss": 0.0024, "step": 893 }, { "epoch": 1.428115015974441, "grad_norm": 1.0886672735214233, "learning_rate": 5e-07, "loss": 0.0033, "step": 894 }, { "epoch": 1.4297124600638977, "grad_norm": 0.1294880360364914, "learning_rate": 5e-07, "loss": 0.0028, "step": 895 }, { "epoch": 1.4313099041533546, "grad_norm": 0.1127467229962349, "learning_rate": 5e-07, "loss": 0.0016, "step": 896 }, { "epoch": 1.4329073482428116, "grad_norm": 0.14215292036533356, "learning_rate": 5e-07, "loss": 0.002, "step": 897 }, { "epoch": 1.4345047923322682, "grad_norm": 0.11055054515600204, "learning_rate": 5e-07, "loss": 0.0018, "step": 898 }, { "epoch": 1.4361022364217253, "grad_norm": 0.8130955100059509, "learning_rate": 5e-07, "loss": 0.0029, "step": 899 }, { "epoch": 1.4376996805111821, "grad_norm": 0.6208978295326233, "learning_rate": 5e-07, "loss": 0.0026, "step": 900 }, { "epoch": 1.439297124600639, "grad_norm": 0.10077293962240219, "learning_rate": 5e-07, "loss": 0.0013, "step": 901 }, { "epoch": 1.4408945686900958, "grad_norm": 0.15302683413028717, "learning_rate": 5e-07, "loss": 0.0016, "step": 902 }, { "epoch": 1.4424920127795526, "grad_norm": 0.2711629867553711, "learning_rate": 5e-07, "loss": 0.0019, "step": 903 }, { "epoch": 1.4440894568690097, "grad_norm": 0.6740185022354126, "learning_rate": 5e-07, "loss": 0.0032, "step": 904 }, { "epoch": 1.4456869009584665, "grad_norm": 0.09207923710346222, "learning_rate": 5e-07, "loss": 0.0011, "step": 905 }, { "epoch": 1.4472843450479234, "grad_norm": 0.13100707530975342, "learning_rate": 5e-07, "loss": 0.0022, "step": 906 }, { "epoch": 1.4488817891373802, "grad_norm": 0.22404153645038605, "learning_rate": 5e-07, "loss": 0.0015, "step": 907 }, { "epoch": 1.450479233226837, "grad_norm": 0.19576634466648102, "learning_rate": 5e-07, "loss": 0.0015, "step": 908 }, { "epoch": 1.4520766773162939, "grad_norm": 0.3745575547218323, "learning_rate": 5e-07, "loss": 0.0027, "step": 909 }, { "epoch": 1.4536741214057507, "grad_norm": 0.3619185984134674, "learning_rate": 5e-07, "loss": 0.0049, "step": 910 }, { "epoch": 1.4552715654952078, "grad_norm": 0.11963102966547012, "learning_rate": 5e-07, "loss": 0.0021, "step": 911 }, { "epoch": 1.4568690095846646, "grad_norm": 0.6312240958213806, "learning_rate": 5e-07, "loss": 0.0028, "step": 912 }, { "epoch": 1.4584664536741214, "grad_norm": 0.44361862540245056, "learning_rate": 5e-07, "loss": 0.0029, "step": 913 }, { "epoch": 1.4600638977635783, "grad_norm": 0.40353521704673767, "learning_rate": 5e-07, "loss": 0.0025, "step": 914 }, { "epoch": 1.461661341853035, "grad_norm": 0.14119647443294525, "learning_rate": 5e-07, "loss": 0.0027, "step": 915 }, { "epoch": 1.4632587859424921, "grad_norm": 0.07256528735160828, "learning_rate": 5e-07, "loss": 0.0018, "step": 916 }, { "epoch": 1.4648562300319488, "grad_norm": 0.4794658422470093, "learning_rate": 5e-07, "loss": 0.0021, "step": 917 }, { "epoch": 1.4664536741214058, "grad_norm": 0.2629548907279968, "learning_rate": 5e-07, "loss": 0.0014, "step": 918 }, { "epoch": 1.4680511182108626, "grad_norm": 0.6776370406150818, "learning_rate": 5e-07, "loss": 0.0041, "step": 919 }, { "epoch": 1.4696485623003195, "grad_norm": 0.14644776284694672, "learning_rate": 5e-07, "loss": 0.0013, "step": 920 }, { "epoch": 1.4712460063897763, "grad_norm": 0.09278970211744308, "learning_rate": 5e-07, "loss": 0.0014, "step": 921 }, { "epoch": 1.4728434504792332, "grad_norm": 0.08292179554700851, "learning_rate": 5e-07, "loss": 0.0013, "step": 922 }, { "epoch": 1.4744408945686902, "grad_norm": 0.9622533321380615, "learning_rate": 5e-07, "loss": 0.0013, "step": 923 }, { "epoch": 1.476038338658147, "grad_norm": 0.17527416348457336, "learning_rate": 5e-07, "loss": 0.0018, "step": 924 }, { "epoch": 1.4776357827476039, "grad_norm": 0.07049129158258438, "learning_rate": 5e-07, "loss": 0.0014, "step": 925 }, { "epoch": 1.4792332268370607, "grad_norm": 0.07232940942049026, "learning_rate": 5e-07, "loss": 0.001, "step": 926 }, { "epoch": 1.4808306709265175, "grad_norm": 0.06436185538768768, "learning_rate": 5e-07, "loss": 0.0012, "step": 927 }, { "epoch": 1.4824281150159744, "grad_norm": 0.24670840799808502, "learning_rate": 5e-07, "loss": 0.0024, "step": 928 }, { "epoch": 1.4840255591054312, "grad_norm": 0.7605392932891846, "learning_rate": 5e-07, "loss": 0.0041, "step": 929 }, { "epoch": 1.4856230031948883, "grad_norm": 0.06602557003498077, "learning_rate": 5e-07, "loss": 0.0008, "step": 930 }, { "epoch": 1.487220447284345, "grad_norm": 0.15429018437862396, "learning_rate": 5e-07, "loss": 0.0021, "step": 931 }, { "epoch": 1.488817891373802, "grad_norm": 0.155389204621315, "learning_rate": 5e-07, "loss": 0.002, "step": 932 }, { "epoch": 1.4904153354632588, "grad_norm": 0.3137715458869934, "learning_rate": 5e-07, "loss": 0.003, "step": 933 }, { "epoch": 1.4920127795527156, "grad_norm": 0.13918346166610718, "learning_rate": 5e-07, "loss": 0.0016, "step": 934 }, { "epoch": 1.4936102236421724, "grad_norm": 0.161741241812706, "learning_rate": 5e-07, "loss": 0.002, "step": 935 }, { "epoch": 1.4952076677316293, "grad_norm": 0.2871551513671875, "learning_rate": 5e-07, "loss": 0.0018, "step": 936 }, { "epoch": 1.4968051118210863, "grad_norm": 0.3078411817550659, "learning_rate": 5e-07, "loss": 0.0015, "step": 937 }, { "epoch": 1.4984025559105432, "grad_norm": 0.1792358011007309, "learning_rate": 5e-07, "loss": 0.0018, "step": 938 }, { "epoch": 1.5, "grad_norm": 0.08885855227708817, "learning_rate": 5e-07, "loss": 0.0018, "step": 939 }, { "epoch": 1.5015974440894568, "grad_norm": 0.10388941317796707, "learning_rate": 5e-07, "loss": 0.0024, "step": 940 }, { "epoch": 1.5031948881789137, "grad_norm": 0.1879977583885193, "learning_rate": 5e-07, "loss": 0.0035, "step": 941 }, { "epoch": 1.5047923322683707, "grad_norm": 0.08109968155622482, "learning_rate": 5e-07, "loss": 0.0014, "step": 942 }, { "epoch": 1.5063897763578273, "grad_norm": 0.14684544503688812, "learning_rate": 5e-07, "loss": 0.001, "step": 943 }, { "epoch": 1.5079872204472844, "grad_norm": 0.250420480966568, "learning_rate": 5e-07, "loss": 0.0019, "step": 944 }, { "epoch": 1.5095846645367412, "grad_norm": 0.0925571396946907, "learning_rate": 5e-07, "loss": 0.0008, "step": 945 }, { "epoch": 1.511182108626198, "grad_norm": 0.1837424635887146, "learning_rate": 5e-07, "loss": 0.0024, "step": 946 }, { "epoch": 1.5127795527156551, "grad_norm": 0.17538028955459595, "learning_rate": 5e-07, "loss": 0.0021, "step": 947 }, { "epoch": 1.5143769968051117, "grad_norm": 0.08512210845947266, "learning_rate": 5e-07, "loss": 0.0017, "step": 948 }, { "epoch": 1.5159744408945688, "grad_norm": 0.08973109722137451, "learning_rate": 5e-07, "loss": 0.0009, "step": 949 }, { "epoch": 1.5175718849840254, "grad_norm": 0.1241975948214531, "learning_rate": 5e-07, "loss": 0.0018, "step": 950 }, { "epoch": 1.5191693290734825, "grad_norm": 3.0702695846557617, "learning_rate": 5e-07, "loss": 0.0019, "step": 951 }, { "epoch": 1.5207667731629393, "grad_norm": 0.07310276478528976, "learning_rate": 5e-07, "loss": 0.0006, "step": 952 }, { "epoch": 1.5223642172523961, "grad_norm": 0.2920989692211151, "learning_rate": 5e-07, "loss": 0.0035, "step": 953 }, { "epoch": 1.5239616613418532, "grad_norm": 0.5082775950431824, "learning_rate": 5e-07, "loss": 0.0026, "step": 954 }, { "epoch": 1.5255591054313098, "grad_norm": 0.1549331098794937, "learning_rate": 5e-07, "loss": 0.0024, "step": 955 }, { "epoch": 1.5271565495207668, "grad_norm": 0.1703399121761322, "learning_rate": 5e-07, "loss": 0.0017, "step": 956 }, { "epoch": 1.5287539936102237, "grad_norm": 0.21957652270793915, "learning_rate": 5e-07, "loss": 0.0021, "step": 957 }, { "epoch": 1.5303514376996805, "grad_norm": 0.3131657838821411, "learning_rate": 5e-07, "loss": 0.0021, "step": 958 }, { "epoch": 1.5319488817891374, "grad_norm": 0.3629818260669708, "learning_rate": 5e-07, "loss": 0.003, "step": 959 }, { "epoch": 1.5335463258785942, "grad_norm": 0.2978671193122864, "learning_rate": 5e-07, "loss": 0.002, "step": 960 }, { "epoch": 1.5351437699680512, "grad_norm": 0.26085761189460754, "learning_rate": 5e-07, "loss": 0.0038, "step": 961 }, { "epoch": 1.5367412140575079, "grad_norm": 0.13067355751991272, "learning_rate": 5e-07, "loss": 0.0032, "step": 962 }, { "epoch": 1.538338658146965, "grad_norm": 1.0939770936965942, "learning_rate": 5e-07, "loss": 0.0075, "step": 963 }, { "epoch": 1.5399361022364217, "grad_norm": 0.4009106755256653, "learning_rate": 5e-07, "loss": 0.0041, "step": 964 }, { "epoch": 1.5415335463258786, "grad_norm": 0.69194495677948, "learning_rate": 5e-07, "loss": 0.0071, "step": 965 }, { "epoch": 1.5431309904153354, "grad_norm": 0.15541456639766693, "learning_rate": 5e-07, "loss": 0.0019, "step": 966 }, { "epoch": 1.5447284345047922, "grad_norm": 0.1687278002500534, "learning_rate": 5e-07, "loss": 0.0024, "step": 967 }, { "epoch": 1.5463258785942493, "grad_norm": 0.08637956529855728, "learning_rate": 5e-07, "loss": 0.0004, "step": 968 }, { "epoch": 1.547923322683706, "grad_norm": 0.49107223749160767, "learning_rate": 5e-07, "loss": 0.0025, "step": 969 }, { "epoch": 1.549520766773163, "grad_norm": 0.40780824422836304, "learning_rate": 5e-07, "loss": 0.0022, "step": 970 }, { "epoch": 1.5511182108626198, "grad_norm": 0.14083029329776764, "learning_rate": 5e-07, "loss": 0.0023, "step": 971 }, { "epoch": 1.5527156549520766, "grad_norm": 0.1509598046541214, "learning_rate": 5e-07, "loss": 0.0025, "step": 972 }, { "epoch": 1.5543130990415337, "grad_norm": 0.22124463319778442, "learning_rate": 5e-07, "loss": 0.0017, "step": 973 }, { "epoch": 1.5559105431309903, "grad_norm": 0.29571980237960815, "learning_rate": 5e-07, "loss": 0.0026, "step": 974 }, { "epoch": 1.5575079872204474, "grad_norm": 0.12444217503070831, "learning_rate": 5e-07, "loss": 0.0026, "step": 975 }, { "epoch": 1.5591054313099042, "grad_norm": 0.19605819880962372, "learning_rate": 5e-07, "loss": 0.0032, "step": 976 }, { "epoch": 1.560702875399361, "grad_norm": 0.36710816621780396, "learning_rate": 5e-07, "loss": 0.0021, "step": 977 }, { "epoch": 1.5623003194888179, "grad_norm": 0.4394298493862152, "learning_rate": 5e-07, "loss": 0.0026, "step": 978 }, { "epoch": 1.5638977635782747, "grad_norm": 3.3451671600341797, "learning_rate": 5e-07, "loss": 0.0031, "step": 979 }, { "epoch": 1.5654952076677318, "grad_norm": 0.0950264111161232, "learning_rate": 5e-07, "loss": 0.0004, "step": 980 }, { "epoch": 1.5670926517571884, "grad_norm": 0.5253363847732544, "learning_rate": 5e-07, "loss": 0.004, "step": 981 }, { "epoch": 1.5686900958466454, "grad_norm": 0.08567991107702255, "learning_rate": 5e-07, "loss": 0.0024, "step": 982 }, { "epoch": 1.5702875399361023, "grad_norm": 0.21886466443538666, "learning_rate": 5e-07, "loss": 0.0017, "step": 983 }, { "epoch": 1.571884984025559, "grad_norm": 0.12182791531085968, "learning_rate": 5e-07, "loss": 0.0018, "step": 984 }, { "epoch": 1.573482428115016, "grad_norm": 0.10883834213018417, "learning_rate": 5e-07, "loss": 0.002, "step": 985 }, { "epoch": 1.5750798722044728, "grad_norm": 0.2287680059671402, "learning_rate": 5e-07, "loss": 0.0022, "step": 986 }, { "epoch": 1.5766773162939298, "grad_norm": 0.7266914248466492, "learning_rate": 5e-07, "loss": 0.0046, "step": 987 }, { "epoch": 1.5782747603833864, "grad_norm": 0.11474244296550751, "learning_rate": 5e-07, "loss": 0.0026, "step": 988 }, { "epoch": 1.5798722044728435, "grad_norm": 0.23904937505722046, "learning_rate": 5e-07, "loss": 0.0013, "step": 989 }, { "epoch": 1.5814696485623003, "grad_norm": 0.055184707045555115, "learning_rate": 5e-07, "loss": 0.0015, "step": 990 }, { "epoch": 1.5830670926517572, "grad_norm": 0.22353950142860413, "learning_rate": 5e-07, "loss": 0.0025, "step": 991 }, { "epoch": 1.5846645367412142, "grad_norm": 0.19132664799690247, "learning_rate": 5e-07, "loss": 0.0025, "step": 992 }, { "epoch": 1.5862619808306708, "grad_norm": 0.2778601348400116, "learning_rate": 5e-07, "loss": 0.0028, "step": 993 }, { "epoch": 1.5878594249201279, "grad_norm": 0.05275161191821098, "learning_rate": 5e-07, "loss": 0.0008, "step": 994 }, { "epoch": 1.5894568690095847, "grad_norm": 0.2460019290447235, "learning_rate": 5e-07, "loss": 0.0024, "step": 995 }, { "epoch": 1.5910543130990416, "grad_norm": 0.3010371923446655, "learning_rate": 5e-07, "loss": 0.0024, "step": 996 }, { "epoch": 1.5926517571884984, "grad_norm": 0.10904088616371155, "learning_rate": 5e-07, "loss": 0.0023, "step": 997 }, { "epoch": 1.5942492012779552, "grad_norm": 0.05686507746577263, "learning_rate": 5e-07, "loss": 0.0013, "step": 998 }, { "epoch": 1.5958466453674123, "grad_norm": 0.20314623415470123, "learning_rate": 5e-07, "loss": 0.0021, "step": 999 }, { "epoch": 1.5974440894568689, "grad_norm": 1.588057279586792, "learning_rate": 5e-07, "loss": 0.0023, "step": 1000 }, { "epoch": 1.599041533546326, "grad_norm": 0.48463931679725647, "learning_rate": 5e-07, "loss": 0.004, "step": 1001 }, { "epoch": 1.6006389776357828, "grad_norm": 0.6624436974525452, "learning_rate": 5e-07, "loss": 0.0026, "step": 1002 }, { "epoch": 1.6022364217252396, "grad_norm": 0.14726407825946808, "learning_rate": 5e-07, "loss": 0.0014, "step": 1003 }, { "epoch": 1.6038338658146964, "grad_norm": 0.7131458520889282, "learning_rate": 5e-07, "loss": 0.0024, "step": 1004 }, { "epoch": 1.6054313099041533, "grad_norm": 0.46038198471069336, "learning_rate": 5e-07, "loss": 0.0043, "step": 1005 }, { "epoch": 1.6070287539936103, "grad_norm": 0.2227628529071808, "learning_rate": 5e-07, "loss": 0.0009, "step": 1006 }, { "epoch": 1.608626198083067, "grad_norm": 0.13341805338859558, "learning_rate": 5e-07, "loss": 0.0007, "step": 1007 }, { "epoch": 1.610223642172524, "grad_norm": 0.07892493903636932, "learning_rate": 5e-07, "loss": 0.0009, "step": 1008 }, { "epoch": 1.6118210862619808, "grad_norm": 0.12278909236192703, "learning_rate": 5e-07, "loss": 0.0013, "step": 1009 }, { "epoch": 1.6134185303514377, "grad_norm": 0.5151563882827759, "learning_rate": 5e-07, "loss": 0.0032, "step": 1010 }, { "epoch": 1.6150159744408947, "grad_norm": 0.23579691350460052, "learning_rate": 5e-07, "loss": 0.0026, "step": 1011 }, { "epoch": 1.6166134185303513, "grad_norm": 0.05294935405254364, "learning_rate": 5e-07, "loss": 0.0014, "step": 1012 }, { "epoch": 1.6182108626198084, "grad_norm": 0.12360315769910812, "learning_rate": 5e-07, "loss": 0.0027, "step": 1013 }, { "epoch": 1.619808306709265, "grad_norm": 0.35527291893959045, "learning_rate": 5e-07, "loss": 0.0024, "step": 1014 }, { "epoch": 1.621405750798722, "grad_norm": 0.25084754824638367, "learning_rate": 5e-07, "loss": 0.003, "step": 1015 }, { "epoch": 1.623003194888179, "grad_norm": 0.13500788807868958, "learning_rate": 5e-07, "loss": 0.0033, "step": 1016 }, { "epoch": 1.6246006389776357, "grad_norm": 0.17942559719085693, "learning_rate": 5e-07, "loss": 0.0026, "step": 1017 }, { "epoch": 1.6261980830670928, "grad_norm": 0.30760741233825684, "learning_rate": 5e-07, "loss": 0.0023, "step": 1018 }, { "epoch": 1.6277955271565494, "grad_norm": 0.12108216434717178, "learning_rate": 5e-07, "loss": 0.0021, "step": 1019 }, { "epoch": 1.6293929712460065, "grad_norm": 0.36486342549324036, "learning_rate": 5e-07, "loss": 0.0026, "step": 1020 }, { "epoch": 1.6309904153354633, "grad_norm": 0.21493099629878998, "learning_rate": 5e-07, "loss": 0.0024, "step": 1021 }, { "epoch": 1.6325878594249201, "grad_norm": 0.2136039435863495, "learning_rate": 5e-07, "loss": 0.0018, "step": 1022 }, { "epoch": 1.634185303514377, "grad_norm": 0.19852106273174286, "learning_rate": 5e-07, "loss": 0.0027, "step": 1023 }, { "epoch": 1.6357827476038338, "grad_norm": 0.07942666113376617, "learning_rate": 5e-07, "loss": 0.001, "step": 1024 }, { "epoch": 1.6373801916932909, "grad_norm": 0.14488880336284637, "learning_rate": 5e-07, "loss": 0.0025, "step": 1025 }, { "epoch": 1.6389776357827475, "grad_norm": 0.07575450092554092, "learning_rate": 5e-07, "loss": 0.0016, "step": 1026 }, { "epoch": 1.6405750798722045, "grad_norm": 0.18136127293109894, "learning_rate": 5e-07, "loss": 0.0019, "step": 1027 }, { "epoch": 1.6421725239616614, "grad_norm": 0.09243427962064743, "learning_rate": 5e-07, "loss": 0.0015, "step": 1028 }, { "epoch": 1.6437699680511182, "grad_norm": 0.15622855722904205, "learning_rate": 5e-07, "loss": 0.0009, "step": 1029 }, { "epoch": 1.645367412140575, "grad_norm": 0.20699208974838257, "learning_rate": 5e-07, "loss": 0.0024, "step": 1030 }, { "epoch": 1.6469648562300319, "grad_norm": 0.17613987624645233, "learning_rate": 5e-07, "loss": 0.0024, "step": 1031 }, { "epoch": 1.648562300319489, "grad_norm": 0.10338564217090607, "learning_rate": 5e-07, "loss": 0.0024, "step": 1032 }, { "epoch": 1.6501597444089455, "grad_norm": 0.07882916927337646, "learning_rate": 5e-07, "loss": 0.0015, "step": 1033 }, { "epoch": 1.6517571884984026, "grad_norm": 0.14468686282634735, "learning_rate": 5e-07, "loss": 0.0014, "step": 1034 }, { "epoch": 1.6533546325878594, "grad_norm": 0.33097293972969055, "learning_rate": 5e-07, "loss": 0.0037, "step": 1035 }, { "epoch": 1.6549520766773163, "grad_norm": 0.46003833413124084, "learning_rate": 5e-07, "loss": 0.0009, "step": 1036 }, { "epoch": 1.6565495207667733, "grad_norm": 0.38343682885169983, "learning_rate": 5e-07, "loss": 0.0013, "step": 1037 }, { "epoch": 1.65814696485623, "grad_norm": 0.22435209155082703, "learning_rate": 5e-07, "loss": 0.002, "step": 1038 }, { "epoch": 1.659744408945687, "grad_norm": 0.10138783603906631, "learning_rate": 5e-07, "loss": 0.0014, "step": 1039 }, { "epoch": 1.6613418530351438, "grad_norm": 0.42318132519721985, "learning_rate": 5e-07, "loss": 0.0033, "step": 1040 }, { "epoch": 1.6629392971246006, "grad_norm": 0.10508925467729568, "learning_rate": 5e-07, "loss": 0.0019, "step": 1041 }, { "epoch": 1.6645367412140575, "grad_norm": 0.3548614978790283, "learning_rate": 5e-07, "loss": 0.0033, "step": 1042 }, { "epoch": 1.6661341853035143, "grad_norm": 0.22214291989803314, "learning_rate": 5e-07, "loss": 0.0019, "step": 1043 }, { "epoch": 1.6677316293929714, "grad_norm": 0.23011279106140137, "learning_rate": 5e-07, "loss": 0.0022, "step": 1044 }, { "epoch": 1.669329073482428, "grad_norm": 0.08691424876451492, "learning_rate": 5e-07, "loss": 0.001, "step": 1045 }, { "epoch": 1.670926517571885, "grad_norm": 0.6495136618614197, "learning_rate": 5e-07, "loss": 0.005, "step": 1046 }, { "epoch": 1.6725239616613419, "grad_norm": 0.20477008819580078, "learning_rate": 5e-07, "loss": 0.0017, "step": 1047 }, { "epoch": 1.6741214057507987, "grad_norm": 0.3025433123111725, "learning_rate": 5e-07, "loss": 0.0023, "step": 1048 }, { "epoch": 1.6757188498402555, "grad_norm": 0.1731894165277481, "learning_rate": 5e-07, "loss": 0.0017, "step": 1049 }, { "epoch": 1.6773162939297124, "grad_norm": 0.06371040642261505, "learning_rate": 5e-07, "loss": 0.0022, "step": 1050 }, { "epoch": 1.6789137380191694, "grad_norm": 0.14911451935768127, "learning_rate": 5e-07, "loss": 0.0012, "step": 1051 }, { "epoch": 1.680511182108626, "grad_norm": 0.09349862486124039, "learning_rate": 5e-07, "loss": 0.0012, "step": 1052 }, { "epoch": 1.682108626198083, "grad_norm": 0.21330159902572632, "learning_rate": 5e-07, "loss": 0.0026, "step": 1053 }, { "epoch": 1.68370607028754, "grad_norm": 0.13860629498958588, "learning_rate": 5e-07, "loss": 0.0018, "step": 1054 }, { "epoch": 1.6853035143769968, "grad_norm": 0.1365477740764618, "learning_rate": 5e-07, "loss": 0.0016, "step": 1055 }, { "epoch": 1.6869009584664538, "grad_norm": 0.17503461241722107, "learning_rate": 5e-07, "loss": 0.0023, "step": 1056 }, { "epoch": 1.6884984025559104, "grad_norm": 0.111385278403759, "learning_rate": 5e-07, "loss": 0.0029, "step": 1057 }, { "epoch": 1.6900958466453675, "grad_norm": 0.06377507001161575, "learning_rate": 5e-07, "loss": 0.0014, "step": 1058 }, { "epoch": 1.6916932907348243, "grad_norm": 0.0641062781214714, "learning_rate": 5e-07, "loss": 0.0023, "step": 1059 }, { "epoch": 1.6932907348242812, "grad_norm": 0.09682592004537582, "learning_rate": 5e-07, "loss": 0.0015, "step": 1060 }, { "epoch": 1.694888178913738, "grad_norm": 0.1440698206424713, "learning_rate": 5e-07, "loss": 0.0018, "step": 1061 }, { "epoch": 1.6964856230031948, "grad_norm": 0.04743053764104843, "learning_rate": 5e-07, "loss": 0.0012, "step": 1062 }, { "epoch": 1.6980830670926519, "grad_norm": 0.18199731409549713, "learning_rate": 5e-07, "loss": 0.0036, "step": 1063 }, { "epoch": 1.6996805111821085, "grad_norm": 0.053704094141721725, "learning_rate": 5e-07, "loss": 0.001, "step": 1064 }, { "epoch": 1.7012779552715656, "grad_norm": 0.10925064235925674, "learning_rate": 5e-07, "loss": 0.0017, "step": 1065 }, { "epoch": 1.7028753993610224, "grad_norm": 0.09475322812795639, "learning_rate": 5e-07, "loss": 0.0005, "step": 1066 }, { "epoch": 1.7044728434504792, "grad_norm": 0.14424358308315277, "learning_rate": 5e-07, "loss": 0.0021, "step": 1067 }, { "epoch": 1.706070287539936, "grad_norm": 0.07309069484472275, "learning_rate": 5e-07, "loss": 0.0023, "step": 1068 }, { "epoch": 1.707667731629393, "grad_norm": 0.17733772099018097, "learning_rate": 5e-07, "loss": 0.0035, "step": 1069 }, { "epoch": 1.70926517571885, "grad_norm": 0.9875695705413818, "learning_rate": 5e-07, "loss": 0.0086, "step": 1070 }, { "epoch": 1.7108626198083066, "grad_norm": 0.408796101808548, "learning_rate": 5e-07, "loss": 0.002, "step": 1071 }, { "epoch": 1.7124600638977636, "grad_norm": 0.17582127451896667, "learning_rate": 5e-07, "loss": 0.0016, "step": 1072 }, { "epoch": 1.7140575079872205, "grad_norm": 0.1303548812866211, "learning_rate": 5e-07, "loss": 0.0023, "step": 1073 }, { "epoch": 1.7156549520766773, "grad_norm": 0.10945620387792587, "learning_rate": 5e-07, "loss": 0.0023, "step": 1074 }, { "epoch": 1.7172523961661343, "grad_norm": 0.07630528509616852, "learning_rate": 5e-07, "loss": 0.0016, "step": 1075 }, { "epoch": 1.718849840255591, "grad_norm": 0.07517647743225098, "learning_rate": 5e-07, "loss": 0.0022, "step": 1076 }, { "epoch": 1.720447284345048, "grad_norm": 0.0669432058930397, "learning_rate": 5e-07, "loss": 0.0021, "step": 1077 }, { "epoch": 1.7220447284345048, "grad_norm": 0.1877792626619339, "learning_rate": 5e-07, "loss": 0.0013, "step": 1078 }, { "epoch": 1.7236421725239617, "grad_norm": 0.04336528107523918, "learning_rate": 5e-07, "loss": 0.001, "step": 1079 }, { "epoch": 1.7252396166134185, "grad_norm": 0.06632347404956818, "learning_rate": 5e-07, "loss": 0.0024, "step": 1080 }, { "epoch": 1.7268370607028753, "grad_norm": 0.09459614008665085, "learning_rate": 5e-07, "loss": 0.0021, "step": 1081 }, { "epoch": 1.7284345047923324, "grad_norm": 0.5509624481201172, "learning_rate": 5e-07, "loss": 0.0012, "step": 1082 }, { "epoch": 1.730031948881789, "grad_norm": 0.2771284878253937, "learning_rate": 5e-07, "loss": 0.002, "step": 1083 }, { "epoch": 1.731629392971246, "grad_norm": 0.05486772954463959, "learning_rate": 5e-07, "loss": 0.0015, "step": 1084 }, { "epoch": 1.733226837060703, "grad_norm": 0.2757831811904907, "learning_rate": 5e-07, "loss": 0.0032, "step": 1085 }, { "epoch": 1.7348242811501597, "grad_norm": 0.04002672806382179, "learning_rate": 5e-07, "loss": 0.0013, "step": 1086 }, { "epoch": 1.7364217252396166, "grad_norm": 0.06649244576692581, "learning_rate": 5e-07, "loss": 0.0003, "step": 1087 }, { "epoch": 1.7380191693290734, "grad_norm": 0.11669522523880005, "learning_rate": 5e-07, "loss": 0.0012, "step": 1088 }, { "epoch": 1.7396166134185305, "grad_norm": 0.12358154356479645, "learning_rate": 5e-07, "loss": 0.0023, "step": 1089 }, { "epoch": 1.741214057507987, "grad_norm": 0.0829407200217247, "learning_rate": 5e-07, "loss": 0.0012, "step": 1090 }, { "epoch": 1.7428115015974441, "grad_norm": 0.09355901926755905, "learning_rate": 5e-07, "loss": 0.0025, "step": 1091 }, { "epoch": 1.744408945686901, "grad_norm": 0.0956120640039444, "learning_rate": 5e-07, "loss": 0.0018, "step": 1092 }, { "epoch": 1.7460063897763578, "grad_norm": 0.4703699052333832, "learning_rate": 5e-07, "loss": 0.0055, "step": 1093 }, { "epoch": 1.7476038338658149, "grad_norm": 0.3924386203289032, "learning_rate": 5e-07, "loss": 0.0039, "step": 1094 }, { "epoch": 1.7492012779552715, "grad_norm": 0.16191145777702332, "learning_rate": 5e-07, "loss": 0.0019, "step": 1095 }, { "epoch": 1.7507987220447285, "grad_norm": 0.06510366499423981, "learning_rate": 5e-07, "loss": 0.0016, "step": 1096 }, { "epoch": 1.7523961661341851, "grad_norm": 0.19621287286281586, "learning_rate": 5e-07, "loss": 0.0022, "step": 1097 }, { "epoch": 1.7539936102236422, "grad_norm": 0.1870349496603012, "learning_rate": 5e-07, "loss": 0.0016, "step": 1098 }, { "epoch": 1.755591054313099, "grad_norm": 0.14898639917373657, "learning_rate": 5e-07, "loss": 0.0019, "step": 1099 }, { "epoch": 1.7571884984025559, "grad_norm": 0.15484674274921417, "learning_rate": 5e-07, "loss": 0.0024, "step": 1100 }, { "epoch": 1.758785942492013, "grad_norm": 0.1420629471540451, "learning_rate": 5e-07, "loss": 0.002, "step": 1101 }, { "epoch": 1.7603833865814695, "grad_norm": 0.12516824901103973, "learning_rate": 5e-07, "loss": 0.002, "step": 1102 }, { "epoch": 1.7619808306709266, "grad_norm": 0.09688828140497208, "learning_rate": 5e-07, "loss": 0.0021, "step": 1103 }, { "epoch": 1.7635782747603834, "grad_norm": 0.09412126988172531, "learning_rate": 5e-07, "loss": 0.0023, "step": 1104 }, { "epoch": 1.7651757188498403, "grad_norm": 1.9070310592651367, "learning_rate": 5e-07, "loss": 0.002, "step": 1105 }, { "epoch": 1.766773162939297, "grad_norm": 0.16800744831562042, "learning_rate": 5e-07, "loss": 0.0018, "step": 1106 }, { "epoch": 1.768370607028754, "grad_norm": 0.10442492365837097, "learning_rate": 5e-07, "loss": 0.0012, "step": 1107 }, { "epoch": 1.769968051118211, "grad_norm": 0.16451716423034668, "learning_rate": 5e-07, "loss": 0.0014, "step": 1108 }, { "epoch": 1.7715654952076676, "grad_norm": 0.8484131097793579, "learning_rate": 5e-07, "loss": 0.0039, "step": 1109 }, { "epoch": 1.7731629392971247, "grad_norm": 0.3258044719696045, "learning_rate": 5e-07, "loss": 0.0032, "step": 1110 }, { "epoch": 1.7747603833865815, "grad_norm": 0.17745210230350494, "learning_rate": 5e-07, "loss": 0.0023, "step": 1111 }, { "epoch": 1.7763578274760383, "grad_norm": 0.07777510583400726, "learning_rate": 5e-07, "loss": 0.0015, "step": 1112 }, { "epoch": 1.7779552715654952, "grad_norm": 0.08588631451129913, "learning_rate": 5e-07, "loss": 0.0023, "step": 1113 }, { "epoch": 1.779552715654952, "grad_norm": 0.6624420881271362, "learning_rate": 5e-07, "loss": 0.005, "step": 1114 }, { "epoch": 1.781150159744409, "grad_norm": 0.11576593667268753, "learning_rate": 5e-07, "loss": 0.0024, "step": 1115 }, { "epoch": 1.7827476038338657, "grad_norm": 0.859341561794281, "learning_rate": 5e-07, "loss": 0.0014, "step": 1116 }, { "epoch": 1.7843450479233227, "grad_norm": 0.10600177943706512, "learning_rate": 5e-07, "loss": 0.0021, "step": 1117 }, { "epoch": 1.7859424920127795, "grad_norm": 0.14881321787834167, "learning_rate": 5e-07, "loss": 0.0017, "step": 1118 }, { "epoch": 1.7875399361022364, "grad_norm": 0.05584081634879112, "learning_rate": 5e-07, "loss": 0.0017, "step": 1119 }, { "epoch": 1.7891373801916934, "grad_norm": 0.06437966227531433, "learning_rate": 5e-07, "loss": 0.0017, "step": 1120 }, { "epoch": 1.79073482428115, "grad_norm": 0.2528286874294281, "learning_rate": 5e-07, "loss": 0.0016, "step": 1121 }, { "epoch": 1.792332268370607, "grad_norm": 0.18848662078380585, "learning_rate": 5e-07, "loss": 0.0022, "step": 1122 }, { "epoch": 1.793929712460064, "grad_norm": 0.14006567001342773, "learning_rate": 5e-07, "loss": 0.0007, "step": 1123 }, { "epoch": 1.7955271565495208, "grad_norm": 0.07045773416757584, "learning_rate": 5e-07, "loss": 0.0009, "step": 1124 }, { "epoch": 1.7971246006389776, "grad_norm": 0.0998934954404831, "learning_rate": 5e-07, "loss": 0.0011, "step": 1125 }, { "epoch": 1.7987220447284344, "grad_norm": 0.09063876420259476, "learning_rate": 5e-07, "loss": 0.002, "step": 1126 }, { "epoch": 1.8003194888178915, "grad_norm": 0.2747619152069092, "learning_rate": 5e-07, "loss": 0.0009, "step": 1127 }, { "epoch": 1.8019169329073481, "grad_norm": 0.0671829804778099, "learning_rate": 5e-07, "loss": 0.0016, "step": 1128 }, { "epoch": 1.8035143769968052, "grad_norm": 0.23768068850040436, "learning_rate": 5e-07, "loss": 0.0025, "step": 1129 }, { "epoch": 1.805111821086262, "grad_norm": 0.14285339415073395, "learning_rate": 5e-07, "loss": 0.0023, "step": 1130 }, { "epoch": 1.8067092651757188, "grad_norm": 3.77717661857605, "learning_rate": 5e-07, "loss": 0.005, "step": 1131 }, { "epoch": 1.8083067092651757, "grad_norm": 0.24507032334804535, "learning_rate": 5e-07, "loss": 0.0027, "step": 1132 }, { "epoch": 1.8099041533546325, "grad_norm": 0.15166565775871277, "learning_rate": 5e-07, "loss": 0.0017, "step": 1133 }, { "epoch": 1.8115015974440896, "grad_norm": 0.1067359670996666, "learning_rate": 5e-07, "loss": 0.002, "step": 1134 }, { "epoch": 1.8130990415335462, "grad_norm": 0.18612955510616302, "learning_rate": 5e-07, "loss": 0.0018, "step": 1135 }, { "epoch": 1.8146964856230032, "grad_norm": 0.17414437234401703, "learning_rate": 5e-07, "loss": 0.0012, "step": 1136 }, { "epoch": 1.81629392971246, "grad_norm": 0.3064412474632263, "learning_rate": 5e-07, "loss": 0.0033, "step": 1137 }, { "epoch": 1.817891373801917, "grad_norm": 0.071932353079319, "learning_rate": 5e-07, "loss": 0.0015, "step": 1138 }, { "epoch": 1.819488817891374, "grad_norm": 0.1119665578007698, "learning_rate": 5e-07, "loss": 0.0017, "step": 1139 }, { "epoch": 1.8210862619808306, "grad_norm": 0.20450662076473236, "learning_rate": 5e-07, "loss": 0.0023, "step": 1140 }, { "epoch": 1.8226837060702876, "grad_norm": 0.13636766374111176, "learning_rate": 5e-07, "loss": 0.0021, "step": 1141 }, { "epoch": 1.8242811501597445, "grad_norm": 0.07295648753643036, "learning_rate": 5e-07, "loss": 0.0014, "step": 1142 }, { "epoch": 1.8258785942492013, "grad_norm": 0.17274846136569977, "learning_rate": 5e-07, "loss": 0.003, "step": 1143 }, { "epoch": 1.8274760383386581, "grad_norm": 0.14178737998008728, "learning_rate": 5e-07, "loss": 0.001, "step": 1144 }, { "epoch": 1.829073482428115, "grad_norm": 0.2096203863620758, "learning_rate": 5e-07, "loss": 0.0017, "step": 1145 }, { "epoch": 1.830670926517572, "grad_norm": 0.09562284499406815, "learning_rate": 5e-07, "loss": 0.0012, "step": 1146 }, { "epoch": 1.8322683706070286, "grad_norm": 0.10589710623025894, "learning_rate": 5e-07, "loss": 0.0013, "step": 1147 }, { "epoch": 1.8338658146964857, "grad_norm": 0.2623519003391266, "learning_rate": 5e-07, "loss": 0.0035, "step": 1148 }, { "epoch": 1.8354632587859425, "grad_norm": 0.06434428691864014, "learning_rate": 5e-07, "loss": 0.0009, "step": 1149 }, { "epoch": 1.8370607028753994, "grad_norm": 0.31623294949531555, "learning_rate": 5e-07, "loss": 0.0035, "step": 1150 }, { "epoch": 1.8386581469648562, "grad_norm": 0.11535608768463135, "learning_rate": 5e-07, "loss": 0.0021, "step": 1151 }, { "epoch": 1.840255591054313, "grad_norm": 0.25482088327407837, "learning_rate": 5e-07, "loss": 0.0006, "step": 1152 }, { "epoch": 1.84185303514377, "grad_norm": 0.1523568332195282, "learning_rate": 5e-07, "loss": 0.0018, "step": 1153 }, { "epoch": 1.8434504792332267, "grad_norm": 0.09644993394613266, "learning_rate": 5e-07, "loss": 0.0014, "step": 1154 }, { "epoch": 1.8450479233226837, "grad_norm": 0.30026814341545105, "learning_rate": 5e-07, "loss": 0.0036, "step": 1155 }, { "epoch": 1.8466453674121406, "grad_norm": 5.034646511077881, "learning_rate": 5e-07, "loss": 0.0016, "step": 1156 }, { "epoch": 1.8482428115015974, "grad_norm": 0.1875925064086914, "learning_rate": 5e-07, "loss": 0.0019, "step": 1157 }, { "epoch": 1.8498402555910545, "grad_norm": 0.15650227665901184, "learning_rate": 5e-07, "loss": 0.0016, "step": 1158 }, { "epoch": 1.851437699680511, "grad_norm": 0.17901848256587982, "learning_rate": 5e-07, "loss": 0.0018, "step": 1159 }, { "epoch": 1.8530351437699681, "grad_norm": 0.27782970666885376, "learning_rate": 5e-07, "loss": 0.0024, "step": 1160 }, { "epoch": 1.854632587859425, "grad_norm": 0.11091984808444977, "learning_rate": 5e-07, "loss": 0.0021, "step": 1161 }, { "epoch": 1.8562300319488818, "grad_norm": 0.3836122453212738, "learning_rate": 5e-07, "loss": 0.0016, "step": 1162 }, { "epoch": 1.8578274760383386, "grad_norm": 2.26542067527771, "learning_rate": 5e-07, "loss": 0.0028, "step": 1163 }, { "epoch": 1.8594249201277955, "grad_norm": 0.8616245985031128, "learning_rate": 5e-07, "loss": 0.0051, "step": 1164 }, { "epoch": 1.8610223642172525, "grad_norm": 0.1418013572692871, "learning_rate": 5e-07, "loss": 0.001, "step": 1165 }, { "epoch": 1.8626198083067091, "grad_norm": 0.20582883059978485, "learning_rate": 5e-07, "loss": 0.0028, "step": 1166 }, { "epoch": 1.8642172523961662, "grad_norm": 0.05279695987701416, "learning_rate": 5e-07, "loss": 0.0006, "step": 1167 }, { "epoch": 1.865814696485623, "grad_norm": 0.30152246356010437, "learning_rate": 5e-07, "loss": 0.002, "step": 1168 }, { "epoch": 1.8674121405750799, "grad_norm": 1.215949296951294, "learning_rate": 5e-07, "loss": 0.0042, "step": 1169 }, { "epoch": 1.8690095846645367, "grad_norm": 0.1583871841430664, "learning_rate": 5e-07, "loss": 0.0031, "step": 1170 }, { "epoch": 1.8706070287539935, "grad_norm": 0.18552114069461823, "learning_rate": 5e-07, "loss": 0.0025, "step": 1171 }, { "epoch": 1.8722044728434506, "grad_norm": 0.25958728790283203, "learning_rate": 5e-07, "loss": 0.0021, "step": 1172 }, { "epoch": 1.8738019169329072, "grad_norm": 0.3198787271976471, "learning_rate": 5e-07, "loss": 0.0018, "step": 1173 }, { "epoch": 1.8753993610223643, "grad_norm": 0.4293941557407379, "learning_rate": 5e-07, "loss": 0.003, "step": 1174 }, { "epoch": 1.876996805111821, "grad_norm": 0.12936322391033173, "learning_rate": 5e-07, "loss": 0.0012, "step": 1175 }, { "epoch": 1.878594249201278, "grad_norm": 0.26641571521759033, "learning_rate": 5e-07, "loss": 0.0033, "step": 1176 }, { "epoch": 1.880191693290735, "grad_norm": 1.098074197769165, "learning_rate": 5e-07, "loss": 0.0036, "step": 1177 }, { "epoch": 1.8817891373801916, "grad_norm": 0.11431043595075607, "learning_rate": 5e-07, "loss": 0.0014, "step": 1178 }, { "epoch": 1.8833865814696487, "grad_norm": 0.37071993947029114, "learning_rate": 5e-07, "loss": 0.0024, "step": 1179 }, { "epoch": 1.8849840255591053, "grad_norm": 0.08628468960523605, "learning_rate": 5e-07, "loss": 0.0016, "step": 1180 }, { "epoch": 1.8865814696485623, "grad_norm": 0.14398378133773804, "learning_rate": 5e-07, "loss": 0.0016, "step": 1181 }, { "epoch": 1.8881789137380192, "grad_norm": 0.35993340611457825, "learning_rate": 5e-07, "loss": 0.003, "step": 1182 }, { "epoch": 1.889776357827476, "grad_norm": 0.1559559851884842, "learning_rate": 5e-07, "loss": 0.0015, "step": 1183 }, { "epoch": 1.891373801916933, "grad_norm": 0.2564754784107208, "learning_rate": 5e-07, "loss": 0.0021, "step": 1184 }, { "epoch": 1.8929712460063897, "grad_norm": 0.11279148608446121, "learning_rate": 5e-07, "loss": 0.0005, "step": 1185 }, { "epoch": 1.8945686900958467, "grad_norm": 0.08808083087205887, "learning_rate": 5e-07, "loss": 0.0016, "step": 1186 }, { "epoch": 1.8961661341853036, "grad_norm": 0.10382703691720963, "learning_rate": 5e-07, "loss": 0.0004, "step": 1187 }, { "epoch": 1.8977635782747604, "grad_norm": 0.15366709232330322, "learning_rate": 5e-07, "loss": 0.0026, "step": 1188 }, { "epoch": 1.8993610223642172, "grad_norm": 0.10865357518196106, "learning_rate": 5e-07, "loss": 0.0021, "step": 1189 }, { "epoch": 1.900958466453674, "grad_norm": 0.24182426929473877, "learning_rate": 5e-07, "loss": 0.0014, "step": 1190 }, { "epoch": 1.9025559105431311, "grad_norm": 0.4804140627384186, "learning_rate": 5e-07, "loss": 0.0048, "step": 1191 }, { "epoch": 1.9041533546325877, "grad_norm": 0.23404286801815033, "learning_rate": 5e-07, "loss": 0.0021, "step": 1192 }, { "epoch": 1.9057507987220448, "grad_norm": 0.36086928844451904, "learning_rate": 5e-07, "loss": 0.0016, "step": 1193 }, { "epoch": 1.9073482428115016, "grad_norm": 0.5354902744293213, "learning_rate": 5e-07, "loss": 0.004, "step": 1194 }, { "epoch": 1.9089456869009584, "grad_norm": 0.11621209979057312, "learning_rate": 5e-07, "loss": 0.0019, "step": 1195 }, { "epoch": 1.9105431309904153, "grad_norm": 0.1943303495645523, "learning_rate": 5e-07, "loss": 0.0023, "step": 1196 }, { "epoch": 1.9121405750798721, "grad_norm": 0.2628275454044342, "learning_rate": 5e-07, "loss": 0.0025, "step": 1197 }, { "epoch": 1.9137380191693292, "grad_norm": 0.20775623619556427, "learning_rate": 5e-07, "loss": 0.0018, "step": 1198 }, { "epoch": 1.9153354632587858, "grad_norm": 0.12529809772968292, "learning_rate": 5e-07, "loss": 0.0021, "step": 1199 }, { "epoch": 1.9169329073482428, "grad_norm": 0.09771095961332321, "learning_rate": 5e-07, "loss": 0.0015, "step": 1200 }, { "epoch": 1.9185303514376997, "grad_norm": 0.10013385117053986, "learning_rate": 5e-07, "loss": 0.0016, "step": 1201 }, { "epoch": 1.9201277955271565, "grad_norm": 0.29045844078063965, "learning_rate": 5e-07, "loss": 0.0033, "step": 1202 }, { "epoch": 1.9217252396166136, "grad_norm": 0.15383721888065338, "learning_rate": 5e-07, "loss": 0.001, "step": 1203 }, { "epoch": 1.9233226837060702, "grad_norm": 0.6897266507148743, "learning_rate": 5e-07, "loss": 0.002, "step": 1204 }, { "epoch": 1.9249201277955272, "grad_norm": 0.15652583539485931, "learning_rate": 5e-07, "loss": 0.0014, "step": 1205 }, { "epoch": 1.926517571884984, "grad_norm": 0.0853152722120285, "learning_rate": 5e-07, "loss": 0.0017, "step": 1206 }, { "epoch": 1.928115015974441, "grad_norm": 0.11881982535123825, "learning_rate": 5e-07, "loss": 0.0021, "step": 1207 }, { "epoch": 1.9297124600638977, "grad_norm": 0.10427961498498917, "learning_rate": 5e-07, "loss": 0.0016, "step": 1208 }, { "epoch": 1.9313099041533546, "grad_norm": 0.1068776324391365, "learning_rate": 5e-07, "loss": 0.0014, "step": 1209 }, { "epoch": 1.9329073482428116, "grad_norm": 0.13248351216316223, "learning_rate": 5e-07, "loss": 0.0009, "step": 1210 }, { "epoch": 1.9345047923322682, "grad_norm": 0.25468289852142334, "learning_rate": 5e-07, "loss": 0.0023, "step": 1211 }, { "epoch": 1.9361022364217253, "grad_norm": 0.16791756451129913, "learning_rate": 5e-07, "loss": 0.0013, "step": 1212 }, { "epoch": 1.9376996805111821, "grad_norm": 0.13099579513072968, "learning_rate": 5e-07, "loss": 0.0019, "step": 1213 }, { "epoch": 1.939297124600639, "grad_norm": 0.26930350065231323, "learning_rate": 5e-07, "loss": 0.0035, "step": 1214 }, { "epoch": 1.9408945686900958, "grad_norm": 2.077728271484375, "learning_rate": 5e-07, "loss": 0.0022, "step": 1215 }, { "epoch": 1.9424920127795526, "grad_norm": 0.31897902488708496, "learning_rate": 5e-07, "loss": 0.0021, "step": 1216 }, { "epoch": 1.9440894568690097, "grad_norm": 0.07247374951839447, "learning_rate": 5e-07, "loss": 0.0014, "step": 1217 }, { "epoch": 1.9456869009584663, "grad_norm": 0.11995717883110046, "learning_rate": 5e-07, "loss": 0.0011, "step": 1218 }, { "epoch": 1.9472843450479234, "grad_norm": 0.2621629536151886, "learning_rate": 5e-07, "loss": 0.0017, "step": 1219 }, { "epoch": 1.9488817891373802, "grad_norm": 0.1726490557193756, "learning_rate": 5e-07, "loss": 0.0027, "step": 1220 }, { "epoch": 1.950479233226837, "grad_norm": 0.16272880136966705, "learning_rate": 5e-07, "loss": 0.0023, "step": 1221 }, { "epoch": 1.952076677316294, "grad_norm": 0.10036841779947281, "learning_rate": 5e-07, "loss": 0.0008, "step": 1222 }, { "epoch": 1.9536741214057507, "grad_norm": 0.7988063097000122, "learning_rate": 5e-07, "loss": 0.0046, "step": 1223 }, { "epoch": 1.9552715654952078, "grad_norm": 0.2782542407512665, "learning_rate": 5e-07, "loss": 0.0023, "step": 1224 }, { "epoch": 1.9568690095846646, "grad_norm": 0.2055627703666687, "learning_rate": 5e-07, "loss": 0.0008, "step": 1225 }, { "epoch": 1.9584664536741214, "grad_norm": 0.19908927381038666, "learning_rate": 5e-07, "loss": 0.0017, "step": 1226 }, { "epoch": 1.9600638977635783, "grad_norm": 0.21738676726818085, "learning_rate": 5e-07, "loss": 0.0017, "step": 1227 }, { "epoch": 1.961661341853035, "grad_norm": 0.16997866332530975, "learning_rate": 5e-07, "loss": 0.0021, "step": 1228 }, { "epoch": 1.9632587859424921, "grad_norm": 0.10853756219148636, "learning_rate": 5e-07, "loss": 0.0018, "step": 1229 }, { "epoch": 1.9648562300319488, "grad_norm": 0.2967356741428375, "learning_rate": 5e-07, "loss": 0.0025, "step": 1230 }, { "epoch": 1.9664536741214058, "grad_norm": 0.24150174856185913, "learning_rate": 5e-07, "loss": 0.0035, "step": 1231 }, { "epoch": 1.9680511182108626, "grad_norm": 0.09346190094947815, "learning_rate": 5e-07, "loss": 0.0012, "step": 1232 }, { "epoch": 1.9696485623003195, "grad_norm": 0.16779378056526184, "learning_rate": 5e-07, "loss": 0.002, "step": 1233 }, { "epoch": 1.9712460063897763, "grad_norm": 0.1114976555109024, "learning_rate": 5e-07, "loss": 0.0011, "step": 1234 }, { "epoch": 1.9728434504792332, "grad_norm": 0.45826414227485657, "learning_rate": 5e-07, "loss": 0.0032, "step": 1235 }, { "epoch": 1.9744408945686902, "grad_norm": 0.4743911921977997, "learning_rate": 5e-07, "loss": 0.0034, "step": 1236 }, { "epoch": 1.9760383386581468, "grad_norm": 0.144659623503685, "learning_rate": 5e-07, "loss": 0.0027, "step": 1237 }, { "epoch": 1.9776357827476039, "grad_norm": 0.9392029643058777, "learning_rate": 5e-07, "loss": 0.0033, "step": 1238 }, { "epoch": 1.9792332268370607, "grad_norm": 0.5866590738296509, "learning_rate": 5e-07, "loss": 0.0033, "step": 1239 }, { "epoch": 1.9808306709265175, "grad_norm": 0.2669076919555664, "learning_rate": 5e-07, "loss": 0.0033, "step": 1240 }, { "epoch": 1.9824281150159746, "grad_norm": 0.34604692459106445, "learning_rate": 5e-07, "loss": 0.0022, "step": 1241 }, { "epoch": 1.9840255591054312, "grad_norm": 0.27668237686157227, "learning_rate": 5e-07, "loss": 0.0012, "step": 1242 }, { "epoch": 1.9856230031948883, "grad_norm": 0.1560208648443222, "learning_rate": 5e-07, "loss": 0.0012, "step": 1243 }, { "epoch": 1.9872204472843449, "grad_norm": 0.35375985503196716, "learning_rate": 5e-07, "loss": 0.002, "step": 1244 }, { "epoch": 1.988817891373802, "grad_norm": 0.134053036570549, "learning_rate": 5e-07, "loss": 0.0014, "step": 1245 }, { "epoch": 1.9904153354632588, "grad_norm": 0.10250476002693176, "learning_rate": 5e-07, "loss": 0.0014, "step": 1246 }, { "epoch": 1.9920127795527156, "grad_norm": 0.13041843473911285, "learning_rate": 5e-07, "loss": 0.0011, "step": 1247 }, { "epoch": 1.9936102236421727, "grad_norm": 0.062465403228998184, "learning_rate": 5e-07, "loss": 0.0017, "step": 1248 }, { "epoch": 1.9952076677316293, "grad_norm": 0.10093759000301361, "learning_rate": 5e-07, "loss": 0.0022, "step": 1249 }, { "epoch": 1.9968051118210863, "grad_norm": 0.08476297557353973, "learning_rate": 5e-07, "loss": 0.0016, "step": 1250 }, { "epoch": 1.9984025559105432, "grad_norm": 0.22871960699558258, "learning_rate": 5e-07, "loss": 0.0018, "step": 1251 }, { "epoch": 2.0, "grad_norm": 0.19058428704738617, "learning_rate": 5e-07, "loss": 0.0019, "step": 1252 }, { "epoch": 2.001597444089457, "grad_norm": 0.2544306218624115, "learning_rate": 5e-07, "loss": 0.0023, "step": 1253 }, { "epoch": 2.0031948881789137, "grad_norm": 0.14264468848705292, "learning_rate": 5e-07, "loss": 0.0029, "step": 1254 }, { "epoch": 2.0047923322683707, "grad_norm": 0.18681064248085022, "learning_rate": 5e-07, "loss": 0.0018, "step": 1255 }, { "epoch": 2.0063897763578273, "grad_norm": 0.17501874268054962, "learning_rate": 5e-07, "loss": 0.0011, "step": 1256 }, { "epoch": 2.0079872204472844, "grad_norm": 0.1236346885561943, "learning_rate": 5e-07, "loss": 0.0022, "step": 1257 }, { "epoch": 2.009584664536741, "grad_norm": 0.08623294532299042, "learning_rate": 5e-07, "loss": 0.0008, "step": 1258 }, { "epoch": 2.011182108626198, "grad_norm": 0.12931625545024872, "learning_rate": 5e-07, "loss": 0.0019, "step": 1259 }, { "epoch": 2.012779552715655, "grad_norm": 0.10413195192813873, "learning_rate": 5e-07, "loss": 0.002, "step": 1260 }, { "epoch": 2.0143769968051117, "grad_norm": 0.23485882580280304, "learning_rate": 5e-07, "loss": 0.0024, "step": 1261 }, { "epoch": 2.015974440894569, "grad_norm": 0.07863974571228027, "learning_rate": 5e-07, "loss": 0.0005, "step": 1262 }, { "epoch": 2.0175718849840254, "grad_norm": 0.33484408259391785, "learning_rate": 5e-07, "loss": 0.0025, "step": 1263 }, { "epoch": 2.0191693290734825, "grad_norm": 0.31939253211021423, "learning_rate": 5e-07, "loss": 0.0028, "step": 1264 }, { "epoch": 2.0207667731629395, "grad_norm": 0.10941127687692642, "learning_rate": 5e-07, "loss": 0.0027, "step": 1265 }, { "epoch": 2.022364217252396, "grad_norm": 0.10784590244293213, "learning_rate": 5e-07, "loss": 0.0016, "step": 1266 }, { "epoch": 2.023961661341853, "grad_norm": 0.16436795890331268, "learning_rate": 5e-07, "loss": 0.001, "step": 1267 }, { "epoch": 2.02555910543131, "grad_norm": 0.09606944024562836, "learning_rate": 5e-07, "loss": 0.002, "step": 1268 }, { "epoch": 2.027156549520767, "grad_norm": 0.17230142652988434, "learning_rate": 5e-07, "loss": 0.0024, "step": 1269 }, { "epoch": 2.0287539936102235, "grad_norm": 0.5120985507965088, "learning_rate": 5e-07, "loss": 0.0031, "step": 1270 }, { "epoch": 2.0303514376996805, "grad_norm": 0.06792975217103958, "learning_rate": 5e-07, "loss": 0.0031, "step": 1271 }, { "epoch": 2.0319488817891376, "grad_norm": 0.12373680621385574, "learning_rate": 5e-07, "loss": 0.0014, "step": 1272 }, { "epoch": 2.033546325878594, "grad_norm": 0.10998155921697617, "learning_rate": 5e-07, "loss": 0.0011, "step": 1273 }, { "epoch": 2.0351437699680512, "grad_norm": 0.15938100218772888, "learning_rate": 5e-07, "loss": 0.0013, "step": 1274 }, { "epoch": 2.036741214057508, "grad_norm": 0.07384390383958817, "learning_rate": 5e-07, "loss": 0.0007, "step": 1275 }, { "epoch": 2.038338658146965, "grad_norm": 0.23505450785160065, "learning_rate": 5e-07, "loss": 0.0016, "step": 1276 }, { "epoch": 2.0399361022364215, "grad_norm": 1.6043593883514404, "learning_rate": 5e-07, "loss": 0.0044, "step": 1277 }, { "epoch": 2.0415335463258786, "grad_norm": 0.08369333297014236, "learning_rate": 5e-07, "loss": 0.001, "step": 1278 }, { "epoch": 2.0431309904153356, "grad_norm": 0.15266691148281097, "learning_rate": 5e-07, "loss": 0.0017, "step": 1279 }, { "epoch": 2.0447284345047922, "grad_norm": 0.1948157548904419, "learning_rate": 5e-07, "loss": 0.0011, "step": 1280 }, { "epoch": 2.0463258785942493, "grad_norm": 0.33584949374198914, "learning_rate": 5e-07, "loss": 0.0025, "step": 1281 }, { "epoch": 2.047923322683706, "grad_norm": 0.6691318154335022, "learning_rate": 5e-07, "loss": 0.0017, "step": 1282 }, { "epoch": 2.049520766773163, "grad_norm": 0.1382388472557068, "learning_rate": 5e-07, "loss": 0.0016, "step": 1283 }, { "epoch": 2.0511182108626196, "grad_norm": 0.12497832626104355, "learning_rate": 5e-07, "loss": 0.0027, "step": 1284 }, { "epoch": 2.0527156549520766, "grad_norm": 0.27157703042030334, "learning_rate": 5e-07, "loss": 0.0017, "step": 1285 }, { "epoch": 2.0543130990415337, "grad_norm": 0.2626846134662628, "learning_rate": 5e-07, "loss": 0.0029, "step": 1286 }, { "epoch": 2.0559105431309903, "grad_norm": 0.16861018538475037, "learning_rate": 5e-07, "loss": 0.0015, "step": 1287 }, { "epoch": 2.0575079872204474, "grad_norm": 0.049949146807193756, "learning_rate": 5e-07, "loss": 0.0004, "step": 1288 }, { "epoch": 2.059105431309904, "grad_norm": 0.0886000543832779, "learning_rate": 5e-07, "loss": 0.0013, "step": 1289 }, { "epoch": 2.060702875399361, "grad_norm": 0.20434832572937012, "learning_rate": 5e-07, "loss": 0.0022, "step": 1290 }, { "epoch": 2.062300319488818, "grad_norm": 0.11731091886758804, "learning_rate": 5e-07, "loss": 0.0015, "step": 1291 }, { "epoch": 2.0638977635782747, "grad_norm": 0.20026597380638123, "learning_rate": 5e-07, "loss": 0.0022, "step": 1292 }, { "epoch": 2.0654952076677318, "grad_norm": 0.27669885754585266, "learning_rate": 5e-07, "loss": 0.0026, "step": 1293 }, { "epoch": 2.0670926517571884, "grad_norm": 0.2785263657569885, "learning_rate": 5e-07, "loss": 0.0035, "step": 1294 }, { "epoch": 2.0686900958466454, "grad_norm": 0.15540768206119537, "learning_rate": 5e-07, "loss": 0.0011, "step": 1295 }, { "epoch": 2.070287539936102, "grad_norm": 0.09532786905765533, "learning_rate": 5e-07, "loss": 0.0027, "step": 1296 }, { "epoch": 2.071884984025559, "grad_norm": 0.1427212655544281, "learning_rate": 5e-07, "loss": 0.002, "step": 1297 }, { "epoch": 2.073482428115016, "grad_norm": 0.3582955300807953, "learning_rate": 5e-07, "loss": 0.003, "step": 1298 }, { "epoch": 2.0750798722044728, "grad_norm": 0.3676643669605255, "learning_rate": 5e-07, "loss": 0.0033, "step": 1299 }, { "epoch": 2.07667731629393, "grad_norm": 0.15875279903411865, "learning_rate": 5e-07, "loss": 0.002, "step": 1300 }, { "epoch": 2.0782747603833864, "grad_norm": 0.17880000174045563, "learning_rate": 5e-07, "loss": 0.002, "step": 1301 }, { "epoch": 2.0798722044728435, "grad_norm": 0.09481951594352722, "learning_rate": 5e-07, "loss": 0.0011, "step": 1302 }, { "epoch": 2.0814696485623, "grad_norm": 0.0720115378499031, "learning_rate": 5e-07, "loss": 0.001, "step": 1303 }, { "epoch": 2.083067092651757, "grad_norm": 0.1885913908481598, "learning_rate": 5e-07, "loss": 0.0012, "step": 1304 }, { "epoch": 2.084664536741214, "grad_norm": 0.07471724599599838, "learning_rate": 5e-07, "loss": 0.0029, "step": 1305 }, { "epoch": 2.086261980830671, "grad_norm": 0.10295652598142624, "learning_rate": 5e-07, "loss": 0.0019, "step": 1306 }, { "epoch": 2.087859424920128, "grad_norm": 0.48676663637161255, "learning_rate": 5e-07, "loss": 0.0021, "step": 1307 }, { "epoch": 2.0894568690095845, "grad_norm": 0.06689424067735672, "learning_rate": 5e-07, "loss": 0.0014, "step": 1308 }, { "epoch": 2.0910543130990416, "grad_norm": 0.3771530091762543, "learning_rate": 5e-07, "loss": 0.0028, "step": 1309 }, { "epoch": 2.0926517571884986, "grad_norm": 0.37652865052223206, "learning_rate": 5e-07, "loss": 0.0018, "step": 1310 }, { "epoch": 2.094249201277955, "grad_norm": 0.0743250697851181, "learning_rate": 5e-07, "loss": 0.0016, "step": 1311 }, { "epoch": 2.0958466453674123, "grad_norm": 0.1055615022778511, "learning_rate": 5e-07, "loss": 0.001, "step": 1312 }, { "epoch": 2.097444089456869, "grad_norm": 0.25487279891967773, "learning_rate": 5e-07, "loss": 0.0021, "step": 1313 }, { "epoch": 2.099041533546326, "grad_norm": 0.13495591282844543, "learning_rate": 5e-07, "loss": 0.0024, "step": 1314 }, { "epoch": 2.1006389776357826, "grad_norm": 0.17297884821891785, "learning_rate": 5e-07, "loss": 0.0017, "step": 1315 }, { "epoch": 2.1022364217252396, "grad_norm": 0.14259670674800873, "learning_rate": 5e-07, "loss": 0.0014, "step": 1316 }, { "epoch": 2.1038338658146967, "grad_norm": 0.06441140174865723, "learning_rate": 5e-07, "loss": 0.0025, "step": 1317 }, { "epoch": 2.1054313099041533, "grad_norm": 0.0832604393362999, "learning_rate": 5e-07, "loss": 0.0021, "step": 1318 }, { "epoch": 2.1070287539936103, "grad_norm": 0.6112750172615051, "learning_rate": 5e-07, "loss": 0.0016, "step": 1319 }, { "epoch": 2.108626198083067, "grad_norm": 0.14800317585468292, "learning_rate": 5e-07, "loss": 0.0012, "step": 1320 }, { "epoch": 2.110223642172524, "grad_norm": 0.389436811208725, "learning_rate": 5e-07, "loss": 0.0025, "step": 1321 }, { "epoch": 2.1118210862619806, "grad_norm": 0.1899157464504242, "learning_rate": 5e-07, "loss": 0.002, "step": 1322 }, { "epoch": 2.1134185303514377, "grad_norm": 0.08389966934919357, "learning_rate": 5e-07, "loss": 0.0017, "step": 1323 }, { "epoch": 2.1150159744408947, "grad_norm": 0.10899113118648529, "learning_rate": 5e-07, "loss": 0.0014, "step": 1324 }, { "epoch": 2.1166134185303513, "grad_norm": 0.10662740468978882, "learning_rate": 5e-07, "loss": 0.0018, "step": 1325 }, { "epoch": 2.1182108626198084, "grad_norm": 0.11692536622285843, "learning_rate": 5e-07, "loss": 0.0015, "step": 1326 }, { "epoch": 2.119808306709265, "grad_norm": 0.09900128096342087, "learning_rate": 5e-07, "loss": 0.0022, "step": 1327 }, { "epoch": 2.121405750798722, "grad_norm": 0.08334943652153015, "learning_rate": 5e-07, "loss": 0.0007, "step": 1328 }, { "epoch": 2.123003194888179, "grad_norm": 0.10552877187728882, "learning_rate": 5e-07, "loss": 0.0021, "step": 1329 }, { "epoch": 2.1246006389776357, "grad_norm": 0.8137688040733337, "learning_rate": 5e-07, "loss": 0.0014, "step": 1330 }, { "epoch": 2.126198083067093, "grad_norm": 0.06990483403205872, "learning_rate": 5e-07, "loss": 0.0009, "step": 1331 }, { "epoch": 2.1277955271565494, "grad_norm": 0.1675483137369156, "learning_rate": 5e-07, "loss": 0.0021, "step": 1332 }, { "epoch": 2.1293929712460065, "grad_norm": 0.2427297979593277, "learning_rate": 5e-07, "loss": 0.0024, "step": 1333 }, { "epoch": 2.130990415335463, "grad_norm": 0.1828804910182953, "learning_rate": 5e-07, "loss": 0.002, "step": 1334 }, { "epoch": 2.13258785942492, "grad_norm": 0.35362717509269714, "learning_rate": 5e-07, "loss": 0.0029, "step": 1335 }, { "epoch": 2.134185303514377, "grad_norm": 0.21310366690158844, "learning_rate": 5e-07, "loss": 0.0016, "step": 1336 }, { "epoch": 2.135782747603834, "grad_norm": 0.14100836217403412, "learning_rate": 5e-07, "loss": 0.0021, "step": 1337 }, { "epoch": 2.137380191693291, "grad_norm": 0.10196174681186676, "learning_rate": 5e-07, "loss": 0.0009, "step": 1338 }, { "epoch": 2.1389776357827475, "grad_norm": 0.2148902863264084, "learning_rate": 5e-07, "loss": 0.0016, "step": 1339 }, { "epoch": 2.1405750798722045, "grad_norm": 0.05017423257231712, "learning_rate": 5e-07, "loss": 0.0007, "step": 1340 }, { "epoch": 2.142172523961661, "grad_norm": 2.62032151222229, "learning_rate": 5e-07, "loss": 0.0035, "step": 1341 }, { "epoch": 2.143769968051118, "grad_norm": 0.23405258357524872, "learning_rate": 5e-07, "loss": 0.0018, "step": 1342 }, { "epoch": 2.1453674121405752, "grad_norm": 0.05571659281849861, "learning_rate": 5e-07, "loss": 0.0018, "step": 1343 }, { "epoch": 2.146964856230032, "grad_norm": 0.17637008428573608, "learning_rate": 5e-07, "loss": 0.0011, "step": 1344 }, { "epoch": 2.148562300319489, "grad_norm": 0.10391382873058319, "learning_rate": 5e-07, "loss": 0.0011, "step": 1345 }, { "epoch": 2.1501597444089455, "grad_norm": 0.5443282723426819, "learning_rate": 5e-07, "loss": 0.0027, "step": 1346 }, { "epoch": 2.1517571884984026, "grad_norm": 0.09875518828630447, "learning_rate": 5e-07, "loss": 0.0019, "step": 1347 }, { "epoch": 2.1533546325878596, "grad_norm": 0.3345401883125305, "learning_rate": 5e-07, "loss": 0.0022, "step": 1348 }, { "epoch": 2.1549520766773163, "grad_norm": 0.14103494584560394, "learning_rate": 5e-07, "loss": 0.0012, "step": 1349 }, { "epoch": 2.1565495207667733, "grad_norm": 0.1581650972366333, "learning_rate": 5e-07, "loss": 0.0023, "step": 1350 }, { "epoch": 2.15814696485623, "grad_norm": 0.1288004070520401, "learning_rate": 5e-07, "loss": 0.0009, "step": 1351 }, { "epoch": 2.159744408945687, "grad_norm": 0.19039283692836761, "learning_rate": 5e-07, "loss": 0.0026, "step": 1352 }, { "epoch": 2.1613418530351436, "grad_norm": 0.15048496425151825, "learning_rate": 5e-07, "loss": 0.0009, "step": 1353 }, { "epoch": 2.1629392971246006, "grad_norm": 0.14754633605480194, "learning_rate": 5e-07, "loss": 0.0021, "step": 1354 }, { "epoch": 2.1645367412140577, "grad_norm": 0.20571894943714142, "learning_rate": 5e-07, "loss": 0.0033, "step": 1355 }, { "epoch": 2.1661341853035143, "grad_norm": 0.12298233062028885, "learning_rate": 5e-07, "loss": 0.0018, "step": 1356 }, { "epoch": 2.1677316293929714, "grad_norm": 0.060658592730760574, "learning_rate": 5e-07, "loss": 0.0007, "step": 1357 }, { "epoch": 2.169329073482428, "grad_norm": 0.2829779088497162, "learning_rate": 5e-07, "loss": 0.0015, "step": 1358 }, { "epoch": 2.170926517571885, "grad_norm": 0.1882810741662979, "learning_rate": 5e-07, "loss": 0.0011, "step": 1359 }, { "epoch": 2.1725239616613417, "grad_norm": 0.5703380703926086, "learning_rate": 5e-07, "loss": 0.0039, "step": 1360 }, { "epoch": 2.1741214057507987, "grad_norm": 0.7452375292778015, "learning_rate": 5e-07, "loss": 0.0052, "step": 1361 }, { "epoch": 2.1757188498402558, "grad_norm": 0.593784749507904, "learning_rate": 5e-07, "loss": 0.0041, "step": 1362 }, { "epoch": 2.1773162939297124, "grad_norm": 0.472791850566864, "learning_rate": 5e-07, "loss": 0.0041, "step": 1363 }, { "epoch": 2.1789137380191694, "grad_norm": 0.2769353687763214, "learning_rate": 5e-07, "loss": 0.0018, "step": 1364 }, { "epoch": 2.180511182108626, "grad_norm": 0.5703862309455872, "learning_rate": 5e-07, "loss": 0.0031, "step": 1365 }, { "epoch": 2.182108626198083, "grad_norm": 0.19744770228862762, "learning_rate": 5e-07, "loss": 0.0015, "step": 1366 }, { "epoch": 2.18370607028754, "grad_norm": 0.7886673212051392, "learning_rate": 5e-07, "loss": 0.0043, "step": 1367 }, { "epoch": 2.1853035143769968, "grad_norm": 0.13506360352039337, "learning_rate": 5e-07, "loss": 0.001, "step": 1368 }, { "epoch": 2.186900958466454, "grad_norm": 0.2344270497560501, "learning_rate": 5e-07, "loss": 0.0016, "step": 1369 }, { "epoch": 2.1884984025559104, "grad_norm": 0.19468432664871216, "learning_rate": 5e-07, "loss": 0.002, "step": 1370 }, { "epoch": 2.1900958466453675, "grad_norm": 0.1934957057237625, "learning_rate": 5e-07, "loss": 0.0023, "step": 1371 }, { "epoch": 2.191693290734824, "grad_norm": 0.2196214497089386, "learning_rate": 5e-07, "loss": 0.0016, "step": 1372 }, { "epoch": 2.193290734824281, "grad_norm": 0.13796208798885345, "learning_rate": 5e-07, "loss": 0.0028, "step": 1373 }, { "epoch": 2.194888178913738, "grad_norm": 0.27561473846435547, "learning_rate": 5e-07, "loss": 0.0038, "step": 1374 }, { "epoch": 2.196485623003195, "grad_norm": 0.12160798162221909, "learning_rate": 5e-07, "loss": 0.0013, "step": 1375 }, { "epoch": 2.198083067092652, "grad_norm": 0.10912802815437317, "learning_rate": 5e-07, "loss": 0.0022, "step": 1376 }, { "epoch": 2.1996805111821085, "grad_norm": 0.09718818962574005, "learning_rate": 5e-07, "loss": 0.0008, "step": 1377 }, { "epoch": 2.2012779552715656, "grad_norm": 0.11051057279109955, "learning_rate": 5e-07, "loss": 0.0006, "step": 1378 }, { "epoch": 2.202875399361022, "grad_norm": 0.12758323550224304, "learning_rate": 5e-07, "loss": 0.0017, "step": 1379 }, { "epoch": 2.2044728434504792, "grad_norm": 0.08249010145664215, "learning_rate": 5e-07, "loss": 0.0012, "step": 1380 }, { "epoch": 2.2060702875399363, "grad_norm": 0.1392274647951126, "learning_rate": 5e-07, "loss": 0.0017, "step": 1381 }, { "epoch": 2.207667731629393, "grad_norm": 0.10520713776350021, "learning_rate": 5e-07, "loss": 0.0023, "step": 1382 }, { "epoch": 2.20926517571885, "grad_norm": 0.10856632888317108, "learning_rate": 5e-07, "loss": 0.0038, "step": 1383 }, { "epoch": 2.2108626198083066, "grad_norm": 0.08818315714597702, "learning_rate": 5e-07, "loss": 0.002, "step": 1384 }, { "epoch": 2.2124600638977636, "grad_norm": 0.13421356678009033, "learning_rate": 5e-07, "loss": 0.0016, "step": 1385 }, { "epoch": 2.2140575079872207, "grad_norm": 0.32427483797073364, "learning_rate": 5e-07, "loss": 0.0025, "step": 1386 }, { "epoch": 2.2156549520766773, "grad_norm": 0.2926160991191864, "learning_rate": 5e-07, "loss": 0.0019, "step": 1387 }, { "epoch": 2.2172523961661343, "grad_norm": 0.2298290878534317, "learning_rate": 5e-07, "loss": 0.0032, "step": 1388 }, { "epoch": 2.218849840255591, "grad_norm": 0.09151104092597961, "learning_rate": 5e-07, "loss": 0.0015, "step": 1389 }, { "epoch": 2.220447284345048, "grad_norm": 0.10674551874399185, "learning_rate": 5e-07, "loss": 0.0021, "step": 1390 }, { "epoch": 2.2220447284345046, "grad_norm": 0.2997196614742279, "learning_rate": 5e-07, "loss": 0.0025, "step": 1391 }, { "epoch": 2.2236421725239617, "grad_norm": 0.11480361223220825, "learning_rate": 5e-07, "loss": 0.0028, "step": 1392 }, { "epoch": 2.2252396166134187, "grad_norm": 0.07671017944812775, "learning_rate": 5e-07, "loss": 0.0019, "step": 1393 }, { "epoch": 2.2268370607028753, "grad_norm": 0.1539941430091858, "learning_rate": 5e-07, "loss": 0.0018, "step": 1394 }, { "epoch": 2.2284345047923324, "grad_norm": 0.12852609157562256, "learning_rate": 5e-07, "loss": 0.0008, "step": 1395 }, { "epoch": 2.230031948881789, "grad_norm": 0.13677293062210083, "learning_rate": 5e-07, "loss": 0.0018, "step": 1396 }, { "epoch": 2.231629392971246, "grad_norm": 0.05409352481365204, "learning_rate": 5e-07, "loss": 0.0011, "step": 1397 }, { "epoch": 2.2332268370607027, "grad_norm": 0.07923945039510727, "learning_rate": 5e-07, "loss": 0.0018, "step": 1398 }, { "epoch": 2.2348242811501597, "grad_norm": 0.10353945195674896, "learning_rate": 5e-07, "loss": 0.0017, "step": 1399 }, { "epoch": 2.236421725239617, "grad_norm": 0.24928894639015198, "learning_rate": 5e-07, "loss": 0.0019, "step": 1400 } ], "logging_steps": 1.0, "max_steps": 3130, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 387658364567552.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }