{ "best_metric": 1.22141695022583, "best_model_checkpoint": "miner_id_24/checkpoint-500", "epoch": 0.07017599618608716, "eval_steps": 100, "global_step": 552, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0001271304278733463, "grad_norm": 0.5326213836669922, "learning_rate": 2e-05, "loss": 1.7856, "step": 1 }, { "epoch": 0.0001271304278733463, "eval_loss": 1.7346340417861938, "eval_runtime": 1249.3361, "eval_samples_per_second": 4.002, "eval_steps_per_second": 1.001, "step": 1 }, { "epoch": 0.0002542608557466926, "grad_norm": 0.5624520778656006, "learning_rate": 4e-05, "loss": 1.7626, "step": 2 }, { "epoch": 0.00038139128362003893, "grad_norm": 0.5890633463859558, "learning_rate": 6e-05, "loss": 1.739, "step": 3 }, { "epoch": 0.0005085217114933852, "grad_norm": 0.5437400937080383, "learning_rate": 8e-05, "loss": 1.6671, "step": 4 }, { "epoch": 0.0006356521393667316, "grad_norm": 0.6639446020126343, "learning_rate": 0.0001, "loss": 1.7684, "step": 5 }, { "epoch": 0.0007627825672400779, "grad_norm": 0.7031175494194031, "learning_rate": 0.00012, "loss": 1.7247, "step": 6 }, { "epoch": 0.0008899129951134241, "grad_norm": 0.5311002731323242, "learning_rate": 0.00014, "loss": 1.6195, "step": 7 }, { "epoch": 0.0010170434229867704, "grad_norm": 0.25101518630981445, "learning_rate": 0.00016, "loss": 1.5182, "step": 8 }, { "epoch": 0.0011441738508601168, "grad_norm": 0.8389205932617188, "learning_rate": 0.00018, "loss": 1.6646, "step": 9 }, { "epoch": 0.0012713042787334632, "grad_norm": 0.9317983388900757, "learning_rate": 0.0002, "loss": 1.6395, "step": 10 }, { "epoch": 0.0013984347066068094, "grad_norm": 0.48066481947898865, "learning_rate": 0.00019999832015210023, "loss": 1.5921, "step": 11 }, { "epoch": 0.0015255651344801557, "grad_norm": 0.2073744535446167, "learning_rate": 0.00019999328066483865, "loss": 1.4335, "step": 12 }, { "epoch": 0.0016526955623535021, "grad_norm": 0.22661754488945007, "learning_rate": 0.0001999848817075267, "loss": 1.477, "step": 13 }, { "epoch": 0.0017798259902268483, "grad_norm": 0.3149760663509369, "learning_rate": 0.00019997312356234386, "loss": 1.5713, "step": 14 }, { "epoch": 0.0019069564181001947, "grad_norm": 0.31392771005630493, "learning_rate": 0.00019995800662432798, "loss": 1.5414, "step": 15 }, { "epoch": 0.002034086845973541, "grad_norm": 0.2748984396457672, "learning_rate": 0.0001999395314013622, "loss": 1.5452, "step": 16 }, { "epoch": 0.0021612172738468874, "grad_norm": 0.19064395129680634, "learning_rate": 0.00019991769851415781, "loss": 1.5742, "step": 17 }, { "epoch": 0.0022883477017202336, "grad_norm": 0.1578415036201477, "learning_rate": 0.00019989250869623343, "loss": 1.5214, "step": 18 }, { "epoch": 0.0024154781295935798, "grad_norm": 0.20229928195476532, "learning_rate": 0.0001998639627938903, "loss": 1.3921, "step": 19 }, { "epoch": 0.0025426085574669264, "grad_norm": 0.2705669403076172, "learning_rate": 0.00019983206176618388, "loss": 1.4712, "step": 20 }, { "epoch": 0.0026697389853402725, "grad_norm": 0.27215054631233215, "learning_rate": 0.00019979680668489165, "loss": 1.4969, "step": 21 }, { "epoch": 0.0027968694132136187, "grad_norm": 0.20431619882583618, "learning_rate": 0.00019975819873447717, "loss": 1.431, "step": 22 }, { "epoch": 0.0029239998410869653, "grad_norm": 0.14068549871444702, "learning_rate": 0.00019971623921205005, "loss": 1.4543, "step": 23 }, { "epoch": 0.0030511302689603115, "grad_norm": 0.1594925820827484, "learning_rate": 0.00019967092952732264, "loss": 1.364, "step": 24 }, { "epoch": 0.0031782606968336576, "grad_norm": 0.17738410830497742, "learning_rate": 0.00019962227120256252, "loss": 1.4377, "step": 25 }, { "epoch": 0.0033053911247070042, "grad_norm": 0.1752498745918274, "learning_rate": 0.00019957026587254134, "loss": 1.3827, "step": 26 }, { "epoch": 0.0034325215525803504, "grad_norm": 0.18004052340984344, "learning_rate": 0.00019951491528448004, "loss": 1.3867, "step": 27 }, { "epoch": 0.0035596519804536966, "grad_norm": 0.1525774598121643, "learning_rate": 0.00019945622129799, "loss": 1.4164, "step": 28 }, { "epoch": 0.003686782408327043, "grad_norm": 0.1710219830274582, "learning_rate": 0.00019939418588501057, "loss": 1.4155, "step": 29 }, { "epoch": 0.0038139128362003893, "grad_norm": 0.20903100073337555, "learning_rate": 0.000199328811129743, "loss": 1.5239, "step": 30 }, { "epoch": 0.003941043264073736, "grad_norm": 0.18399456143379211, "learning_rate": 0.00019926009922858006, "loss": 1.3889, "step": 31 }, { "epoch": 0.004068173691947082, "grad_norm": 0.13330113887786865, "learning_rate": 0.0001991880524900327, "loss": 1.3587, "step": 32 }, { "epoch": 0.004195304119820428, "grad_norm": 0.13807597756385803, "learning_rate": 0.00019911267333465218, "loss": 1.4211, "step": 33 }, { "epoch": 0.004322434547693775, "grad_norm": 0.1550034135580063, "learning_rate": 0.0001990339642949488, "loss": 1.4317, "step": 34 }, { "epoch": 0.004449564975567121, "grad_norm": 0.1827971190214157, "learning_rate": 0.00019895192801530685, "loss": 1.4176, "step": 35 }, { "epoch": 0.004576695403440467, "grad_norm": 0.16028065979480743, "learning_rate": 0.00019886656725189575, "loss": 1.4122, "step": 36 }, { "epoch": 0.004703825831313814, "grad_norm": 0.14322146773338318, "learning_rate": 0.00019877788487257753, "loss": 1.423, "step": 37 }, { "epoch": 0.0048309562591871595, "grad_norm": 0.1685505211353302, "learning_rate": 0.00019868588385681032, "loss": 1.3702, "step": 38 }, { "epoch": 0.004958086687060506, "grad_norm": 0.1632552444934845, "learning_rate": 0.00019859056729554844, "loss": 1.3164, "step": 39 }, { "epoch": 0.005085217114933853, "grad_norm": 0.17149530351161957, "learning_rate": 0.00019849193839113833, "loss": 1.2799, "step": 40 }, { "epoch": 0.0052123475428071985, "grad_norm": 0.14993086457252502, "learning_rate": 0.00019839000045721118, "loss": 1.3412, "step": 41 }, { "epoch": 0.005339477970680545, "grad_norm": 0.15981823205947876, "learning_rate": 0.00019828475691857145, "loss": 1.3698, "step": 42 }, { "epoch": 0.005466608398553892, "grad_norm": 0.15311439335346222, "learning_rate": 0.00019817621131108196, "loss": 1.3792, "step": 43 }, { "epoch": 0.005593738826427237, "grad_norm": 0.19757720828056335, "learning_rate": 0.00019806436728154485, "loss": 1.4082, "step": 44 }, { "epoch": 0.005720869254300584, "grad_norm": 0.15765132009983063, "learning_rate": 0.00019794922858757928, "loss": 1.282, "step": 45 }, { "epoch": 0.005847999682173931, "grad_norm": 0.15940701961517334, "learning_rate": 0.00019783079909749515, "loss": 1.4016, "step": 46 }, { "epoch": 0.005975130110047276, "grad_norm": 0.1622331291437149, "learning_rate": 0.00019770908279016309, "loss": 1.3624, "step": 47 }, { "epoch": 0.006102260537920623, "grad_norm": 0.14866457879543304, "learning_rate": 0.00019758408375488071, "loss": 1.2807, "step": 48 }, { "epoch": 0.0062293909657939696, "grad_norm": 0.16648173332214355, "learning_rate": 0.00019745580619123535, "loss": 1.3617, "step": 49 }, { "epoch": 0.006356521393667315, "grad_norm": 0.16083629429340363, "learning_rate": 0.00019732425440896297, "loss": 1.3903, "step": 50 }, { "epoch": 0.006483651821540662, "grad_norm": 0.18012581765651703, "learning_rate": 0.00019718943282780323, "loss": 1.3472, "step": 51 }, { "epoch": 0.0066107822494140085, "grad_norm": 0.16914351284503937, "learning_rate": 0.00019705134597735113, "loss": 1.3765, "step": 52 }, { "epoch": 0.006737912677287354, "grad_norm": 0.15967325866222382, "learning_rate": 0.00019690999849690484, "loss": 1.3312, "step": 53 }, { "epoch": 0.006865043105160701, "grad_norm": 0.15996071696281433, "learning_rate": 0.00019676539513530968, "loss": 1.4227, "step": 54 }, { "epoch": 0.006992173533034047, "grad_norm": 0.1715613752603531, "learning_rate": 0.0001966175407507987, "loss": 1.3634, "step": 55 }, { "epoch": 0.007119303960907393, "grad_norm": 0.16633041203022003, "learning_rate": 0.00019646644031082948, "loss": 1.3279, "step": 56 }, { "epoch": 0.00724643438878074, "grad_norm": 0.16247525811195374, "learning_rate": 0.00019631209889191712, "loss": 1.3721, "step": 57 }, { "epoch": 0.007373564816654086, "grad_norm": 0.1603326052427292, "learning_rate": 0.00019615452167946385, "loss": 1.3212, "step": 58 }, { "epoch": 0.007500695244527432, "grad_norm": 0.16569744050502777, "learning_rate": 0.00019599371396758456, "loss": 1.3224, "step": 59 }, { "epoch": 0.007627825672400779, "grad_norm": 0.16010916233062744, "learning_rate": 0.0001958296811589293, "loss": 1.3022, "step": 60 }, { "epoch": 0.007754956100274125, "grad_norm": 0.1680569052696228, "learning_rate": 0.00019566242876450137, "loss": 1.3197, "step": 61 }, { "epoch": 0.007882086528147472, "grad_norm": 0.16432645916938782, "learning_rate": 0.00019549196240347248, "loss": 1.3167, "step": 62 }, { "epoch": 0.008009216956020818, "grad_norm": 0.17412547767162323, "learning_rate": 0.00019531828780299383, "loss": 1.3196, "step": 63 }, { "epoch": 0.008136347383894163, "grad_norm": 0.1736118346452713, "learning_rate": 0.0001951414107980036, "loss": 1.2966, "step": 64 }, { "epoch": 0.00826347781176751, "grad_norm": 0.16254910826683044, "learning_rate": 0.00019496133733103112, "loss": 1.3416, "step": 65 }, { "epoch": 0.008390608239640857, "grad_norm": 0.16831637918949127, "learning_rate": 0.00019477807345199714, "loss": 1.3396, "step": 66 }, { "epoch": 0.008517738667514202, "grad_norm": 0.1759282648563385, "learning_rate": 0.00019459162531801046, "loss": 1.3101, "step": 67 }, { "epoch": 0.00864486909538755, "grad_norm": 0.17314572632312775, "learning_rate": 0.00019440199919316123, "loss": 1.4026, "step": 68 }, { "epoch": 0.008771999523260895, "grad_norm": 0.17074303328990936, "learning_rate": 0.00019420920144831044, "loss": 1.3088, "step": 69 }, { "epoch": 0.008899129951134241, "grad_norm": 0.17773644626140594, "learning_rate": 0.0001940132385608757, "loss": 1.32, "step": 70 }, { "epoch": 0.009026260379007589, "grad_norm": 0.1736891269683838, "learning_rate": 0.0001938141171146141, "loss": 1.2865, "step": 71 }, { "epoch": 0.009153390806880934, "grad_norm": 0.17593072354793549, "learning_rate": 0.0001936118437994003, "loss": 1.3276, "step": 72 }, { "epoch": 0.00928052123475428, "grad_norm": 0.16799978911876678, "learning_rate": 0.00019340642541100248, "loss": 1.2585, "step": 73 }, { "epoch": 0.009407651662627628, "grad_norm": 0.17271657288074493, "learning_rate": 0.00019319786885085364, "loss": 1.3838, "step": 74 }, { "epoch": 0.009534782090500973, "grad_norm": 0.18318656086921692, "learning_rate": 0.0001929861811258197, "loss": 1.3857, "step": 75 }, { "epoch": 0.009661912518374319, "grad_norm": 0.1850346177816391, "learning_rate": 0.0001927713693479643, "loss": 1.3884, "step": 76 }, { "epoch": 0.009789042946247667, "grad_norm": 0.1707659363746643, "learning_rate": 0.0001925534407343097, "loss": 1.2674, "step": 77 }, { "epoch": 0.009916173374121012, "grad_norm": 0.1803124099969864, "learning_rate": 0.0001923324026065944, "loss": 1.2899, "step": 78 }, { "epoch": 0.010043303801994358, "grad_norm": 0.18143856525421143, "learning_rate": 0.0001921082623910271, "loss": 1.2967, "step": 79 }, { "epoch": 0.010170434229867705, "grad_norm": 0.17903882265090942, "learning_rate": 0.00019188102761803717, "loss": 1.2913, "step": 80 }, { "epoch": 0.010297564657741051, "grad_norm": 0.181906595826149, "learning_rate": 0.00019165070592202173, "loss": 1.2568, "step": 81 }, { "epoch": 0.010424695085614397, "grad_norm": 0.1655733585357666, "learning_rate": 0.00019141730504108922, "loss": 1.2758, "step": 82 }, { "epoch": 0.010551825513487744, "grad_norm": 0.17457562685012817, "learning_rate": 0.00019118083281679913, "loss": 1.2506, "step": 83 }, { "epoch": 0.01067895594136109, "grad_norm": 0.18457446992397308, "learning_rate": 0.00019094129719389886, "loss": 1.3701, "step": 84 }, { "epoch": 0.010806086369234436, "grad_norm": 0.17702506482601166, "learning_rate": 0.0001906987062200567, "loss": 1.3071, "step": 85 }, { "epoch": 0.010933216797107783, "grad_norm": 0.1763213723897934, "learning_rate": 0.0001904530680455914, "loss": 1.2996, "step": 86 }, { "epoch": 0.011060347224981129, "grad_norm": 0.17649979889392853, "learning_rate": 0.0001902043909231984, "loss": 1.314, "step": 87 }, { "epoch": 0.011187477652854475, "grad_norm": 0.1817278414964676, "learning_rate": 0.00018995268320767252, "loss": 1.315, "step": 88 }, { "epoch": 0.011314608080727822, "grad_norm": 0.17668454349040985, "learning_rate": 0.0001896979533556273, "loss": 1.2914, "step": 89 }, { "epoch": 0.011441738508601168, "grad_norm": 0.17107272148132324, "learning_rate": 0.0001894402099252109, "loss": 1.2884, "step": 90 }, { "epoch": 0.011568868936474514, "grad_norm": 0.18352022767066956, "learning_rate": 0.0001891794615758185, "loss": 1.3404, "step": 91 }, { "epoch": 0.011695999364347861, "grad_norm": 0.17999856173992157, "learning_rate": 0.00018891571706780146, "loss": 1.3001, "step": 92 }, { "epoch": 0.011823129792221207, "grad_norm": 0.17374040186405182, "learning_rate": 0.00018864898526217293, "loss": 1.266, "step": 93 }, { "epoch": 0.011950260220094553, "grad_norm": 0.18675245344638824, "learning_rate": 0.0001883792751203102, "loss": 1.3347, "step": 94 }, { "epoch": 0.0120773906479679, "grad_norm": 0.18314674496650696, "learning_rate": 0.0001881065957036536, "loss": 1.3224, "step": 95 }, { "epoch": 0.012204521075841246, "grad_norm": 0.17483913898468018, "learning_rate": 0.00018783095617340193, "loss": 1.2926, "step": 96 }, { "epoch": 0.012331651503714592, "grad_norm": 0.19491428136825562, "learning_rate": 0.00018755236579020502, "loss": 1.2636, "step": 97 }, { "epoch": 0.012458781931587939, "grad_norm": 0.17128659784793854, "learning_rate": 0.0001872708339138522, "loss": 1.2653, "step": 98 }, { "epoch": 0.012585912359461285, "grad_norm": 0.172018900513649, "learning_rate": 0.00018698637000295816, "loss": 1.2686, "step": 99 }, { "epoch": 0.01271304278733463, "grad_norm": 0.18891726434230804, "learning_rate": 0.0001866989836146449, "loss": 1.4058, "step": 100 }, { "epoch": 0.01271304278733463, "eval_loss": 1.2906723022460938, "eval_runtime": 1258.4463, "eval_samples_per_second": 3.973, "eval_steps_per_second": 0.993, "step": 100 }, { "epoch": 0.012840173215207978, "grad_norm": 0.1781390905380249, "learning_rate": 0.0001864086844042209, "loss": 1.3021, "step": 101 }, { "epoch": 0.012967303643081324, "grad_norm": 0.17100100219249725, "learning_rate": 0.00018611548212485647, "loss": 1.2574, "step": 102 }, { "epoch": 0.01309443407095467, "grad_norm": 0.18398095667362213, "learning_rate": 0.00018581938662725632, "loss": 1.2839, "step": 103 }, { "epoch": 0.013221564498828017, "grad_norm": 0.18981115520000458, "learning_rate": 0.00018552040785932845, "loss": 1.3149, "step": 104 }, { "epoch": 0.013348694926701363, "grad_norm": 0.18872378766536713, "learning_rate": 0.00018521855586584995, "loss": 1.279, "step": 105 }, { "epoch": 0.013475825354574708, "grad_norm": 0.1824631690979004, "learning_rate": 0.00018491384078812959, "loss": 1.2743, "step": 106 }, { "epoch": 0.013602955782448056, "grad_norm": 0.1971443146467209, "learning_rate": 0.000184606272863667, "loss": 1.3365, "step": 107 }, { "epoch": 0.013730086210321402, "grad_norm": 0.19964328408241272, "learning_rate": 0.00018429586242580884, "loss": 1.3184, "step": 108 }, { "epoch": 0.013857216638194747, "grad_norm": 0.17624543607234955, "learning_rate": 0.00018398261990340152, "loss": 1.2755, "step": 109 }, { "epoch": 0.013984347066068095, "grad_norm": 0.18599238991737366, "learning_rate": 0.00018366655582044094, "loss": 1.3025, "step": 110 }, { "epoch": 0.01411147749394144, "grad_norm": 0.19051305949687958, "learning_rate": 0.00018334768079571884, "loss": 1.351, "step": 111 }, { "epoch": 0.014238607921814786, "grad_norm": 0.1858106255531311, "learning_rate": 0.00018302600554246601, "loss": 1.2386, "step": 112 }, { "epoch": 0.014365738349688134, "grad_norm": 0.17598244547843933, "learning_rate": 0.00018270154086799239, "loss": 1.2687, "step": 113 }, { "epoch": 0.01449286877756148, "grad_norm": 0.18105947971343994, "learning_rate": 0.00018237429767332405, "loss": 1.2843, "step": 114 }, { "epoch": 0.014619999205434825, "grad_norm": 0.18796177208423615, "learning_rate": 0.00018204428695283687, "loss": 1.2999, "step": 115 }, { "epoch": 0.014747129633308173, "grad_norm": 0.18702763319015503, "learning_rate": 0.00018171151979388714, "loss": 1.2391, "step": 116 }, { "epoch": 0.014874260061181518, "grad_norm": 0.17469799518585205, "learning_rate": 0.00018137600737643913, "loss": 1.2915, "step": 117 }, { "epoch": 0.015001390489054864, "grad_norm": 0.1871766746044159, "learning_rate": 0.00018103776097268942, "loss": 1.2429, "step": 118 }, { "epoch": 0.015128520916928212, "grad_norm": 0.18426093459129333, "learning_rate": 0.00018069679194668826, "loss": 1.2678, "step": 119 }, { "epoch": 0.015255651344801557, "grad_norm": 0.1830713450908661, "learning_rate": 0.0001803531117539577, "loss": 1.3231, "step": 120 }, { "epoch": 0.015382781772674903, "grad_norm": 0.19156108796596527, "learning_rate": 0.00018000673194110668, "loss": 1.3426, "step": 121 }, { "epoch": 0.01550991220054825, "grad_norm": 0.18232569098472595, "learning_rate": 0.00017965766414544326, "loss": 1.2227, "step": 122 }, { "epoch": 0.015637042628421596, "grad_norm": 0.18696987628936768, "learning_rate": 0.00017930592009458352, "loss": 1.2933, "step": 123 }, { "epoch": 0.015764173056294944, "grad_norm": 0.18148070573806763, "learning_rate": 0.00017895151160605757, "loss": 1.3598, "step": 124 }, { "epoch": 0.015891303484168288, "grad_norm": 0.1859319657087326, "learning_rate": 0.00017859445058691247, "loss": 1.2688, "step": 125 }, { "epoch": 0.016018433912041635, "grad_norm": 0.18133966624736786, "learning_rate": 0.00017823474903331233, "loss": 1.2912, "step": 126 }, { "epoch": 0.016145564339914983, "grad_norm": 0.16695751249790192, "learning_rate": 0.0001778724190301351, "loss": 1.2772, "step": 127 }, { "epoch": 0.016272694767788327, "grad_norm": 0.17694084346294403, "learning_rate": 0.0001775074727505667, "loss": 1.2998, "step": 128 }, { "epoch": 0.016399825195661674, "grad_norm": 0.18545518815517426, "learning_rate": 0.0001771399224556919, "loss": 1.2996, "step": 129 }, { "epoch": 0.01652695562353502, "grad_norm": 0.1763446033000946, "learning_rate": 0.00017676978049408263, "loss": 1.2942, "step": 130 }, { "epoch": 0.016654086051408366, "grad_norm": 0.1751178801059723, "learning_rate": 0.00017639705930138272, "loss": 1.2491, "step": 131 }, { "epoch": 0.016781216479281713, "grad_norm": 0.17463481426239014, "learning_rate": 0.00017602177139989044, "loss": 1.3015, "step": 132 }, { "epoch": 0.01690834690715506, "grad_norm": 0.1884208619594574, "learning_rate": 0.0001756439293981377, "loss": 1.2555, "step": 133 }, { "epoch": 0.017035477335028405, "grad_norm": 0.1824871301651001, "learning_rate": 0.00017526354599046635, "loss": 1.3321, "step": 134 }, { "epoch": 0.017162607762901752, "grad_norm": 0.17852945625782013, "learning_rate": 0.00017488063395660177, "loss": 1.2134, "step": 135 }, { "epoch": 0.0172897381907751, "grad_norm": 0.17903351783752441, "learning_rate": 0.00017449520616122344, "loss": 1.202, "step": 136 }, { "epoch": 0.017416868618648444, "grad_norm": 0.19624289870262146, "learning_rate": 0.00017410727555353282, "loss": 1.2983, "step": 137 }, { "epoch": 0.01754399904652179, "grad_norm": 0.20271572470664978, "learning_rate": 0.00017371685516681825, "loss": 1.331, "step": 138 }, { "epoch": 0.01767112947439514, "grad_norm": 0.19160455465316772, "learning_rate": 0.00017332395811801707, "loss": 1.2325, "step": 139 }, { "epoch": 0.017798259902268482, "grad_norm": 0.19286282360553741, "learning_rate": 0.00017292859760727493, "loss": 1.3632, "step": 140 }, { "epoch": 0.01792539033014183, "grad_norm": 0.18525561690330505, "learning_rate": 0.00017253078691750227, "loss": 1.302, "step": 141 }, { "epoch": 0.018052520758015177, "grad_norm": 0.17999610304832458, "learning_rate": 0.00017213053941392818, "loss": 1.2617, "step": 142 }, { "epoch": 0.01817965118588852, "grad_norm": 0.1817435920238495, "learning_rate": 0.00017172786854365116, "loss": 1.285, "step": 143 }, { "epoch": 0.01830678161376187, "grad_norm": 0.18393941223621368, "learning_rate": 0.00017132278783518756, "loss": 1.2033, "step": 144 }, { "epoch": 0.018433912041635216, "grad_norm": 0.18280182778835297, "learning_rate": 0.00017091531089801694, "loss": 1.2454, "step": 145 }, { "epoch": 0.01856104246950856, "grad_norm": 0.17269238829612732, "learning_rate": 0.00017050545142212483, "loss": 1.2137, "step": 146 }, { "epoch": 0.018688172897381908, "grad_norm": 0.18515561521053314, "learning_rate": 0.00017009322317754278, "loss": 1.2876, "step": 147 }, { "epoch": 0.018815303325255255, "grad_norm": 0.18649280071258545, "learning_rate": 0.0001696786400138859, "loss": 1.3279, "step": 148 }, { "epoch": 0.0189424337531286, "grad_norm": 0.18008284270763397, "learning_rate": 0.00016926171585988727, "loss": 1.1943, "step": 149 }, { "epoch": 0.019069564181001947, "grad_norm": 0.18855896592140198, "learning_rate": 0.00016884246472293016, "loss": 1.3458, "step": 150 }, { "epoch": 0.019196694608875294, "grad_norm": 0.18721222877502441, "learning_rate": 0.00016842090068857742, "loss": 1.205, "step": 151 }, { "epoch": 0.019323825036748638, "grad_norm": 0.18609726428985596, "learning_rate": 0.00016799703792009827, "loss": 1.3147, "step": 152 }, { "epoch": 0.019450955464621986, "grad_norm": 0.18827542662620544, "learning_rate": 0.00016757089065799226, "loss": 1.2053, "step": 153 }, { "epoch": 0.019578085892495333, "grad_norm": 0.19211921095848083, "learning_rate": 0.00016714247321951106, "loss": 1.2881, "step": 154 }, { "epoch": 0.019705216320368677, "grad_norm": 0.1911146342754364, "learning_rate": 0.0001667117999981774, "loss": 1.2841, "step": 155 }, { "epoch": 0.019832346748242025, "grad_norm": 0.1876746416091919, "learning_rate": 0.00016627888546330138, "loss": 1.2795, "step": 156 }, { "epoch": 0.019959477176115372, "grad_norm": 0.18275220692157745, "learning_rate": 0.00016584374415949443, "loss": 1.2646, "step": 157 }, { "epoch": 0.020086607603988716, "grad_norm": 0.19240595400333405, "learning_rate": 0.0001654063907061807, "loss": 1.2286, "step": 158 }, { "epoch": 0.020213738031862064, "grad_norm": 0.17621144652366638, "learning_rate": 0.00016496683979710575, "loss": 1.2623, "step": 159 }, { "epoch": 0.02034086845973541, "grad_norm": 0.18566247820854187, "learning_rate": 0.000164525106199843, "loss": 1.2915, "step": 160 }, { "epoch": 0.020467998887608755, "grad_norm": 0.19843867421150208, "learning_rate": 0.00016408120475529763, "loss": 1.1703, "step": 161 }, { "epoch": 0.020595129315482102, "grad_norm": 0.20230089128017426, "learning_rate": 0.00016363515037720773, "loss": 1.274, "step": 162 }, { "epoch": 0.02072225974335545, "grad_norm": 0.1874382644891739, "learning_rate": 0.00016318695805164359, "loss": 1.267, "step": 163 }, { "epoch": 0.020849390171228794, "grad_norm": 0.19301468133926392, "learning_rate": 0.0001627366428365039, "loss": 1.3385, "step": 164 }, { "epoch": 0.02097652059910214, "grad_norm": 0.1960678994655609, "learning_rate": 0.00016228421986101005, "loss": 1.2469, "step": 165 }, { "epoch": 0.02110365102697549, "grad_norm": 0.2149035483598709, "learning_rate": 0.00016182970432519772, "loss": 1.2695, "step": 166 }, { "epoch": 0.021230781454848833, "grad_norm": 0.1928316354751587, "learning_rate": 0.00016137311149940633, "loss": 1.2581, "step": 167 }, { "epoch": 0.02135791188272218, "grad_norm": 0.18403369188308716, "learning_rate": 0.0001609144567237658, "loss": 1.2872, "step": 168 }, { "epoch": 0.021485042310595528, "grad_norm": 0.18688054382801056, "learning_rate": 0.00016045375540768136, "loss": 1.2762, "step": 169 }, { "epoch": 0.021612172738468872, "grad_norm": 0.19875864684581757, "learning_rate": 0.00015999102302931585, "loss": 1.2773, "step": 170 }, { "epoch": 0.02173930316634222, "grad_norm": 0.19474861025810242, "learning_rate": 0.0001595262751350695, "loss": 1.2329, "step": 171 }, { "epoch": 0.021866433594215567, "grad_norm": 0.1946505606174469, "learning_rate": 0.00015905952733905775, "loss": 1.1726, "step": 172 }, { "epoch": 0.02199356402208891, "grad_norm": 0.18479324877262115, "learning_rate": 0.00015859079532258677, "loss": 1.3177, "step": 173 }, { "epoch": 0.022120694449962258, "grad_norm": 0.19268646836280823, "learning_rate": 0.00015812009483362642, "loss": 1.2721, "step": 174 }, { "epoch": 0.022247824877835606, "grad_norm": 0.18371957540512085, "learning_rate": 0.0001576474416862812, "loss": 1.3083, "step": 175 }, { "epoch": 0.02237495530570895, "grad_norm": 0.1987624615430832, "learning_rate": 0.00015717285176025913, "loss": 1.2582, "step": 176 }, { "epoch": 0.022502085733582297, "grad_norm": 0.19360652565956116, "learning_rate": 0.00015669634100033797, "loss": 1.2597, "step": 177 }, { "epoch": 0.022629216161455645, "grad_norm": 0.1875244826078415, "learning_rate": 0.00015621792541582966, "loss": 1.2637, "step": 178 }, { "epoch": 0.02275634658932899, "grad_norm": 0.19594229757785797, "learning_rate": 0.00015573762108004262, "loss": 1.2907, "step": 179 }, { "epoch": 0.022883477017202336, "grad_norm": 0.1935066133737564, "learning_rate": 0.00015525544412974132, "loss": 1.2446, "step": 180 }, { "epoch": 0.023010607445075684, "grad_norm": 0.19178606569766998, "learning_rate": 0.0001547714107646046, "loss": 1.2644, "step": 181 }, { "epoch": 0.023137737872949028, "grad_norm": 0.18824580311775208, "learning_rate": 0.00015428553724668103, "loss": 1.2592, "step": 182 }, { "epoch": 0.023264868300822375, "grad_norm": 0.1857818067073822, "learning_rate": 0.00015379783989984277, "loss": 1.2547, "step": 183 }, { "epoch": 0.023391998728695722, "grad_norm": 0.18491147458553314, "learning_rate": 0.00015330833510923718, "loss": 1.3073, "step": 184 }, { "epoch": 0.023519129156569066, "grad_norm": 0.19134363532066345, "learning_rate": 0.00015281703932073612, "loss": 1.2456, "step": 185 }, { "epoch": 0.023646259584442414, "grad_norm": 0.18579505383968353, "learning_rate": 0.0001523239690403835, "loss": 1.2626, "step": 186 }, { "epoch": 0.02377339001231576, "grad_norm": 0.18687140941619873, "learning_rate": 0.0001518291408338409, "loss": 1.2795, "step": 187 }, { "epoch": 0.023900520440189105, "grad_norm": 0.1869836449623108, "learning_rate": 0.00015133257132583073, "loss": 1.2111, "step": 188 }, { "epoch": 0.024027650868062453, "grad_norm": 0.18433886766433716, "learning_rate": 0.00015083427719957793, "loss": 1.1969, "step": 189 }, { "epoch": 0.0241547812959358, "grad_norm": 0.19012001156806946, "learning_rate": 0.0001503342751962493, "loss": 1.2973, "step": 190 }, { "epoch": 0.024281911723809144, "grad_norm": 0.18975861370563507, "learning_rate": 0.00014983258211439117, "loss": 1.2964, "step": 191 }, { "epoch": 0.024409042151682492, "grad_norm": 0.17685554921627045, "learning_rate": 0.0001493292148093649, "loss": 1.2763, "step": 192 }, { "epoch": 0.02453617257955584, "grad_norm": 0.19333194196224213, "learning_rate": 0.00014882419019278075, "loss": 1.3203, "step": 193 }, { "epoch": 0.024663303007429183, "grad_norm": 0.19778768718242645, "learning_rate": 0.00014831752523192948, "loss": 1.3204, "step": 194 }, { "epoch": 0.02479043343530253, "grad_norm": 0.1869363635778427, "learning_rate": 0.00014780923694921255, "loss": 1.2258, "step": 195 }, { "epoch": 0.024917563863175878, "grad_norm": 0.17671674489974976, "learning_rate": 0.00014729934242157004, "loss": 1.1667, "step": 196 }, { "epoch": 0.025044694291049222, "grad_norm": 0.1893490105867386, "learning_rate": 0.00014678785877990697, "loss": 1.3572, "step": 197 }, { "epoch": 0.02517182471892257, "grad_norm": 0.19606593251228333, "learning_rate": 0.00014627480320851774, "loss": 1.2507, "step": 198 }, { "epoch": 0.025298955146795917, "grad_norm": 0.20087891817092896, "learning_rate": 0.00014576019294450888, "loss": 1.3149, "step": 199 }, { "epoch": 0.02542608557466926, "grad_norm": 0.1857730895280838, "learning_rate": 0.00014524404527721977, "loss": 1.2893, "step": 200 }, { "epoch": 0.02542608557466926, "eval_loss": 1.2551084756851196, "eval_runtime": 1258.1994, "eval_samples_per_second": 3.974, "eval_steps_per_second": 0.993, "step": 200 }, { "epoch": 0.02555321600254261, "grad_norm": 0.18368631601333618, "learning_rate": 0.00014472637754764196, "loss": 1.2125, "step": 201 }, { "epoch": 0.025680346430415956, "grad_norm": 0.18972043693065643, "learning_rate": 0.00014420720714783636, "loss": 1.2131, "step": 202 }, { "epoch": 0.0258074768582893, "grad_norm": 0.18747109174728394, "learning_rate": 0.00014368655152034908, "loss": 1.2224, "step": 203 }, { "epoch": 0.025934607286162648, "grad_norm": 0.18962696194648743, "learning_rate": 0.00014316442815762544, "loss": 1.2613, "step": 204 }, { "epoch": 0.026061737714035995, "grad_norm": 0.18641987442970276, "learning_rate": 0.00014264085460142202, "loss": 1.2525, "step": 205 }, { "epoch": 0.02618886814190934, "grad_norm": 0.19106072187423706, "learning_rate": 0.0001421158484422177, "loss": 1.2549, "step": 206 }, { "epoch": 0.026315998569782686, "grad_norm": 0.19771872460842133, "learning_rate": 0.0001415894273186223, "loss": 1.2612, "step": 207 }, { "epoch": 0.026443128997656034, "grad_norm": 0.18108506500720978, "learning_rate": 0.0001410616089167842, "loss": 1.2114, "step": 208 }, { "epoch": 0.026570259425529378, "grad_norm": 0.17011211812496185, "learning_rate": 0.0001405324109697961, "loss": 1.2695, "step": 209 }, { "epoch": 0.026697389853402725, "grad_norm": 0.1930396556854248, "learning_rate": 0.00014000185125709918, "loss": 1.211, "step": 210 }, { "epoch": 0.026824520281276073, "grad_norm": 0.19416122138500214, "learning_rate": 0.00013946994760388582, "loss": 1.1772, "step": 211 }, { "epoch": 0.026951650709149417, "grad_norm": 0.18353648483753204, "learning_rate": 0.00013893671788050074, "loss": 1.2672, "step": 212 }, { "epoch": 0.027078781137022764, "grad_norm": 0.18951141834259033, "learning_rate": 0.00013840218000184053, "loss": 1.3209, "step": 213 }, { "epoch": 0.027205911564896112, "grad_norm": 0.19500471651554108, "learning_rate": 0.00013786635192675184, "loss": 1.2519, "step": 214 }, { "epoch": 0.027333041992769456, "grad_norm": 0.1958056539297104, "learning_rate": 0.00013732925165742805, "loss": 1.208, "step": 215 }, { "epoch": 0.027460172420642803, "grad_norm": 0.1859259307384491, "learning_rate": 0.00013679089723880427, "loss": 1.2715, "step": 216 }, { "epoch": 0.02758730284851615, "grad_norm": 0.18673139810562134, "learning_rate": 0.00013625130675795134, "loss": 1.292, "step": 217 }, { "epoch": 0.027714433276389495, "grad_norm": 0.17959700524806976, "learning_rate": 0.00013571049834346799, "loss": 1.1896, "step": 218 }, { "epoch": 0.027841563704262842, "grad_norm": 0.1903347671031952, "learning_rate": 0.0001351684901648718, "loss": 1.3381, "step": 219 }, { "epoch": 0.02796869413213619, "grad_norm": 0.18993370234966278, "learning_rate": 0.00013462530043198873, "loss": 1.2739, "step": 220 }, { "epoch": 0.028095824560009534, "grad_norm": 0.18846477568149567, "learning_rate": 0.0001340809473943415, "loss": 1.2399, "step": 221 }, { "epoch": 0.02822295498788288, "grad_norm": 0.18699532747268677, "learning_rate": 0.00013353544934053616, "loss": 1.2061, "step": 222 }, { "epoch": 0.02835008541575623, "grad_norm": 0.19469809532165527, "learning_rate": 0.00013298882459764798, "loss": 1.2455, "step": 223 }, { "epoch": 0.028477215843629573, "grad_norm": 0.19830243289470673, "learning_rate": 0.00013244109153060548, "loss": 1.2542, "step": 224 }, { "epoch": 0.02860434627150292, "grad_norm": 0.20483078062534332, "learning_rate": 0.0001318922685415735, "loss": 1.2287, "step": 225 }, { "epoch": 0.028731476699376268, "grad_norm": 0.190695121884346, "learning_rate": 0.00013134237406933492, "loss": 1.2165, "step": 226 }, { "epoch": 0.02885860712724961, "grad_norm": 0.19430223107337952, "learning_rate": 0.00013079142658867124, "loss": 1.2922, "step": 227 }, { "epoch": 0.02898573755512296, "grad_norm": 0.1994917094707489, "learning_rate": 0.00013023944460974183, "loss": 1.2402, "step": 228 }, { "epoch": 0.029112867982996306, "grad_norm": 0.20195803046226501, "learning_rate": 0.00012968644667746206, "loss": 1.2594, "step": 229 }, { "epoch": 0.02923999841086965, "grad_norm": 0.19695453345775604, "learning_rate": 0.00012913245137088024, "loss": 1.2762, "step": 230 }, { "epoch": 0.029367128838742998, "grad_norm": 0.19726547598838806, "learning_rate": 0.00012857747730255338, "loss": 1.2494, "step": 231 }, { "epoch": 0.029494259266616345, "grad_norm": 0.19146564602851868, "learning_rate": 0.00012802154311792197, "loss": 1.2312, "step": 232 }, { "epoch": 0.02962138969448969, "grad_norm": 0.19849611818790436, "learning_rate": 0.00012746466749468345, "loss": 1.2186, "step": 233 }, { "epoch": 0.029748520122363037, "grad_norm": 0.18684804439544678, "learning_rate": 0.00012690686914216474, "loss": 1.1775, "step": 234 }, { "epoch": 0.029875650550236384, "grad_norm": 0.19955115020275116, "learning_rate": 0.0001263481668006937, "loss": 1.2476, "step": 235 }, { "epoch": 0.03000278097810973, "grad_norm": 0.20034034550189972, "learning_rate": 0.00012578857924096934, "loss": 1.2307, "step": 236 }, { "epoch": 0.030129911405983076, "grad_norm": 0.1980581283569336, "learning_rate": 0.00012522812526343148, "loss": 1.2332, "step": 237 }, { "epoch": 0.030257041833856423, "grad_norm": 0.1966305524110794, "learning_rate": 0.00012466682369762882, "loss": 1.2219, "step": 238 }, { "epoch": 0.030384172261729767, "grad_norm": 0.19543439149856567, "learning_rate": 0.00012410469340158655, "loss": 1.2998, "step": 239 }, { "epoch": 0.030511302689603115, "grad_norm": 0.19517168402671814, "learning_rate": 0.00012354175326117253, "loss": 1.2451, "step": 240 }, { "epoch": 0.030638433117476462, "grad_norm": 0.18800078332424164, "learning_rate": 0.00012297802218946306, "loss": 1.2349, "step": 241 }, { "epoch": 0.030765563545349806, "grad_norm": 0.20408718287944794, "learning_rate": 0.00012241351912610726, "loss": 1.3123, "step": 242 }, { "epoch": 0.030892693973223154, "grad_norm": 0.19570188224315643, "learning_rate": 0.00012184826303669083, "loss": 1.2181, "step": 243 }, { "epoch": 0.0310198244010965, "grad_norm": 0.1854136884212494, "learning_rate": 0.00012128227291209891, "loss": 1.2298, "step": 244 }, { "epoch": 0.031146954828969845, "grad_norm": 0.19532455503940582, "learning_rate": 0.00012071556776787786, "loss": 1.3124, "step": 245 }, { "epoch": 0.03127408525684319, "grad_norm": 0.18832701444625854, "learning_rate": 0.00012014816664359671, "loss": 1.1565, "step": 246 }, { "epoch": 0.03140121568471654, "grad_norm": 0.19303160905838013, "learning_rate": 0.0001195800886022071, "loss": 1.2329, "step": 247 }, { "epoch": 0.03152834611258989, "grad_norm": 0.1881171315908432, "learning_rate": 0.0001190113527294032, "loss": 1.2456, "step": 248 }, { "epoch": 0.03165547654046323, "grad_norm": 0.19656208157539368, "learning_rate": 0.00011844197813298017, "loss": 1.2959, "step": 249 }, { "epoch": 0.031782606968336576, "grad_norm": 0.19458794593811035, "learning_rate": 0.0001178719839421925, "loss": 1.2968, "step": 250 }, { "epoch": 0.031909737396209926, "grad_norm": 0.1953679323196411, "learning_rate": 0.00011730138930711101, "loss": 1.3225, "step": 251 }, { "epoch": 0.03203686782408327, "grad_norm": 0.18624471127986908, "learning_rate": 0.00011673021339797967, "loss": 1.2895, "step": 252 }, { "epoch": 0.032163998251956614, "grad_norm": 0.1975700408220291, "learning_rate": 0.00011615847540457157, "loss": 1.2272, "step": 253 }, { "epoch": 0.032291128679829965, "grad_norm": 0.18464815616607666, "learning_rate": 0.000115586194535544, "loss": 1.1589, "step": 254 }, { "epoch": 0.03241825910770331, "grad_norm": 0.19085770845413208, "learning_rate": 0.00011501339001779332, "loss": 1.2129, "step": 255 }, { "epoch": 0.03254538953557665, "grad_norm": 0.19415773451328278, "learning_rate": 0.00011444008109580884, "loss": 1.2209, "step": 256 }, { "epoch": 0.032672519963450004, "grad_norm": 0.20239484310150146, "learning_rate": 0.00011386628703102633, "loss": 1.2872, "step": 257 }, { "epoch": 0.03279965039132335, "grad_norm": 0.18557807803153992, "learning_rate": 0.00011329202710118088, "loss": 1.2661, "step": 258 }, { "epoch": 0.03292678081919669, "grad_norm": 0.19118450582027435, "learning_rate": 0.00011271732059965925, "loss": 1.2781, "step": 259 }, { "epoch": 0.03305391124707004, "grad_norm": 0.1958242654800415, "learning_rate": 0.00011214218683485158, "loss": 1.2579, "step": 260 }, { "epoch": 0.03318104167494339, "grad_norm": 0.1829763948917389, "learning_rate": 0.00011156664512950287, "loss": 1.2359, "step": 261 }, { "epoch": 0.03330817210281673, "grad_norm": 0.18603093922138214, "learning_rate": 0.00011099071482006361, "loss": 1.2487, "step": 262 }, { "epoch": 0.03343530253069008, "grad_norm": 0.18507151305675507, "learning_rate": 0.00011041441525604014, "loss": 1.2339, "step": 263 }, { "epoch": 0.033562432958563426, "grad_norm": 0.198081374168396, "learning_rate": 0.00010983776579934482, "loss": 1.1937, "step": 264 }, { "epoch": 0.03368956338643677, "grad_norm": 0.19567249715328217, "learning_rate": 0.00010926078582364514, "loss": 1.2447, "step": 265 }, { "epoch": 0.03381669381431012, "grad_norm": 0.1892256885766983, "learning_rate": 0.00010868349471371315, "loss": 1.2011, "step": 266 }, { "epoch": 0.033943824242183465, "grad_norm": 0.19196678698062897, "learning_rate": 0.000108105911864774, "loss": 1.2341, "step": 267 }, { "epoch": 0.03407095467005681, "grad_norm": 0.20349054038524628, "learning_rate": 0.00010752805668185442, "loss": 1.2582, "step": 268 }, { "epoch": 0.03419808509793016, "grad_norm": 0.19051875174045563, "learning_rate": 0.0001069499485791307, "loss": 1.2593, "step": 269 }, { "epoch": 0.034325215525803504, "grad_norm": 0.18554829061031342, "learning_rate": 0.00010637160697927651, "loss": 1.1395, "step": 270 }, { "epoch": 0.03445234595367685, "grad_norm": 0.18687109649181366, "learning_rate": 0.00010579305131281025, "loss": 1.2079, "step": 271 }, { "epoch": 0.0345794763815502, "grad_norm": 0.19098562002182007, "learning_rate": 0.00010521430101744239, "loss": 1.2147, "step": 272 }, { "epoch": 0.03470660680942354, "grad_norm": 0.19619682431221008, "learning_rate": 0.00010463537553742225, "loss": 1.1458, "step": 273 }, { "epoch": 0.03483373723729689, "grad_norm": 0.1930547058582306, "learning_rate": 0.00010405629432288488, "loss": 1.2704, "step": 274 }, { "epoch": 0.03496086766517024, "grad_norm": 0.19838818907737732, "learning_rate": 0.00010347707682919754, "loss": 1.2228, "step": 275 }, { "epoch": 0.03508799809304358, "grad_norm": 0.1933777928352356, "learning_rate": 0.00010289774251630602, "loss": 1.189, "step": 276 }, { "epoch": 0.035215128520916926, "grad_norm": 0.20109447836875916, "learning_rate": 0.0001023183108480809, "loss": 1.2199, "step": 277 }, { "epoch": 0.03534225894879028, "grad_norm": 0.18909600377082825, "learning_rate": 0.00010173880129166358, "loss": 1.1529, "step": 278 }, { "epoch": 0.03546938937666362, "grad_norm": 0.18809406459331512, "learning_rate": 0.00010115923331681232, "loss": 1.2183, "step": 279 }, { "epoch": 0.035596519804536965, "grad_norm": 0.191794291138649, "learning_rate": 0.00010057962639524798, "loss": 1.2621, "step": 280 }, { "epoch": 0.035723650232410316, "grad_norm": 0.19512364268302917, "learning_rate": 0.0001, "loss": 1.1575, "step": 281 }, { "epoch": 0.03585078066028366, "grad_norm": 0.18720309436321259, "learning_rate": 9.942037360475205e-05, "loss": 1.2435, "step": 282 }, { "epoch": 0.035977911088157004, "grad_norm": 0.20254142582416534, "learning_rate": 9.884076668318773e-05, "loss": 1.3042, "step": 283 }, { "epoch": 0.036105041516030355, "grad_norm": 0.1859116405248642, "learning_rate": 9.826119870833643e-05, "loss": 1.2607, "step": 284 }, { "epoch": 0.0362321719439037, "grad_norm": 0.20392954349517822, "learning_rate": 9.768168915191913e-05, "loss": 1.2478, "step": 285 }, { "epoch": 0.03635930237177704, "grad_norm": 0.20078176259994507, "learning_rate": 9.710225748369401e-05, "loss": 1.2359, "step": 286 }, { "epoch": 0.036486432799650394, "grad_norm": 0.1939856857061386, "learning_rate": 9.65229231708025e-05, "loss": 1.2484, "step": 287 }, { "epoch": 0.03661356322752374, "grad_norm": 0.19322216510772705, "learning_rate": 9.594370567711513e-05, "loss": 1.3249, "step": 288 }, { "epoch": 0.03674069365539708, "grad_norm": 0.1958608329296112, "learning_rate": 9.536462446257776e-05, "loss": 1.2212, "step": 289 }, { "epoch": 0.03686782408327043, "grad_norm": 0.2111774981021881, "learning_rate": 9.478569898255765e-05, "loss": 1.3251, "step": 290 }, { "epoch": 0.03699495451114378, "grad_norm": 0.20403634011745453, "learning_rate": 9.420694868718977e-05, "loss": 1.1831, "step": 291 }, { "epoch": 0.03712208493901712, "grad_norm": 0.20421482622623444, "learning_rate": 9.362839302072354e-05, "loss": 1.2513, "step": 292 }, { "epoch": 0.03724921536689047, "grad_norm": 0.1911601573228836, "learning_rate": 9.305005142086932e-05, "loss": 1.1875, "step": 293 }, { "epoch": 0.037376345794763816, "grad_norm": 0.20054025948047638, "learning_rate": 9.247194331814562e-05, "loss": 1.214, "step": 294 }, { "epoch": 0.03750347622263716, "grad_norm": 0.198801651597023, "learning_rate": 9.1894088135226e-05, "loss": 1.1993, "step": 295 }, { "epoch": 0.03763060665051051, "grad_norm": 0.1995508074760437, "learning_rate": 9.131650528628687e-05, "loss": 1.1986, "step": 296 }, { "epoch": 0.037757737078383855, "grad_norm": 0.20325009524822235, "learning_rate": 9.073921417635486e-05, "loss": 1.2292, "step": 297 }, { "epoch": 0.0378848675062572, "grad_norm": 0.193388894200325, "learning_rate": 9.016223420065519e-05, "loss": 1.2304, "step": 298 }, { "epoch": 0.03801199793413055, "grad_norm": 0.1861080676317215, "learning_rate": 8.958558474395987e-05, "loss": 1.2843, "step": 299 }, { "epoch": 0.03813912836200389, "grad_norm": 0.19273944199085236, "learning_rate": 8.900928517993644e-05, "loss": 1.111, "step": 300 }, { "epoch": 0.03813912836200389, "eval_loss": 1.2362135648727417, "eval_runtime": 1257.3039, "eval_samples_per_second": 3.977, "eval_steps_per_second": 0.994, "step": 300 }, { "epoch": 0.03826625878987724, "grad_norm": 0.19571375846862793, "learning_rate": 8.843335487049712e-05, "loss": 1.2259, "step": 301 }, { "epoch": 0.03839338921775059, "grad_norm": 0.19907177984714508, "learning_rate": 8.785781316514841e-05, "loss": 1.1946, "step": 302 }, { "epoch": 0.03852051964562393, "grad_norm": 0.19650404155254364, "learning_rate": 8.728267940034078e-05, "loss": 1.1977, "step": 303 }, { "epoch": 0.038647650073497276, "grad_norm": 0.1937037855386734, "learning_rate": 8.670797289881915e-05, "loss": 1.1719, "step": 304 }, { "epoch": 0.03877478050137063, "grad_norm": 0.2028638869524002, "learning_rate": 8.61337129689737e-05, "loss": 1.224, "step": 305 }, { "epoch": 0.03890191092924397, "grad_norm": 0.19181466102600098, "learning_rate": 8.555991890419117e-05, "loss": 1.2375, "step": 306 }, { "epoch": 0.039029041357117315, "grad_norm": 0.1887066513299942, "learning_rate": 8.498660998220669e-05, "loss": 1.1786, "step": 307 }, { "epoch": 0.039156171784990666, "grad_norm": 0.18820159137248993, "learning_rate": 8.441380546445603e-05, "loss": 1.2536, "step": 308 }, { "epoch": 0.03928330221286401, "grad_norm": 0.18667809665203094, "learning_rate": 8.384152459542848e-05, "loss": 1.2834, "step": 309 }, { "epoch": 0.039410432640737354, "grad_norm": 0.19369632005691528, "learning_rate": 8.326978660202034e-05, "loss": 1.2989, "step": 310 }, { "epoch": 0.039537563068610705, "grad_norm": 0.18864746391773224, "learning_rate": 8.269861069288903e-05, "loss": 1.292, "step": 311 }, { "epoch": 0.03966469349648405, "grad_norm": 0.19906029105186462, "learning_rate": 8.212801605780753e-05, "loss": 1.2855, "step": 312 }, { "epoch": 0.03979182392435739, "grad_norm": 0.19700987637043, "learning_rate": 8.155802186701984e-05, "loss": 1.1771, "step": 313 }, { "epoch": 0.039918954352230744, "grad_norm": 0.20134317874908447, "learning_rate": 8.098864727059685e-05, "loss": 1.1995, "step": 314 }, { "epoch": 0.04004608478010409, "grad_norm": 0.1883343607187271, "learning_rate": 8.04199113977929e-05, "loss": 1.2433, "step": 315 }, { "epoch": 0.04017321520797743, "grad_norm": 0.19041708111763, "learning_rate": 7.985183335640331e-05, "loss": 1.2538, "step": 316 }, { "epoch": 0.04030034563585078, "grad_norm": 0.1838679164648056, "learning_rate": 7.928443223212215e-05, "loss": 1.2025, "step": 317 }, { "epoch": 0.04042747606372413, "grad_norm": 0.19493237137794495, "learning_rate": 7.871772708790114e-05, "loss": 1.2553, "step": 318 }, { "epoch": 0.04055460649159747, "grad_norm": 0.197859525680542, "learning_rate": 7.815173696330919e-05, "loss": 1.2661, "step": 319 }, { "epoch": 0.04068173691947082, "grad_norm": 0.19427183270454407, "learning_rate": 7.758648087389277e-05, "loss": 1.2121, "step": 320 }, { "epoch": 0.040808867347344166, "grad_norm": 0.19236573576927185, "learning_rate": 7.702197781053696e-05, "loss": 1.2375, "step": 321 }, { "epoch": 0.04093599777521751, "grad_norm": 0.19856838881969452, "learning_rate": 7.645824673882748e-05, "loss": 1.2648, "step": 322 }, { "epoch": 0.04106312820309086, "grad_norm": 0.20721471309661865, "learning_rate": 7.589530659841349e-05, "loss": 1.2503, "step": 323 }, { "epoch": 0.041190258630964205, "grad_norm": 0.19413287937641144, "learning_rate": 7.533317630237117e-05, "loss": 1.265, "step": 324 }, { "epoch": 0.04131738905883755, "grad_norm": 0.1948065459728241, "learning_rate": 7.477187473656853e-05, "loss": 1.2581, "step": 325 }, { "epoch": 0.0414445194867109, "grad_norm": 0.19630448520183563, "learning_rate": 7.421142075903067e-05, "loss": 1.2013, "step": 326 }, { "epoch": 0.041571649914584244, "grad_norm": 0.18867121636867523, "learning_rate": 7.365183319930635e-05, "loss": 1.1628, "step": 327 }, { "epoch": 0.04169878034245759, "grad_norm": 0.2017098367214203, "learning_rate": 7.309313085783524e-05, "loss": 1.1882, "step": 328 }, { "epoch": 0.04182591077033094, "grad_norm": 0.19574840366840363, "learning_rate": 7.253533250531656e-05, "loss": 1.1917, "step": 329 }, { "epoch": 0.04195304119820428, "grad_norm": 0.2003111094236374, "learning_rate": 7.197845688207805e-05, "loss": 1.3069, "step": 330 }, { "epoch": 0.04208017162607763, "grad_norm": 0.19444699585437775, "learning_rate": 7.142252269744665e-05, "loss": 1.1623, "step": 331 }, { "epoch": 0.04220730205395098, "grad_norm": 0.19306592643260956, "learning_rate": 7.086754862911982e-05, "loss": 1.2512, "step": 332 }, { "epoch": 0.04233443248182432, "grad_norm": 0.1940678060054779, "learning_rate": 7.031355332253795e-05, "loss": 1.2404, "step": 333 }, { "epoch": 0.042461562909697666, "grad_norm": 0.20104342699050903, "learning_rate": 6.976055539025818e-05, "loss": 1.1826, "step": 334 }, { "epoch": 0.04258869333757102, "grad_norm": 0.20509012043476105, "learning_rate": 6.92085734113288e-05, "loss": 1.2247, "step": 335 }, { "epoch": 0.04271582376544436, "grad_norm": 0.2038545161485672, "learning_rate": 6.865762593066513e-05, "loss": 1.25, "step": 336 }, { "epoch": 0.042842954193317705, "grad_norm": 0.20035366714000702, "learning_rate": 6.810773145842653e-05, "loss": 1.2243, "step": 337 }, { "epoch": 0.042970084621191056, "grad_norm": 0.20092110335826874, "learning_rate": 6.755890846939454e-05, "loss": 1.2279, "step": 338 }, { "epoch": 0.0430972150490644, "grad_norm": 0.20261281728744507, "learning_rate": 6.701117540235204e-05, "loss": 1.2418, "step": 339 }, { "epoch": 0.043224345476937744, "grad_norm": 0.20916980504989624, "learning_rate": 6.646455065946386e-05, "loss": 1.2205, "step": 340 }, { "epoch": 0.043351475904811095, "grad_norm": 0.1868792027235031, "learning_rate": 6.591905260565852e-05, "loss": 1.2149, "step": 341 }, { "epoch": 0.04347860633268444, "grad_norm": 0.19856908917427063, "learning_rate": 6.537469956801128e-05, "loss": 1.2518, "step": 342 }, { "epoch": 0.04360573676055778, "grad_norm": 0.19585344195365906, "learning_rate": 6.483150983512823e-05, "loss": 1.2202, "step": 343 }, { "epoch": 0.04373286718843113, "grad_norm": 0.19705970585346222, "learning_rate": 6.428950165653204e-05, "loss": 1.2701, "step": 344 }, { "epoch": 0.04385999761630448, "grad_norm": 0.19830965995788574, "learning_rate": 6.374869324204869e-05, "loss": 1.2132, "step": 345 }, { "epoch": 0.04398712804417782, "grad_norm": 0.20360921323299408, "learning_rate": 6.320910276119576e-05, "loss": 1.1979, "step": 346 }, { "epoch": 0.04411425847205117, "grad_norm": 0.20261693000793457, "learning_rate": 6.267074834257199e-05, "loss": 1.2231, "step": 347 }, { "epoch": 0.044241388899924516, "grad_norm": 0.19419489800930023, "learning_rate": 6.213364807324818e-05, "loss": 1.1575, "step": 348 }, { "epoch": 0.04436851932779786, "grad_norm": 0.2025313675403595, "learning_rate": 6.15978199981595e-05, "loss": 1.2566, "step": 349 }, { "epoch": 0.04449564975567121, "grad_norm": 0.19754880666732788, "learning_rate": 6.106328211949928e-05, "loss": 1.2821, "step": 350 }, { "epoch": 0.044622780183544555, "grad_norm": 0.20343464612960815, "learning_rate": 6.053005239611418e-05, "loss": 1.2204, "step": 351 }, { "epoch": 0.0447499106114179, "grad_norm": 0.20527192950248718, "learning_rate": 5.999814874290084e-05, "loss": 1.2513, "step": 352 }, { "epoch": 0.04487704103929125, "grad_norm": 0.19321362674236298, "learning_rate": 5.946758903020393e-05, "loss": 1.2466, "step": 353 }, { "epoch": 0.045004171467164594, "grad_norm": 0.20470896363258362, "learning_rate": 5.893839108321584e-05, "loss": 1.2846, "step": 354 }, { "epoch": 0.04513130189503794, "grad_norm": 0.19084323942661285, "learning_rate": 5.841057268137771e-05, "loss": 1.2126, "step": 355 }, { "epoch": 0.04525843232291129, "grad_norm": 0.19587008655071259, "learning_rate": 5.7884151557782305e-05, "loss": 1.1983, "step": 356 }, { "epoch": 0.04538556275078463, "grad_norm": 0.20390859246253967, "learning_rate": 5.735914539857798e-05, "loss": 1.1981, "step": 357 }, { "epoch": 0.04551269317865798, "grad_norm": 0.19584935903549194, "learning_rate": 5.68355718423746e-05, "loss": 1.2039, "step": 358 }, { "epoch": 0.04563982360653133, "grad_norm": 0.19530071318149567, "learning_rate": 5.6313448479650946e-05, "loss": 1.236, "step": 359 }, { "epoch": 0.04576695403440467, "grad_norm": 0.19659969210624695, "learning_rate": 5.579279285216369e-05, "loss": 1.1936, "step": 360 }, { "epoch": 0.045894084462278016, "grad_norm": 0.1933298110961914, "learning_rate": 5.527362245235805e-05, "loss": 1.227, "step": 361 }, { "epoch": 0.04602121489015137, "grad_norm": 0.20280398428440094, "learning_rate": 5.475595472278024e-05, "loss": 1.2644, "step": 362 }, { "epoch": 0.04614834531802471, "grad_norm": 0.1918189376592636, "learning_rate": 5.4239807055491135e-05, "loss": 1.1495, "step": 363 }, { "epoch": 0.046275475745898055, "grad_norm": 0.2044762223958969, "learning_rate": 5.372519679148227e-05, "loss": 1.241, "step": 364 }, { "epoch": 0.046402606173771406, "grad_norm": 0.1972542256116867, "learning_rate": 5.321214122009306e-05, "loss": 1.1419, "step": 365 }, { "epoch": 0.04652973660164475, "grad_norm": 0.20039339363574982, "learning_rate": 5.270065757843e-05, "loss": 1.2718, "step": 366 }, { "epoch": 0.046656867029518094, "grad_norm": 0.1938110589981079, "learning_rate": 5.219076305078749e-05, "loss": 1.1947, "step": 367 }, { "epoch": 0.046783997457391445, "grad_norm": 0.20640990138053894, "learning_rate": 5.168247476807053e-05, "loss": 1.1526, "step": 368 }, { "epoch": 0.04691112788526479, "grad_norm": 0.198054239153862, "learning_rate": 5.11758098072193e-05, "loss": 1.1965, "step": 369 }, { "epoch": 0.04703825831313813, "grad_norm": 0.19604484736919403, "learning_rate": 5.067078519063514e-05, "loss": 1.2568, "step": 370 }, { "epoch": 0.047165388741011484, "grad_norm": 0.2095029056072235, "learning_rate": 5.016741788560889e-05, "loss": 1.2822, "step": 371 }, { "epoch": 0.04729251916888483, "grad_norm": 0.20356985926628113, "learning_rate": 4.9665724803750756e-05, "loss": 1.1434, "step": 372 }, { "epoch": 0.04741964959675817, "grad_norm": 0.19989654421806335, "learning_rate": 4.9165722800422096e-05, "loss": 1.2767, "step": 373 }, { "epoch": 0.04754678002463152, "grad_norm": 0.19582509994506836, "learning_rate": 4.86674286741693e-05, "loss": 1.2693, "step": 374 }, { "epoch": 0.04767391045250487, "grad_norm": 0.1962389498949051, "learning_rate": 4.8170859166159144e-05, "loss": 1.3266, "step": 375 }, { "epoch": 0.04780104088037821, "grad_norm": 0.2056453377008438, "learning_rate": 4.7676030959616526e-05, "loss": 1.3004, "step": 376 }, { "epoch": 0.04792817130825156, "grad_norm": 0.19587452709674835, "learning_rate": 4.71829606792639e-05, "loss": 1.2154, "step": 377 }, { "epoch": 0.048055301736124906, "grad_norm": 0.19662117958068848, "learning_rate": 4.669166489076283e-05, "loss": 1.2434, "step": 378 }, { "epoch": 0.04818243216399825, "grad_norm": 0.19508899748325348, "learning_rate": 4.620216010015724e-05, "loss": 1.2319, "step": 379 }, { "epoch": 0.0483095625918716, "grad_norm": 0.19653861224651337, "learning_rate": 4.571446275331903e-05, "loss": 1.2006, "step": 380 }, { "epoch": 0.048436693019744945, "grad_norm": 0.1971856951713562, "learning_rate": 4.5228589235395436e-05, "loss": 1.2937, "step": 381 }, { "epoch": 0.04856382344761829, "grad_norm": 0.2165059596300125, "learning_rate": 4.4744555870258694e-05, "loss": 1.2722, "step": 382 }, { "epoch": 0.04869095387549164, "grad_norm": 0.20020557940006256, "learning_rate": 4.4262378919957413e-05, "loss": 1.1947, "step": 383 }, { "epoch": 0.048818084303364984, "grad_norm": 0.19455446302890778, "learning_rate": 4.378207458417035e-05, "loss": 1.1956, "step": 384 }, { "epoch": 0.04894521473123833, "grad_norm": 0.202660471200943, "learning_rate": 4.3303658999662086e-05, "loss": 1.2553, "step": 385 }, { "epoch": 0.04907234515911168, "grad_norm": 0.19681531190872192, "learning_rate": 4.282714823974088e-05, "loss": 1.2031, "step": 386 }, { "epoch": 0.04919947558698502, "grad_norm": 0.20613734424114227, "learning_rate": 4.2352558313718795e-05, "loss": 1.2384, "step": 387 }, { "epoch": 0.04932660601485837, "grad_norm": 0.1990024596452713, "learning_rate": 4.1879905166373614e-05, "loss": 1.2184, "step": 388 }, { "epoch": 0.04945373644273172, "grad_norm": 0.21309691667556763, "learning_rate": 4.140920467741325e-05, "loss": 1.1853, "step": 389 }, { "epoch": 0.04958086687060506, "grad_norm": 0.19488035142421722, "learning_rate": 4.094047266094225e-05, "loss": 1.1804, "step": 390 }, { "epoch": 0.049707997298478405, "grad_norm": 0.19738849997520447, "learning_rate": 4.047372486493054e-05, "loss": 1.2534, "step": 391 }, { "epoch": 0.049835127726351756, "grad_norm": 0.20008018612861633, "learning_rate": 4.0008976970684176e-05, "loss": 1.2723, "step": 392 }, { "epoch": 0.0499622581542251, "grad_norm": 0.19521461427211761, "learning_rate": 3.954624459231866e-05, "loss": 1.1705, "step": 393 }, { "epoch": 0.050089388582098444, "grad_norm": 0.20466111600399017, "learning_rate": 3.908554327623425e-05, "loss": 1.154, "step": 394 }, { "epoch": 0.050216519009971795, "grad_norm": 0.2047969251871109, "learning_rate": 3.8626888500593695e-05, "loss": 1.2139, "step": 395 }, { "epoch": 0.05034364943784514, "grad_norm": 0.1980600655078888, "learning_rate": 3.817029567480228e-05, "loss": 1.279, "step": 396 }, { "epoch": 0.05047077986571848, "grad_norm": 0.20217813551425934, "learning_rate": 3.771578013898996e-05, "loss": 1.2561, "step": 397 }, { "epoch": 0.050597910293591834, "grad_norm": 0.1985122561454773, "learning_rate": 3.726335716349612e-05, "loss": 1.2778, "step": 398 }, { "epoch": 0.05072504072146518, "grad_norm": 0.19889195263385773, "learning_rate": 3.681304194835641e-05, "loss": 1.3225, "step": 399 }, { "epoch": 0.05085217114933852, "grad_norm": 0.19213935732841492, "learning_rate": 3.6364849622792266e-05, "loss": 1.2308, "step": 400 }, { "epoch": 0.05085217114933852, "eval_loss": 1.2253398895263672, "eval_runtime": 1257.7786, "eval_samples_per_second": 3.975, "eval_steps_per_second": 0.994, "step": 400 }, { "epoch": 0.05097930157721187, "grad_norm": 0.19073964655399323, "learning_rate": 3.5918795244702396e-05, "loss": 1.1798, "step": 401 }, { "epoch": 0.05110643200508522, "grad_norm": 0.19348806142807007, "learning_rate": 3.547489380015701e-05, "loss": 1.2429, "step": 402 }, { "epoch": 0.05123356243295856, "grad_norm": 0.201893150806427, "learning_rate": 3.503316020289429e-05, "loss": 1.2302, "step": 403 }, { "epoch": 0.05136069286083191, "grad_norm": 0.20248207449913025, "learning_rate": 3.459360929381931e-05, "loss": 1.2295, "step": 404 }, { "epoch": 0.051487823288705256, "grad_norm": 0.20291946828365326, "learning_rate": 3.415625584050557e-05, "loss": 1.2925, "step": 405 }, { "epoch": 0.0516149537165786, "grad_norm": 0.19170844554901123, "learning_rate": 3.372111453669864e-05, "loss": 1.1825, "step": 406 }, { "epoch": 0.05174208414445195, "grad_norm": 0.1890149712562561, "learning_rate": 3.328820000182262e-05, "loss": 1.149, "step": 407 }, { "epoch": 0.051869214572325295, "grad_norm": 0.20486074686050415, "learning_rate": 3.285752678048892e-05, "loss": 1.1458, "step": 408 }, { "epoch": 0.05199634500019864, "grad_norm": 0.20006342232227325, "learning_rate": 3.242910934200775e-05, "loss": 1.2031, "step": 409 }, { "epoch": 0.05212347542807199, "grad_norm": 0.20113137364387512, "learning_rate": 3.2002962079901744e-05, "loss": 1.2474, "step": 410 }, { "epoch": 0.052250605855945334, "grad_norm": 0.1928662657737732, "learning_rate": 3.157909931142257e-05, "loss": 1.2189, "step": 411 }, { "epoch": 0.05237773628381868, "grad_norm": 0.19416049122810364, "learning_rate": 3.115753527706986e-05, "loss": 1.2506, "step": 412 }, { "epoch": 0.05250486671169203, "grad_norm": 0.19863539934158325, "learning_rate": 3.073828414011274e-05, "loss": 1.2019, "step": 413 }, { "epoch": 0.05263199713956537, "grad_norm": 0.19736243784427643, "learning_rate": 3.0321359986114096e-05, "loss": 1.2718, "step": 414 }, { "epoch": 0.05275912756743872, "grad_norm": 0.200786292552948, "learning_rate": 2.9906776822457205e-05, "loss": 1.2523, "step": 415 }, { "epoch": 0.05288625799531207, "grad_norm": 0.19629159569740295, "learning_rate": 2.9494548577875192e-05, "loss": 1.2156, "step": 416 }, { "epoch": 0.05301338842318541, "grad_norm": 0.19293853640556335, "learning_rate": 2.9084689101983075e-05, "loss": 1.2422, "step": 417 }, { "epoch": 0.053140518851058756, "grad_norm": 0.19834856688976288, "learning_rate": 2.8677212164812462e-05, "loss": 1.1569, "step": 418 }, { "epoch": 0.05326764927893211, "grad_norm": 0.19939422607421875, "learning_rate": 2.827213145634887e-05, "loss": 1.18, "step": 419 }, { "epoch": 0.05339477970680545, "grad_norm": 0.20803573727607727, "learning_rate": 2.7869460586071873e-05, "loss": 1.2536, "step": 420 }, { "epoch": 0.053521910134678795, "grad_norm": 0.20102088153362274, "learning_rate": 2.7469213082497736e-05, "loss": 1.2664, "step": 421 }, { "epoch": 0.053649040562552146, "grad_norm": 0.20080140233039856, "learning_rate": 2.7071402392725097e-05, "loss": 1.2931, "step": 422 }, { "epoch": 0.05377617099042549, "grad_norm": 0.19506552815437317, "learning_rate": 2.6676041881982962e-05, "loss": 1.2461, "step": 423 }, { "epoch": 0.053903301418298834, "grad_norm": 0.19534236192703247, "learning_rate": 2.6283144833181783e-05, "loss": 1.1971, "step": 424 }, { "epoch": 0.054030431846172185, "grad_norm": 0.20050783455371857, "learning_rate": 2.589272444646723e-05, "loss": 1.169, "step": 425 }, { "epoch": 0.05415756227404553, "grad_norm": 0.201995849609375, "learning_rate": 2.5504793838776586e-05, "loss": 1.2758, "step": 426 }, { "epoch": 0.05428469270191887, "grad_norm": 0.20075969398021698, "learning_rate": 2.5119366043398264e-05, "loss": 1.2455, "step": 427 }, { "epoch": 0.054411823129792224, "grad_norm": 0.19450737535953522, "learning_rate": 2.473645400953366e-05, "loss": 1.2319, "step": 428 }, { "epoch": 0.05453895355766557, "grad_norm": 0.19657433032989502, "learning_rate": 2.4356070601862324e-05, "loss": 1.2031, "step": 429 }, { "epoch": 0.05466608398553891, "grad_norm": 0.1956198513507843, "learning_rate": 2.3978228600109565e-05, "loss": 1.2345, "step": 430 }, { "epoch": 0.05479321441341226, "grad_norm": 0.19880710542201996, "learning_rate": 2.3602940698617325e-05, "loss": 1.3102, "step": 431 }, { "epoch": 0.05492034484128561, "grad_norm": 0.19965411722660065, "learning_rate": 2.3230219505917426e-05, "loss": 1.2873, "step": 432 }, { "epoch": 0.05504747526915895, "grad_norm": 0.19630952179431915, "learning_rate": 2.2860077544308124e-05, "loss": 1.272, "step": 433 }, { "epoch": 0.0551746056970323, "grad_norm": 0.19560639560222626, "learning_rate": 2.249252724943336e-05, "loss": 1.2593, "step": 434 }, { "epoch": 0.055301736124905645, "grad_norm": 0.20019298791885376, "learning_rate": 2.2127580969864925e-05, "loss": 1.2723, "step": 435 }, { "epoch": 0.05542886655277899, "grad_norm": 0.20072917640209198, "learning_rate": 2.176525096668769e-05, "loss": 1.2069, "step": 436 }, { "epoch": 0.05555599698065234, "grad_norm": 0.20204520225524902, "learning_rate": 2.1405549413087544e-05, "loss": 1.2361, "step": 437 }, { "epoch": 0.055683127408525684, "grad_norm": 0.20441173017024994, "learning_rate": 2.1048488393942454e-05, "loss": 1.1882, "step": 438 }, { "epoch": 0.05581025783639903, "grad_norm": 0.1971142739057541, "learning_rate": 2.0694079905416473e-05, "loss": 1.2168, "step": 439 }, { "epoch": 0.05593738826427238, "grad_norm": 0.20565344393253326, "learning_rate": 2.0342335854556737e-05, "loss": 1.2486, "step": 440 }, { "epoch": 0.05606451869214572, "grad_norm": 0.19731251895427704, "learning_rate": 1.9993268058893343e-05, "loss": 1.116, "step": 441 }, { "epoch": 0.05619164912001907, "grad_norm": 0.19447311758995056, "learning_rate": 1.964688824604234e-05, "loss": 1.2497, "step": 442 }, { "epoch": 0.05631877954789242, "grad_norm": 0.2041405290365219, "learning_rate": 1.930320805331176e-05, "loss": 1.2843, "step": 443 }, { "epoch": 0.05644590997576576, "grad_norm": 0.20058204233646393, "learning_rate": 1.896223902731058e-05, "loss": 1.2685, "step": 444 }, { "epoch": 0.056573040403639106, "grad_norm": 0.20194143056869507, "learning_rate": 1.8623992623560893e-05, "loss": 1.2751, "step": 445 }, { "epoch": 0.05670017083151246, "grad_norm": 0.20336030423641205, "learning_rate": 1.828848020611288e-05, "loss": 1.2362, "step": 446 }, { "epoch": 0.0568273012593858, "grad_norm": 0.2054579108953476, "learning_rate": 1.7955713047163157e-05, "loss": 1.2254, "step": 447 }, { "epoch": 0.056954431687259145, "grad_norm": 0.20549102127552032, "learning_rate": 1.762570232667595e-05, "loss": 1.2574, "step": 448 }, { "epoch": 0.057081562115132496, "grad_norm": 0.2029975950717926, "learning_rate": 1.7298459132007627e-05, "loss": 1.2066, "step": 449 }, { "epoch": 0.05720869254300584, "grad_norm": 0.20551042258739471, "learning_rate": 1.6973994457534026e-05, "loss": 1.2384, "step": 450 }, { "epoch": 0.057335822970879184, "grad_norm": 0.20290708541870117, "learning_rate": 1.6652319204281187e-05, "loss": 1.2257, "step": 451 }, { "epoch": 0.057462953398752535, "grad_norm": 0.202660471200943, "learning_rate": 1.6333444179559078e-05, "loss": 1.1865, "step": 452 }, { "epoch": 0.05759008382662588, "grad_norm": 0.19935418665409088, "learning_rate": 1.601738009659849e-05, "loss": 1.2445, "step": 453 }, { "epoch": 0.05771721425449922, "grad_norm": 0.19397136569023132, "learning_rate": 1.5704137574191203e-05, "loss": 1.2246, "step": 454 }, { "epoch": 0.057844344682372574, "grad_norm": 0.20033963024616241, "learning_rate": 1.5393727136333035e-05, "loss": 1.2452, "step": 455 }, { "epoch": 0.05797147511024592, "grad_norm": 0.19725894927978516, "learning_rate": 1.5086159211870442e-05, "loss": 1.1902, "step": 456 }, { "epoch": 0.05809860553811926, "grad_norm": 0.21174763143062592, "learning_rate": 1.4781444134150047e-05, "loss": 1.1956, "step": 457 }, { "epoch": 0.05822573596599261, "grad_norm": 0.20643867552280426, "learning_rate": 1.447959214067155e-05, "loss": 1.2708, "step": 458 }, { "epoch": 0.05835286639386596, "grad_norm": 0.203439399600029, "learning_rate": 1.4180613372743679e-05, "loss": 1.1935, "step": 459 }, { "epoch": 0.0584799968217393, "grad_norm": 0.2026398777961731, "learning_rate": 1.3884517875143544e-05, "loss": 1.2331, "step": 460 }, { "epoch": 0.05860712724961265, "grad_norm": 0.20543427765369415, "learning_rate": 1.3591315595779108e-05, "loss": 1.2039, "step": 461 }, { "epoch": 0.058734257677485996, "grad_norm": 0.2049439400434494, "learning_rate": 1.3301016385355092e-05, "loss": 1.1847, "step": 462 }, { "epoch": 0.05886138810535934, "grad_norm": 0.20000450313091278, "learning_rate": 1.3013629997041853e-05, "loss": 1.2432, "step": 463 }, { "epoch": 0.05898851853323269, "grad_norm": 0.20123903453350067, "learning_rate": 1.2729166086147803e-05, "loss": 1.2698, "step": 464 }, { "epoch": 0.059115648961106035, "grad_norm": 0.20424886047840118, "learning_rate": 1.2447634209795e-05, "loss": 1.2635, "step": 465 }, { "epoch": 0.05924277938897938, "grad_norm": 0.19165809452533722, "learning_rate": 1.2169043826598058e-05, "loss": 1.1772, "step": 466 }, { "epoch": 0.05936990981685273, "grad_norm": 0.19973435997962952, "learning_rate": 1.1893404296346423e-05, "loss": 1.2326, "step": 467 }, { "epoch": 0.059497040244726074, "grad_norm": 0.20723304152488708, "learning_rate": 1.1620724879689792e-05, "loss": 1.2451, "step": 468 }, { "epoch": 0.05962417067259942, "grad_norm": 0.1965390294790268, "learning_rate": 1.135101473782706e-05, "loss": 1.2154, "step": 469 }, { "epoch": 0.05975130110047277, "grad_norm": 0.19724096357822418, "learning_rate": 1.1084282932198541e-05, "loss": 1.2092, "step": 470 }, { "epoch": 0.05987843152834611, "grad_norm": 0.19244475662708282, "learning_rate": 1.0820538424181515e-05, "loss": 1.12, "step": 471 }, { "epoch": 0.06000556195621946, "grad_norm": 0.20514468848705292, "learning_rate": 1.0559790074789133e-05, "loss": 1.1499, "step": 472 }, { "epoch": 0.06013269238409281, "grad_norm": 0.19560429453849792, "learning_rate": 1.030204664437271e-05, "loss": 1.2266, "step": 473 }, { "epoch": 0.06025982281196615, "grad_norm": 0.20317231118679047, "learning_rate": 1.0047316792327499e-05, "loss": 1.1913, "step": 474 }, { "epoch": 0.060386953239839496, "grad_norm": 0.19489817321300507, "learning_rate": 9.795609076801625e-06, "loss": 1.2461, "step": 475 }, { "epoch": 0.06051408366771285, "grad_norm": 0.1974310576915741, "learning_rate": 9.546931954408622e-06, "loss": 1.1443, "step": 476 }, { "epoch": 0.06064121409558619, "grad_norm": 0.19223681092262268, "learning_rate": 9.301293779943321e-06, "loss": 1.1864, "step": 477 }, { "epoch": 0.060768344523459535, "grad_norm": 0.2029593586921692, "learning_rate": 9.058702806101172e-06, "loss": 1.2081, "step": 478 }, { "epoch": 0.060895474951332886, "grad_norm": 0.20288948714733124, "learning_rate": 8.819167183200905e-06, "loss": 1.2794, "step": 479 }, { "epoch": 0.06102260537920623, "grad_norm": 0.20905464887619019, "learning_rate": 8.58269495891081e-06, "loss": 1.2125, "step": 480 }, { "epoch": 0.061149735807079574, "grad_norm": 0.2005111575126648, "learning_rate": 8.349294077978265e-06, "loss": 1.2183, "step": 481 }, { "epoch": 0.061276866234952924, "grad_norm": 0.20153559744358063, "learning_rate": 8.118972381962853e-06, "loss": 1.2213, "step": 482 }, { "epoch": 0.06140399666282627, "grad_norm": 0.20067910850048065, "learning_rate": 7.891737608972927e-06, "loss": 1.1919, "step": 483 }, { "epoch": 0.06153112709069961, "grad_norm": 0.2007380872964859, "learning_rate": 7.6675973934056e-06, "loss": 1.2182, "step": 484 }, { "epoch": 0.06165825751857296, "grad_norm": 0.19200630486011505, "learning_rate": 7.4465592656903114e-06, "loss": 1.2089, "step": 485 }, { "epoch": 0.06178538794644631, "grad_norm": 0.19756072759628296, "learning_rate": 7.228630652035717e-06, "loss": 1.2236, "step": 486 }, { "epoch": 0.06191251837431965, "grad_norm": 0.20330612361431122, "learning_rate": 7.0138188741803225e-06, "loss": 1.1603, "step": 487 }, { "epoch": 0.062039648802193, "grad_norm": 0.2018778920173645, "learning_rate": 6.802131149146373e-06, "loss": 1.2256, "step": 488 }, { "epoch": 0.062166779230066346, "grad_norm": 0.19810867309570312, "learning_rate": 6.59357458899752e-06, "loss": 1.2123, "step": 489 }, { "epoch": 0.06229390965793969, "grad_norm": 0.19678053259849548, "learning_rate": 6.388156200599726e-06, "loss": 1.2266, "step": 490 }, { "epoch": 0.06242104008581304, "grad_norm": 0.20404191315174103, "learning_rate": 6.185882885385952e-06, "loss": 1.2376, "step": 491 }, { "epoch": 0.06254817051368639, "grad_norm": 0.20588378608226776, "learning_rate": 5.986761439124289e-06, "loss": 1.2604, "step": 492 }, { "epoch": 0.06267530094155974, "grad_norm": 0.19841307401657104, "learning_rate": 5.790798551689592e-06, "loss": 1.2849, "step": 493 }, { "epoch": 0.06280243136943307, "grad_norm": 0.20967555046081543, "learning_rate": 5.598000806838766e-06, "loss": 1.2321, "step": 494 }, { "epoch": 0.06292956179730642, "grad_norm": 0.19931279122829437, "learning_rate": 5.408374681989548e-06, "loss": 1.278, "step": 495 }, { "epoch": 0.06305669222517978, "grad_norm": 0.19611623883247375, "learning_rate": 5.221926548002876e-06, "loss": 1.1999, "step": 496 }, { "epoch": 0.06318382265305311, "grad_norm": 0.19863781332969666, "learning_rate": 5.038662668968886e-06, "loss": 1.2295, "step": 497 }, { "epoch": 0.06331095308092646, "grad_norm": 0.19629532098770142, "learning_rate": 4.858589201996433e-06, "loss": 1.2194, "step": 498 }, { "epoch": 0.06343808350879981, "grad_norm": 0.21103760600090027, "learning_rate": 4.681712197006205e-06, "loss": 1.2861, "step": 499 }, { "epoch": 0.06356521393667315, "grad_norm": 0.20111264288425446, "learning_rate": 4.508037596527526e-06, "loss": 1.2123, "step": 500 }, { "epoch": 0.06356521393667315, "eval_loss": 1.22141695022583, "eval_runtime": 1257.3017, "eval_samples_per_second": 3.977, "eval_steps_per_second": 0.994, "step": 500 }, { "epoch": 0.0636923443645465, "grad_norm": 0.20111538469791412, "learning_rate": 4.337571235498628e-06, "loss": 1.2477, "step": 501 }, { "epoch": 0.06381947479241985, "grad_norm": 0.19941484928131104, "learning_rate": 4.170318841070708e-06, "loss": 1.2335, "step": 502 }, { "epoch": 0.06394660522029319, "grad_norm": 0.19625377655029297, "learning_rate": 4.00628603241544e-06, "loss": 1.2302, "step": 503 }, { "epoch": 0.06407373564816654, "grad_norm": 0.2074848711490631, "learning_rate": 3.845478320536178e-06, "loss": 1.22, "step": 504 }, { "epoch": 0.06420086607603989, "grad_norm": 0.19643086194992065, "learning_rate": 3.687901108082892e-06, "loss": 1.2234, "step": 505 }, { "epoch": 0.06432799650391323, "grad_norm": 0.19488604366779327, "learning_rate": 3.53355968917054e-06, "loss": 1.2088, "step": 506 }, { "epoch": 0.06445512693178658, "grad_norm": 0.20913287997245789, "learning_rate": 3.3824592492013085e-06, "loss": 1.1734, "step": 507 }, { "epoch": 0.06458225735965993, "grad_norm": 0.19445528090000153, "learning_rate": 3.2346048646903494e-06, "loss": 1.2014, "step": 508 }, { "epoch": 0.06470938778753327, "grad_norm": 0.19821353256702423, "learning_rate": 3.0900015030951744e-06, "loss": 1.2446, "step": 509 }, { "epoch": 0.06483651821540662, "grad_norm": 0.20283670723438263, "learning_rate": 2.9486540226488557e-06, "loss": 1.1858, "step": 510 }, { "epoch": 0.06496364864327997, "grad_norm": 0.20396962761878967, "learning_rate": 2.8105671721967875e-06, "loss": 1.2166, "step": 511 }, { "epoch": 0.0650907790711533, "grad_norm": 0.2022467404603958, "learning_rate": 2.6757455910370488e-06, "loss": 1.1738, "step": 512 }, { "epoch": 0.06521790949902666, "grad_norm": 0.1929662674665451, "learning_rate": 2.5441938087646612e-06, "loss": 1.205, "step": 513 }, { "epoch": 0.06534503992690001, "grad_norm": 0.19997857511043549, "learning_rate": 2.4159162451193097e-06, "loss": 1.2103, "step": 514 }, { "epoch": 0.06547217035477335, "grad_norm": 0.1969158947467804, "learning_rate": 2.290917209836918e-06, "loss": 1.1912, "step": 515 }, { "epoch": 0.0655993007826467, "grad_norm": 0.19717784225940704, "learning_rate": 2.1692009025048422e-06, "loss": 1.2374, "step": 516 }, { "epoch": 0.06572643121052005, "grad_norm": 0.20149968564510345, "learning_rate": 2.0507714124207157e-06, "loss": 1.1857, "step": 517 }, { "epoch": 0.06585356163839338, "grad_norm": 0.19946229457855225, "learning_rate": 1.9356327184551714e-06, "loss": 1.1947, "step": 518 }, { "epoch": 0.06598069206626674, "grad_norm": 0.19773566722869873, "learning_rate": 1.8237886889180489e-06, "loss": 1.2825, "step": 519 }, { "epoch": 0.06610782249414009, "grad_norm": 0.20150107145309448, "learning_rate": 1.7152430814285303e-06, "loss": 1.2352, "step": 520 }, { "epoch": 0.06623495292201342, "grad_norm": 0.20660698413848877, "learning_rate": 1.6099995427888315e-06, "loss": 1.2127, "step": 521 }, { "epoch": 0.06636208334988677, "grad_norm": 0.19696985185146332, "learning_rate": 1.5080616088616884e-06, "loss": 1.2197, "step": 522 }, { "epoch": 0.06648921377776013, "grad_norm": 0.20150014758110046, "learning_rate": 1.4094327044515853e-06, "loss": 1.2534, "step": 523 }, { "epoch": 0.06661634420563346, "grad_norm": 0.1950562745332718, "learning_rate": 1.3141161431896808e-06, "loss": 1.165, "step": 524 }, { "epoch": 0.06674347463350681, "grad_norm": 0.20011726021766663, "learning_rate": 1.222115127422485e-06, "loss": 1.2179, "step": 525 }, { "epoch": 0.06687060506138016, "grad_norm": 0.20083405077457428, "learning_rate": 1.1334327481042573e-06, "loss": 1.305, "step": 526 }, { "epoch": 0.0669977354892535, "grad_norm": 0.20200292766094208, "learning_rate": 1.0480719846931774e-06, "loss": 1.2263, "step": 527 }, { "epoch": 0.06712486591712685, "grad_norm": 0.20572660863399506, "learning_rate": 9.660357050512158e-07, "loss": 1.2029, "step": 528 }, { "epoch": 0.0672519963450002, "grad_norm": 0.20432178676128387, "learning_rate": 8.873266653478208e-07, "loss": 1.2703, "step": 529 }, { "epoch": 0.06737912677287354, "grad_norm": 0.1951807290315628, "learning_rate": 8.119475099673036e-07, "loss": 1.2298, "step": 530 }, { "epoch": 0.06750625720074689, "grad_norm": 0.2002389281988144, "learning_rate": 7.399007714199658e-07, "loss": 1.2342, "step": 531 }, { "epoch": 0.06763338762862024, "grad_norm": 0.19941021502017975, "learning_rate": 6.711888702570556e-07, "loss": 1.146, "step": 532 }, { "epoch": 0.06776051805649358, "grad_norm": 0.19345982372760773, "learning_rate": 6.058141149894336e-07, "loss": 1.1954, "step": 533 }, { "epoch": 0.06788764848436693, "grad_norm": 0.1961802840232849, "learning_rate": 5.437787020100115e-07, "loss": 1.2165, "step": 534 }, { "epoch": 0.06801477891224028, "grad_norm": 0.20169439911842346, "learning_rate": 4.850847155199567e-07, "loss": 1.2445, "step": 535 }, { "epoch": 0.06814190934011362, "grad_norm": 0.19823016226291656, "learning_rate": 4.297341274586475e-07, "loss": 1.2371, "step": 536 }, { "epoch": 0.06826903976798697, "grad_norm": 0.2043391764163971, "learning_rate": 3.777287974374932e-07, "loss": 1.2942, "step": 537 }, { "epoch": 0.06839617019586032, "grad_norm": 0.20149071514606476, "learning_rate": 3.290704726773619e-07, "loss": 1.1842, "step": 538 }, { "epoch": 0.06852330062373366, "grad_norm": 0.20085620880126953, "learning_rate": 2.837607879499604e-07, "loss": 1.1982, "step": 539 }, { "epoch": 0.06865043105160701, "grad_norm": 0.2070370465517044, "learning_rate": 2.418012655228452e-07, "loss": 1.211, "step": 540 }, { "epoch": 0.06877756147948036, "grad_norm": 0.2030269056558609, "learning_rate": 2.0319331510835205e-07, "loss": 1.2534, "step": 541 }, { "epoch": 0.0689046919073537, "grad_norm": 0.1966077983379364, "learning_rate": 1.6793823381614505e-07, "loss": 1.1683, "step": 542 }, { "epoch": 0.06903182233522705, "grad_norm": 0.205659419298172, "learning_rate": 1.3603720610972925e-07, "loss": 1.141, "step": 543 }, { "epoch": 0.0691589527631004, "grad_norm": 0.2047136127948761, "learning_rate": 1.0749130376659366e-07, "loss": 1.2415, "step": 544 }, { "epoch": 0.06928608319097374, "grad_norm": 0.2038879543542862, "learning_rate": 8.230148584219554e-08, "loss": 1.2148, "step": 545 }, { "epoch": 0.06941321361884709, "grad_norm": 0.19984766840934753, "learning_rate": 6.046859863781951e-08, "loss": 1.1954, "step": 546 }, { "epoch": 0.06954034404672044, "grad_norm": 0.2025536447763443, "learning_rate": 4.199337567203365e-08, "loss": 1.2076, "step": 547 }, { "epoch": 0.06966747447459377, "grad_norm": 0.2022514045238495, "learning_rate": 2.6876437656153665e-08, "loss": 1.2369, "step": 548 }, { "epoch": 0.06979460490246713, "grad_norm": 0.21228355169296265, "learning_rate": 1.5118292473292885e-08, "loss": 1.2304, "step": 549 }, { "epoch": 0.06992173533034048, "grad_norm": 0.20524141192436218, "learning_rate": 6.719335161364804e-09, "loss": 1.2158, "step": 550 }, { "epoch": 0.07004886575821381, "grad_norm": 0.18994790315628052, "learning_rate": 1.6798478997825939e-09, "loss": 1.1945, "step": 551 }, { "epoch": 0.07017599618608716, "grad_norm": 0.20824959874153137, "learning_rate": 0.0, "loss": 1.241, "step": 552 } ], "logging_steps": 1, "max_steps": 552, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 2, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.720928125466968e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }