{ "best_metric": 1.1766536235809326, "best_model_checkpoint": "miner_id_24/checkpoint-400", "epoch": 0.45610034207525657, "eval_steps": 100, "global_step": 500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0009122006841505131, "grad_norm": 0.25783950090408325, "learning_rate": 5e-06, "loss": 1.2112, "step": 1 }, { "epoch": 0.0009122006841505131, "eval_loss": 1.528485894203186, "eval_runtime": 113.6842, "eval_samples_per_second": 16.238, "eval_steps_per_second": 4.064, "step": 1 }, { "epoch": 0.0018244013683010262, "grad_norm": 0.3564816117286682, "learning_rate": 1e-05, "loss": 1.0953, "step": 2 }, { "epoch": 0.0027366020524515395, "grad_norm": 0.3273909389972687, "learning_rate": 1.5e-05, "loss": 1.3012, "step": 3 }, { "epoch": 0.0036488027366020524, "grad_norm": 0.2606115937232971, "learning_rate": 2e-05, "loss": 1.2529, "step": 4 }, { "epoch": 0.004561003420752566, "grad_norm": 0.3063911199569702, "learning_rate": 2.5e-05, "loss": 1.4867, "step": 5 }, { "epoch": 0.005473204104903079, "grad_norm": 0.30145400762557983, "learning_rate": 3e-05, "loss": 1.4505, "step": 6 }, { "epoch": 0.0063854047890535915, "grad_norm": 0.3016102612018585, "learning_rate": 3.5e-05, "loss": 1.5596, "step": 7 }, { "epoch": 0.007297605473204105, "grad_norm": 0.291785329580307, "learning_rate": 4e-05, "loss": 2.1331, "step": 8 }, { "epoch": 0.008209806157354617, "grad_norm": 0.38654470443725586, "learning_rate": 4.5e-05, "loss": 1.998, "step": 9 }, { "epoch": 0.009122006841505131, "grad_norm": 0.24303749203681946, "learning_rate": 5e-05, "loss": 1.2621, "step": 10 }, { "epoch": 0.010034207525655644, "grad_norm": 0.4852235019207001, "learning_rate": 5.500000000000001e-05, "loss": 2.9854, "step": 11 }, { "epoch": 0.010946408209806158, "grad_norm": 0.2009701430797577, "learning_rate": 6e-05, "loss": 1.2834, "step": 12 }, { "epoch": 0.01185860889395667, "grad_norm": 0.17731308937072754, "learning_rate": 6.500000000000001e-05, "loss": 1.3094, "step": 13 }, { "epoch": 0.012770809578107183, "grad_norm": 0.1488691121339798, "learning_rate": 7e-05, "loss": 1.0258, "step": 14 }, { "epoch": 0.013683010262257697, "grad_norm": 0.15303461253643036, "learning_rate": 7.500000000000001e-05, "loss": 1.3275, "step": 15 }, { "epoch": 0.01459521094640821, "grad_norm": 0.19198919832706451, "learning_rate": 8e-05, "loss": 1.2239, "step": 16 }, { "epoch": 0.015507411630558724, "grad_norm": 0.18219693005084991, "learning_rate": 8.5e-05, "loss": 1.1367, "step": 17 }, { "epoch": 0.016419612314709234, "grad_norm": 0.15564540028572083, "learning_rate": 9e-05, "loss": 1.1671, "step": 18 }, { "epoch": 0.01733181299885975, "grad_norm": 0.15285353362560272, "learning_rate": 9.5e-05, "loss": 1.2621, "step": 19 }, { "epoch": 0.018244013683010263, "grad_norm": 0.1607624590396881, "learning_rate": 0.0001, "loss": 1.3366, "step": 20 }, { "epoch": 0.019156214367160775, "grad_norm": 0.1453542858362198, "learning_rate": 9.999892908320647e-05, "loss": 1.0341, "step": 21 }, { "epoch": 0.020068415051311288, "grad_norm": 0.2023877501487732, "learning_rate": 9.999571637870036e-05, "loss": 1.3088, "step": 22 }, { "epoch": 0.0209806157354618, "grad_norm": 0.14624053239822388, "learning_rate": 9.999036202410325e-05, "loss": 0.9596, "step": 23 }, { "epoch": 0.021892816419612316, "grad_norm": 0.18767771124839783, "learning_rate": 9.998286624877786e-05, "loss": 1.2299, "step": 24 }, { "epoch": 0.02280501710376283, "grad_norm": 0.20108367502689362, "learning_rate": 9.997322937381829e-05, "loss": 1.2757, "step": 25 }, { "epoch": 0.02371721778791334, "grad_norm": 0.2668151259422302, "learning_rate": 9.996145181203615e-05, "loss": 1.2839, "step": 26 }, { "epoch": 0.024629418472063853, "grad_norm": 0.21400977671146393, "learning_rate": 9.994753406794301e-05, "loss": 1.4506, "step": 27 }, { "epoch": 0.025541619156214366, "grad_norm": 0.26276835799217224, "learning_rate": 9.99314767377287e-05, "loss": 1.2257, "step": 28 }, { "epoch": 0.026453819840364882, "grad_norm": 0.23972387611865997, "learning_rate": 9.991328050923581e-05, "loss": 1.3055, "step": 29 }, { "epoch": 0.027366020524515394, "grad_norm": 0.19858954846858978, "learning_rate": 9.989294616193017e-05, "loss": 1.0788, "step": 30 }, { "epoch": 0.028278221208665907, "grad_norm": 0.21251460909843445, "learning_rate": 9.98704745668676e-05, "loss": 1.0723, "step": 31 }, { "epoch": 0.02919042189281642, "grad_norm": 0.25747787952423096, "learning_rate": 9.98458666866564e-05, "loss": 1.2754, "step": 32 }, { "epoch": 0.03010262257696693, "grad_norm": 0.18308940529823303, "learning_rate": 9.981912357541627e-05, "loss": 1.1309, "step": 33 }, { "epoch": 0.031014823261117448, "grad_norm": 0.28769150376319885, "learning_rate": 9.97902463787331e-05, "loss": 1.1491, "step": 34 }, { "epoch": 0.03192702394526796, "grad_norm": 0.33303603529930115, "learning_rate": 9.975923633360985e-05, "loss": 1.5505, "step": 35 }, { "epoch": 0.03283922462941847, "grad_norm": 0.19029738008975983, "learning_rate": 9.972609476841367e-05, "loss": 1.3271, "step": 36 }, { "epoch": 0.033751425313568985, "grad_norm": 0.1386023312807083, "learning_rate": 9.969082310281891e-05, "loss": 1.2999, "step": 37 }, { "epoch": 0.0346636259977195, "grad_norm": 0.13166525959968567, "learning_rate": 9.965342284774632e-05, "loss": 1.2839, "step": 38 }, { "epoch": 0.03557582668187001, "grad_norm": 0.1366724967956543, "learning_rate": 9.961389560529836e-05, "loss": 1.327, "step": 39 }, { "epoch": 0.036488027366020526, "grad_norm": 0.15251226723194122, "learning_rate": 9.957224306869053e-05, "loss": 1.312, "step": 40 }, { "epoch": 0.037400228050171035, "grad_norm": 0.13887174427509308, "learning_rate": 9.952846702217886e-05, "loss": 1.3777, "step": 41 }, { "epoch": 0.03831242873432155, "grad_norm": 0.1658075451850891, "learning_rate": 9.948256934098352e-05, "loss": 1.3959, "step": 42 }, { "epoch": 0.039224629418472066, "grad_norm": 0.16137467324733734, "learning_rate": 9.943455199120837e-05, "loss": 1.3763, "step": 43 }, { "epoch": 0.040136830102622575, "grad_norm": 0.17031268775463104, "learning_rate": 9.938441702975689e-05, "loss": 1.3507, "step": 44 }, { "epoch": 0.04104903078677309, "grad_norm": 0.17008689045906067, "learning_rate": 9.933216660424395e-05, "loss": 1.3843, "step": 45 }, { "epoch": 0.0419612314709236, "grad_norm": 0.23986858129501343, "learning_rate": 9.927780295290389e-05, "loss": 1.3637, "step": 46 }, { "epoch": 0.042873432155074116, "grad_norm": 0.253842830657959, "learning_rate": 9.922132840449459e-05, "loss": 1.4677, "step": 47 }, { "epoch": 0.04378563283922463, "grad_norm": 0.3024607002735138, "learning_rate": 9.916274537819775e-05, "loss": 1.595, "step": 48 }, { "epoch": 0.04469783352337514, "grad_norm": 0.4269244372844696, "learning_rate": 9.91020563835152e-05, "loss": 1.6949, "step": 49 }, { "epoch": 0.04561003420752566, "grad_norm": 1.273447871208191, "learning_rate": 9.903926402016153e-05, "loss": 2.6063, "step": 50 }, { "epoch": 0.046522234891676166, "grad_norm": 0.25115707516670227, "learning_rate": 9.897437097795257e-05, "loss": 1.2458, "step": 51 }, { "epoch": 0.04743443557582668, "grad_norm": 0.2734687328338623, "learning_rate": 9.890738003669029e-05, "loss": 1.1315, "step": 52 }, { "epoch": 0.0483466362599772, "grad_norm": 0.3112577199935913, "learning_rate": 9.883829406604363e-05, "loss": 1.2837, "step": 53 }, { "epoch": 0.04925883694412771, "grad_norm": 0.2956832945346832, "learning_rate": 9.876711602542563e-05, "loss": 1.6056, "step": 54 }, { "epoch": 0.05017103762827822, "grad_norm": 0.27159854769706726, "learning_rate": 9.869384896386668e-05, "loss": 1.4008, "step": 55 }, { "epoch": 0.05108323831242873, "grad_norm": 0.23057343065738678, "learning_rate": 9.861849601988383e-05, "loss": 1.3612, "step": 56 }, { "epoch": 0.05199543899657925, "grad_norm": 0.22093012928962708, "learning_rate": 9.854106042134641e-05, "loss": 1.3945, "step": 57 }, { "epoch": 0.052907639680729764, "grad_norm": 0.24365180730819702, "learning_rate": 9.846154548533773e-05, "loss": 1.4811, "step": 58 }, { "epoch": 0.05381984036488027, "grad_norm": 0.255241721868515, "learning_rate": 9.837995461801299e-05, "loss": 1.3864, "step": 59 }, { "epoch": 0.05473204104903079, "grad_norm": 0.2505890429019928, "learning_rate": 9.829629131445342e-05, "loss": 1.3, "step": 60 }, { "epoch": 0.0556442417331813, "grad_norm": 0.19075927138328552, "learning_rate": 9.821055915851647e-05, "loss": 1.1616, "step": 61 }, { "epoch": 0.05655644241733181, "grad_norm": 0.19343258440494537, "learning_rate": 9.812276182268236e-05, "loss": 1.0507, "step": 62 }, { "epoch": 0.05746864310148233, "grad_norm": 0.14569373428821564, "learning_rate": 9.803290306789676e-05, "loss": 1.1448, "step": 63 }, { "epoch": 0.05838084378563284, "grad_norm": 0.1645849347114563, "learning_rate": 9.794098674340965e-05, "loss": 1.2202, "step": 64 }, { "epoch": 0.059293044469783354, "grad_norm": 0.1657133549451828, "learning_rate": 9.784701678661045e-05, "loss": 0.9998, "step": 65 }, { "epoch": 0.06020524515393386, "grad_norm": 0.14773251116275787, "learning_rate": 9.775099722285935e-05, "loss": 1.0601, "step": 66 }, { "epoch": 0.06111744583808438, "grad_norm": 0.14379927515983582, "learning_rate": 9.765293216531486e-05, "loss": 1.232, "step": 67 }, { "epoch": 0.062029646522234895, "grad_norm": 0.14879746735095978, "learning_rate": 9.755282581475769e-05, "loss": 1.011, "step": 68 }, { "epoch": 0.0629418472063854, "grad_norm": 0.14008712768554688, "learning_rate": 9.74506824594107e-05, "loss": 0.9511, "step": 69 }, { "epoch": 0.06385404789053592, "grad_norm": 0.15255555510520935, "learning_rate": 9.73465064747553e-05, "loss": 0.983, "step": 70 }, { "epoch": 0.06476624857468644, "grad_norm": 0.16574011743068695, "learning_rate": 9.724030232334391e-05, "loss": 1.0218, "step": 71 }, { "epoch": 0.06567844925883694, "grad_norm": 0.16048061847686768, "learning_rate": 9.713207455460894e-05, "loss": 1.1199, "step": 72 }, { "epoch": 0.06659064994298745, "grad_norm": 0.18178284168243408, "learning_rate": 9.702182780466775e-05, "loss": 1.0187, "step": 73 }, { "epoch": 0.06750285062713797, "grad_norm": 0.15241661667823792, "learning_rate": 9.690956679612421e-05, "loss": 1.2068, "step": 74 }, { "epoch": 0.06841505131128849, "grad_norm": 0.1547076255083084, "learning_rate": 9.67952963378663e-05, "loss": 0.9773, "step": 75 }, { "epoch": 0.069327251995439, "grad_norm": 0.1607867181301117, "learning_rate": 9.667902132486009e-05, "loss": 1.1561, "step": 76 }, { "epoch": 0.0702394526795895, "grad_norm": 0.17180593311786652, "learning_rate": 9.656074673794018e-05, "loss": 1.0215, "step": 77 }, { "epoch": 0.07115165336374002, "grad_norm": 0.18671779334545135, "learning_rate": 9.644047764359622e-05, "loss": 1.0732, "step": 78 }, { "epoch": 0.07206385404789054, "grad_norm": 0.22109557688236237, "learning_rate": 9.631821919375591e-05, "loss": 1.2983, "step": 79 }, { "epoch": 0.07297605473204105, "grad_norm": 0.16936635971069336, "learning_rate": 9.619397662556435e-05, "loss": 1.0403, "step": 80 }, { "epoch": 0.07388825541619157, "grad_norm": 0.20985794067382812, "learning_rate": 9.606775526115963e-05, "loss": 1.1105, "step": 81 }, { "epoch": 0.07480045610034207, "grad_norm": 0.24890471994876862, "learning_rate": 9.593956050744492e-05, "loss": 1.0215, "step": 82 }, { "epoch": 0.07571265678449259, "grad_norm": 0.204713374376297, "learning_rate": 9.580939785585681e-05, "loss": 1.1271, "step": 83 }, { "epoch": 0.0766248574686431, "grad_norm": 0.2502953112125397, "learning_rate": 9.567727288213005e-05, "loss": 1.2075, "step": 84 }, { "epoch": 0.07753705815279362, "grad_norm": 0.2343108057975769, "learning_rate": 9.554319124605879e-05, "loss": 1.2663, "step": 85 }, { "epoch": 0.07844925883694413, "grad_norm": 0.15296907722949982, "learning_rate": 9.540715869125407e-05, "loss": 1.2773, "step": 86 }, { "epoch": 0.07936145952109463, "grad_norm": 0.13890579342842102, "learning_rate": 9.526918104489777e-05, "loss": 1.0958, "step": 87 }, { "epoch": 0.08027366020524515, "grad_norm": 0.14783431589603424, "learning_rate": 9.512926421749304e-05, "loss": 1.1379, "step": 88 }, { "epoch": 0.08118586088939567, "grad_norm": 0.1457366943359375, "learning_rate": 9.498741420261108e-05, "loss": 1.1979, "step": 89 }, { "epoch": 0.08209806157354618, "grad_norm": 0.14700822532176971, "learning_rate": 9.484363707663442e-05, "loss": 1.2511, "step": 90 }, { "epoch": 0.0830102622576967, "grad_norm": 0.14076313376426697, "learning_rate": 9.469793899849661e-05, "loss": 1.273, "step": 91 }, { "epoch": 0.0839224629418472, "grad_norm": 0.1587403565645218, "learning_rate": 9.45503262094184e-05, "loss": 1.2475, "step": 92 }, { "epoch": 0.08483466362599772, "grad_norm": 0.1829884648323059, "learning_rate": 9.440080503264037e-05, "loss": 1.3541, "step": 93 }, { "epoch": 0.08574686431014823, "grad_norm": 0.15041080117225647, "learning_rate": 9.42493818731521e-05, "loss": 1.3175, "step": 94 }, { "epoch": 0.08665906499429875, "grad_norm": 0.16183224320411682, "learning_rate": 9.409606321741775e-05, "loss": 1.3011, "step": 95 }, { "epoch": 0.08757126567844926, "grad_norm": 0.17238490283489227, "learning_rate": 9.394085563309827e-05, "loss": 1.441, "step": 96 }, { "epoch": 0.08848346636259977, "grad_norm": 0.1964457482099533, "learning_rate": 9.378376576876999e-05, "loss": 1.3421, "step": 97 }, { "epoch": 0.08939566704675028, "grad_norm": 0.2645294964313507, "learning_rate": 9.362480035363986e-05, "loss": 1.5238, "step": 98 }, { "epoch": 0.0903078677309008, "grad_norm": 0.4722297191619873, "learning_rate": 9.34639661972572e-05, "loss": 1.6759, "step": 99 }, { "epoch": 0.09122006841505131, "grad_norm": 1.4135944843292236, "learning_rate": 9.330127018922194e-05, "loss": 2.271, "step": 100 }, { "epoch": 0.09122006841505131, "eval_loss": 1.2722834348678589, "eval_runtime": 113.025, "eval_samples_per_second": 16.333, "eval_steps_per_second": 4.088, "step": 100 }, { "epoch": 0.09213226909920183, "grad_norm": 0.12369579821825027, "learning_rate": 9.31367192988896e-05, "loss": 0.996, "step": 101 }, { "epoch": 0.09304446978335233, "grad_norm": 0.12814058363437653, "learning_rate": 9.297032057507264e-05, "loss": 1.0837, "step": 102 }, { "epoch": 0.09395667046750285, "grad_norm": 0.14010389149188995, "learning_rate": 9.280208114573859e-05, "loss": 1.2772, "step": 103 }, { "epoch": 0.09486887115165336, "grad_norm": 0.15505646169185638, "learning_rate": 9.263200821770461e-05, "loss": 1.3608, "step": 104 }, { "epoch": 0.09578107183580388, "grad_norm": 0.16221708059310913, "learning_rate": 9.246010907632895e-05, "loss": 1.3107, "step": 105 }, { "epoch": 0.0966932725199544, "grad_norm": 0.15704494714736938, "learning_rate": 9.228639108519868e-05, "loss": 1.1564, "step": 106 }, { "epoch": 0.0976054732041049, "grad_norm": 0.15318067371845245, "learning_rate": 9.211086168581433e-05, "loss": 1.2841, "step": 107 }, { "epoch": 0.09851767388825541, "grad_norm": 0.16971643269062042, "learning_rate": 9.193352839727121e-05, "loss": 1.3776, "step": 108 }, { "epoch": 0.09942987457240593, "grad_norm": 0.17601683735847473, "learning_rate": 9.175439881593716e-05, "loss": 1.1342, "step": 109 }, { "epoch": 0.10034207525655645, "grad_norm": 0.2042451649904251, "learning_rate": 9.157348061512727e-05, "loss": 1.3205, "step": 110 }, { "epoch": 0.10125427594070696, "grad_norm": 0.20541740953922272, "learning_rate": 9.139078154477512e-05, "loss": 1.2256, "step": 111 }, { "epoch": 0.10216647662485746, "grad_norm": 0.1825324296951294, "learning_rate": 9.120630943110077e-05, "loss": 1.1943, "step": 112 }, { "epoch": 0.10307867730900798, "grad_norm": 0.17503520846366882, "learning_rate": 9.102007217627568e-05, "loss": 1.0749, "step": 113 }, { "epoch": 0.1039908779931585, "grad_norm": 0.1890123039484024, "learning_rate": 9.083207775808396e-05, "loss": 1.0601, "step": 114 }, { "epoch": 0.10490307867730901, "grad_norm": 0.15623880922794342, "learning_rate": 9.064233422958077e-05, "loss": 0.8245, "step": 115 }, { "epoch": 0.10581527936145953, "grad_norm": 0.1406368911266327, "learning_rate": 9.045084971874738e-05, "loss": 1.0316, "step": 116 }, { "epoch": 0.10672748004561003, "grad_norm": 0.1501624435186386, "learning_rate": 9.025763242814291e-05, "loss": 1.1908, "step": 117 }, { "epoch": 0.10763968072976055, "grad_norm": 0.13635428249835968, "learning_rate": 9.006269063455304e-05, "loss": 0.9685, "step": 118 }, { "epoch": 0.10855188141391106, "grad_norm": 0.1702578067779541, "learning_rate": 8.986603268863536e-05, "loss": 1.0043, "step": 119 }, { "epoch": 0.10946408209806158, "grad_norm": 0.16892701387405396, "learning_rate": 8.966766701456177e-05, "loss": 1.0515, "step": 120 }, { "epoch": 0.11037628278221209, "grad_norm": 0.1571364402770996, "learning_rate": 8.94676021096575e-05, "loss": 0.9528, "step": 121 }, { "epoch": 0.1112884834663626, "grad_norm": 0.15004131197929382, "learning_rate": 8.926584654403724e-05, "loss": 1.0623, "step": 122 }, { "epoch": 0.11220068415051311, "grad_norm": 0.18099543452262878, "learning_rate": 8.906240896023794e-05, "loss": 1.0798, "step": 123 }, { "epoch": 0.11311288483466363, "grad_norm": 0.16211232542991638, "learning_rate": 8.885729807284856e-05, "loss": 0.9695, "step": 124 }, { "epoch": 0.11402508551881414, "grad_norm": 0.19721828401088715, "learning_rate": 8.865052266813685e-05, "loss": 1.0282, "step": 125 }, { "epoch": 0.11493728620296466, "grad_norm": 0.16071362793445587, "learning_rate": 8.844209160367299e-05, "loss": 1.0692, "step": 126 }, { "epoch": 0.11584948688711516, "grad_norm": 0.20482495427131653, "learning_rate": 8.823201380795001e-05, "loss": 1.1321, "step": 127 }, { "epoch": 0.11676168757126568, "grad_norm": 0.19026656448841095, "learning_rate": 8.802029828000156e-05, "loss": 0.9361, "step": 128 }, { "epoch": 0.11767388825541619, "grad_norm": 0.19571590423583984, "learning_rate": 8.780695408901613e-05, "loss": 1.0152, "step": 129 }, { "epoch": 0.11858608893956671, "grad_norm": 0.2167755663394928, "learning_rate": 8.759199037394887e-05, "loss": 1.1986, "step": 130 }, { "epoch": 0.11949828962371722, "grad_norm": 0.20209772884845734, "learning_rate": 8.737541634312985e-05, "loss": 0.9061, "step": 131 }, { "epoch": 0.12041049030786773, "grad_norm": 0.2215677797794342, "learning_rate": 8.715724127386972e-05, "loss": 0.8564, "step": 132 }, { "epoch": 0.12132269099201824, "grad_norm": 0.2394784688949585, "learning_rate": 8.693747451206232e-05, "loss": 1.1324, "step": 133 }, { "epoch": 0.12223489167616876, "grad_norm": 0.2727131247520447, "learning_rate": 8.671612547178428e-05, "loss": 1.3143, "step": 134 }, { "epoch": 0.12314709236031927, "grad_norm": 0.263237327337265, "learning_rate": 8.649320363489179e-05, "loss": 1.654, "step": 135 }, { "epoch": 0.12405929304446979, "grad_norm": 0.14213962852954865, "learning_rate": 8.626871855061438e-05, "loss": 1.2624, "step": 136 }, { "epoch": 0.12497149372862029, "grad_norm": 0.11428536474704742, "learning_rate": 8.604267983514594e-05, "loss": 1.1345, "step": 137 }, { "epoch": 0.1258836944127708, "grad_norm": 0.11425229161977768, "learning_rate": 8.581509717123273e-05, "loss": 1.0768, "step": 138 }, { "epoch": 0.12679589509692132, "grad_norm": 0.14232636988162994, "learning_rate": 8.558598030775857e-05, "loss": 1.2241, "step": 139 }, { "epoch": 0.12770809578107184, "grad_norm": 0.14870019257068634, "learning_rate": 8.535533905932738e-05, "loss": 1.2843, "step": 140 }, { "epoch": 0.12862029646522236, "grad_norm": 0.14005614817142487, "learning_rate": 8.51231833058426e-05, "loss": 1.348, "step": 141 }, { "epoch": 0.12953249714937287, "grad_norm": 0.14191389083862305, "learning_rate": 8.488952299208401e-05, "loss": 1.3306, "step": 142 }, { "epoch": 0.1304446978335234, "grad_norm": 0.14700360596179962, "learning_rate": 8.46543681272818e-05, "loss": 1.2998, "step": 143 }, { "epoch": 0.13135689851767388, "grad_norm": 0.17052660882472992, "learning_rate": 8.44177287846877e-05, "loss": 1.3757, "step": 144 }, { "epoch": 0.1322690992018244, "grad_norm": 0.17917169630527496, "learning_rate": 8.417961510114356e-05, "loss": 1.276, "step": 145 }, { "epoch": 0.1331812998859749, "grad_norm": 0.1966487616300583, "learning_rate": 8.39400372766471e-05, "loss": 1.2529, "step": 146 }, { "epoch": 0.13409350057012542, "grad_norm": 0.22469089925289154, "learning_rate": 8.36990055739149e-05, "loss": 1.4108, "step": 147 }, { "epoch": 0.13500570125427594, "grad_norm": 0.29152804613113403, "learning_rate": 8.345653031794292e-05, "loss": 1.3991, "step": 148 }, { "epoch": 0.13591790193842646, "grad_norm": 0.44599342346191406, "learning_rate": 8.321262189556409e-05, "loss": 1.5375, "step": 149 }, { "epoch": 0.13683010262257697, "grad_norm": 0.9355916380882263, "learning_rate": 8.296729075500344e-05, "loss": 2.2142, "step": 150 }, { "epoch": 0.1377423033067275, "grad_norm": 0.16966940462589264, "learning_rate": 8.272054740543052e-05, "loss": 1.0935, "step": 151 }, { "epoch": 0.138654503990878, "grad_norm": 0.17415517568588257, "learning_rate": 8.247240241650918e-05, "loss": 1.069, "step": 152 }, { "epoch": 0.13956670467502852, "grad_norm": 0.15520510077476501, "learning_rate": 8.222286641794488e-05, "loss": 1.1164, "step": 153 }, { "epoch": 0.140478905359179, "grad_norm": 0.16199630498886108, "learning_rate": 8.197195009902924e-05, "loss": 1.2711, "step": 154 }, { "epoch": 0.14139110604332952, "grad_norm": 0.20093193650245667, "learning_rate": 8.171966420818228e-05, "loss": 1.3726, "step": 155 }, { "epoch": 0.14230330672748004, "grad_norm": 0.17293791472911835, "learning_rate": 8.146601955249188e-05, "loss": 1.2781, "step": 156 }, { "epoch": 0.14321550741163055, "grad_norm": 0.20883025228977203, "learning_rate": 8.121102699725089e-05, "loss": 1.3765, "step": 157 }, { "epoch": 0.14412770809578107, "grad_norm": 0.1882334053516388, "learning_rate": 8.095469746549172e-05, "loss": 1.2449, "step": 158 }, { "epoch": 0.1450399087799316, "grad_norm": 0.20094580948352814, "learning_rate": 8.069704193751832e-05, "loss": 1.2535, "step": 159 }, { "epoch": 0.1459521094640821, "grad_norm": 0.2629421055316925, "learning_rate": 8.043807145043604e-05, "loss": 1.3591, "step": 160 }, { "epoch": 0.14686431014823262, "grad_norm": 0.6893475651741028, "learning_rate": 8.017779709767858e-05, "loss": 2.445, "step": 161 }, { "epoch": 0.14777651083238313, "grad_norm": 0.14671513438224792, "learning_rate": 7.991623002853296e-05, "loss": 0.9265, "step": 162 }, { "epoch": 0.14868871151653365, "grad_norm": 0.17360147833824158, "learning_rate": 7.965338144766186e-05, "loss": 0.9184, "step": 163 }, { "epoch": 0.14960091220068414, "grad_norm": 0.16476592421531677, "learning_rate": 7.938926261462366e-05, "loss": 1.0991, "step": 164 }, { "epoch": 0.15051311288483465, "grad_norm": 0.18807539343833923, "learning_rate": 7.912388484339012e-05, "loss": 1.1164, "step": 165 }, { "epoch": 0.15142531356898517, "grad_norm": 0.14939163625240326, "learning_rate": 7.88572595018617e-05, "loss": 1.0408, "step": 166 }, { "epoch": 0.1523375142531357, "grad_norm": 0.16947825253009796, "learning_rate": 7.858939801138061e-05, "loss": 0.937, "step": 167 }, { "epoch": 0.1532497149372862, "grad_norm": 0.13935978710651398, "learning_rate": 7.832031184624164e-05, "loss": 0.9394, "step": 168 }, { "epoch": 0.15416191562143672, "grad_norm": 0.16934405267238617, "learning_rate": 7.80500125332005e-05, "loss": 1.1168, "step": 169 }, { "epoch": 0.15507411630558723, "grad_norm": 0.16050076484680176, "learning_rate": 7.777851165098012e-05, "loss": 0.8851, "step": 170 }, { "epoch": 0.15598631698973775, "grad_norm": 0.13356317579746246, "learning_rate": 7.750582082977467e-05, "loss": 0.8304, "step": 171 }, { "epoch": 0.15689851767388827, "grad_norm": 0.14673428237438202, "learning_rate": 7.723195175075136e-05, "loss": 1.1082, "step": 172 }, { "epoch": 0.15781071835803878, "grad_norm": 0.18469220399856567, "learning_rate": 7.695691614555003e-05, "loss": 1.314, "step": 173 }, { "epoch": 0.15872291904218927, "grad_norm": 0.15937888622283936, "learning_rate": 7.668072579578058e-05, "loss": 0.917, "step": 174 }, { "epoch": 0.15963511972633979, "grad_norm": 0.17641136050224304, "learning_rate": 7.64033925325184e-05, "loss": 1.0633, "step": 175 }, { "epoch": 0.1605473204104903, "grad_norm": 0.13878133893013, "learning_rate": 7.612492823579745e-05, "loss": 0.6571, "step": 176 }, { "epoch": 0.16145952109464082, "grad_norm": 0.19035881757736206, "learning_rate": 7.584534483410137e-05, "loss": 1.0225, "step": 177 }, { "epoch": 0.16237172177879133, "grad_norm": 0.19843564927577972, "learning_rate": 7.55646543038526e-05, "loss": 1.0645, "step": 178 }, { "epoch": 0.16328392246294185, "grad_norm": 0.18366238474845886, "learning_rate": 7.528286866889924e-05, "loss": 0.91, "step": 179 }, { "epoch": 0.16419612314709237, "grad_norm": 0.2137911468744278, "learning_rate": 7.500000000000001e-05, "loss": 0.9201, "step": 180 }, { "epoch": 0.16510832383124288, "grad_norm": 0.19734999537467957, "learning_rate": 7.471606041430723e-05, "loss": 1.1128, "step": 181 }, { "epoch": 0.1660205245153934, "grad_norm": 0.21062205731868744, "learning_rate": 7.443106207484776e-05, "loss": 1.0556, "step": 182 }, { "epoch": 0.1669327251995439, "grad_norm": 0.24208199977874756, "learning_rate": 7.414501719000187e-05, "loss": 1.1084, "step": 183 }, { "epoch": 0.1678449258836944, "grad_norm": 0.3003711998462677, "learning_rate": 7.385793801298042e-05, "loss": 1.2188, "step": 184 }, { "epoch": 0.16875712656784492, "grad_norm": 0.12483101338148117, "learning_rate": 7.35698368412999e-05, "loss": 1.242, "step": 185 }, { "epoch": 0.16966932725199543, "grad_norm": 0.10687603801488876, "learning_rate": 7.328072601625557e-05, "loss": 1.0969, "step": 186 }, { "epoch": 0.17058152793614595, "grad_norm": 0.10687136650085449, "learning_rate": 7.2990617922393e-05, "loss": 1.2887, "step": 187 }, { "epoch": 0.17149372862029647, "grad_norm": 0.12314095348119736, "learning_rate": 7.269952498697734e-05, "loss": 1.2609, "step": 188 }, { "epoch": 0.17240592930444698, "grad_norm": 0.1314985156059265, "learning_rate": 7.240745967946113e-05, "loss": 1.2694, "step": 189 }, { "epoch": 0.1733181299885975, "grad_norm": 0.12387102842330933, "learning_rate": 7.211443451095007e-05, "loss": 1.2707, "step": 190 }, { "epoch": 0.174230330672748, "grad_norm": 0.16482983529567719, "learning_rate": 7.18204620336671e-05, "loss": 1.1994, "step": 191 }, { "epoch": 0.17514253135689853, "grad_norm": 0.16161847114562988, "learning_rate": 7.152555484041476e-05, "loss": 1.3605, "step": 192 }, { "epoch": 0.17605473204104904, "grad_norm": 0.16728828847408295, "learning_rate": 7.122972556403567e-05, "loss": 1.1442, "step": 193 }, { "epoch": 0.17696693272519953, "grad_norm": 0.18906857073307037, "learning_rate": 7.09329868768714e-05, "loss": 1.1403, "step": 194 }, { "epoch": 0.17787913340935005, "grad_norm": 0.21785447001457214, "learning_rate": 7.063535149021973e-05, "loss": 1.4139, "step": 195 }, { "epoch": 0.17879133409350056, "grad_norm": 0.21159727871418, "learning_rate": 7.033683215379002e-05, "loss": 1.4149, "step": 196 }, { "epoch": 0.17970353477765108, "grad_norm": 0.2096451222896576, "learning_rate": 7.003744165515705e-05, "loss": 1.1846, "step": 197 }, { "epoch": 0.1806157354618016, "grad_norm": 0.30022695660591125, "learning_rate": 6.973719281921335e-05, "loss": 1.5039, "step": 198 }, { "epoch": 0.1815279361459521, "grad_norm": 0.39245763421058655, "learning_rate": 6.943609850761979e-05, "loss": 1.4439, "step": 199 }, { "epoch": 0.18244013683010263, "grad_norm": 0.7122502326965332, "learning_rate": 6.91341716182545e-05, "loss": 1.7626, "step": 200 }, { "epoch": 0.18244013683010263, "eval_loss": 1.1971056461334229, "eval_runtime": 113.4472, "eval_samples_per_second": 16.272, "eval_steps_per_second": 4.072, "step": 200 }, { "epoch": 0.18335233751425314, "grad_norm": 0.11290186643600464, "learning_rate": 6.883142508466054e-05, "loss": 0.8787, "step": 201 }, { "epoch": 0.18426453819840366, "grad_norm": 0.12823708355426788, "learning_rate": 6.852787187549182e-05, "loss": 0.8172, "step": 202 }, { "epoch": 0.18517673888255415, "grad_norm": 0.1377316415309906, "learning_rate": 6.82235249939575e-05, "loss": 1.0454, "step": 203 }, { "epoch": 0.18608893956670466, "grad_norm": 0.1476827710866928, "learning_rate": 6.7918397477265e-05, "loss": 1.2703, "step": 204 }, { "epoch": 0.18700114025085518, "grad_norm": 0.14928844571113586, "learning_rate": 6.761250239606169e-05, "loss": 1.3711, "step": 205 }, { "epoch": 0.1879133409350057, "grad_norm": 0.16220971941947937, "learning_rate": 6.730585285387465e-05, "loss": 1.1761, "step": 206 }, { "epoch": 0.1888255416191562, "grad_norm": 0.15594278275966644, "learning_rate": 6.699846198654971e-05, "loss": 1.159, "step": 207 }, { "epoch": 0.18973774230330673, "grad_norm": 0.22083112597465515, "learning_rate": 6.669034296168855e-05, "loss": 1.3703, "step": 208 }, { "epoch": 0.19064994298745724, "grad_norm": 0.2218848466873169, "learning_rate": 6.638150897808468e-05, "loss": 1.0984, "step": 209 }, { "epoch": 0.19156214367160776, "grad_norm": 0.3727222979068756, "learning_rate": 6.607197326515808e-05, "loss": 1.4564, "step": 210 }, { "epoch": 0.19247434435575828, "grad_norm": 0.14719393849372864, "learning_rate": 6.57617490823885e-05, "loss": 1.0078, "step": 211 }, { "epoch": 0.1933865450399088, "grad_norm": 0.13103777170181274, "learning_rate": 6.545084971874738e-05, "loss": 0.9745, "step": 212 }, { "epoch": 0.19429874572405928, "grad_norm": 0.13376685976982117, "learning_rate": 6.513928849212873e-05, "loss": 0.9558, "step": 213 }, { "epoch": 0.1952109464082098, "grad_norm": 0.1713114231824875, "learning_rate": 6.482707874877854e-05, "loss": 1.0795, "step": 214 }, { "epoch": 0.1961231470923603, "grad_norm": 0.14341691136360168, "learning_rate": 6.451423386272312e-05, "loss": 1.0165, "step": 215 }, { "epoch": 0.19703534777651083, "grad_norm": 0.14939330518245697, "learning_rate": 6.420076723519614e-05, "loss": 0.9657, "step": 216 }, { "epoch": 0.19794754846066134, "grad_norm": 0.1301591396331787, "learning_rate": 6.388669229406462e-05, "loss": 0.757, "step": 217 }, { "epoch": 0.19885974914481186, "grad_norm": 0.160866379737854, "learning_rate": 6.357202249325371e-05, "loss": 0.965, "step": 218 }, { "epoch": 0.19977194982896238, "grad_norm": 0.21365958452224731, "learning_rate": 6.32567713121704e-05, "loss": 1.1012, "step": 219 }, { "epoch": 0.2006841505131129, "grad_norm": 0.16345688700675964, "learning_rate": 6.294095225512603e-05, "loss": 1.0251, "step": 220 }, { "epoch": 0.2015963511972634, "grad_norm": 0.1504313200712204, "learning_rate": 6.26245788507579e-05, "loss": 1.0471, "step": 221 }, { "epoch": 0.20250855188141392, "grad_norm": 0.15645664930343628, "learning_rate": 6.230766465144967e-05, "loss": 0.932, "step": 222 }, { "epoch": 0.2034207525655644, "grad_norm": 0.16272133588790894, "learning_rate": 6.199022323275083e-05, "loss": 0.9476, "step": 223 }, { "epoch": 0.20433295324971493, "grad_norm": 0.15730346739292145, "learning_rate": 6.167226819279528e-05, "loss": 0.9678, "step": 224 }, { "epoch": 0.20524515393386544, "grad_norm": 0.16858989000320435, "learning_rate": 6.135381315171867e-05, "loss": 1.0503, "step": 225 }, { "epoch": 0.20615735461801596, "grad_norm": 0.18682138621807098, "learning_rate": 6.103487175107507e-05, "loss": 1.0477, "step": 226 }, { "epoch": 0.20706955530216647, "grad_norm": 0.17781555652618408, "learning_rate": 6.071545765325254e-05, "loss": 0.9751, "step": 227 }, { "epoch": 0.207981755986317, "grad_norm": 0.22681044042110443, "learning_rate": 6.0395584540887963e-05, "loss": 1.1294, "step": 228 }, { "epoch": 0.2088939566704675, "grad_norm": 0.20905116200447083, "learning_rate": 6.007526611628086e-05, "loss": 1.0836, "step": 229 }, { "epoch": 0.20980615735461802, "grad_norm": 0.17305400967597961, "learning_rate": 5.9754516100806423e-05, "loss": 1.0323, "step": 230 }, { "epoch": 0.21071835803876854, "grad_norm": 0.1822892129421234, "learning_rate": 5.9433348234327765e-05, "loss": 0.9336, "step": 231 }, { "epoch": 0.21163055872291905, "grad_norm": 0.2400861382484436, "learning_rate": 5.911177627460739e-05, "loss": 1.0643, "step": 232 }, { "epoch": 0.21254275940706954, "grad_norm": 0.19734802842140198, "learning_rate": 5.8789813996717736e-05, "loss": 1.1696, "step": 233 }, { "epoch": 0.21345496009122006, "grad_norm": 0.24006140232086182, "learning_rate": 5.8467475192451226e-05, "loss": 1.3169, "step": 234 }, { "epoch": 0.21436716077537057, "grad_norm": 0.1660918891429901, "learning_rate": 5.814477366972945e-05, "loss": 1.2928, "step": 235 }, { "epoch": 0.2152793614595211, "grad_norm": 0.10857274383306503, "learning_rate": 5.782172325201155e-05, "loss": 1.1304, "step": 236 }, { "epoch": 0.2161915621436716, "grad_norm": 0.09754611551761627, "learning_rate": 5.749833777770225e-05, "loss": 1.0516, "step": 237 }, { "epoch": 0.21710376282782212, "grad_norm": 0.10842633992433548, "learning_rate": 5.717463109955896e-05, "loss": 1.2546, "step": 238 }, { "epoch": 0.21801596351197264, "grad_norm": 0.11293677240610123, "learning_rate": 5.685061708409841e-05, "loss": 1.2007, "step": 239 }, { "epoch": 0.21892816419612315, "grad_norm": 0.11466323584318161, "learning_rate": 5.6526309611002594e-05, "loss": 1.2429, "step": 240 }, { "epoch": 0.21984036488027367, "grad_norm": 0.12489130347967148, "learning_rate": 5.6201722572524275e-05, "loss": 1.2082, "step": 241 }, { "epoch": 0.22075256556442419, "grad_norm": 0.14694885909557343, "learning_rate": 5.587686987289189e-05, "loss": 1.2742, "step": 242 }, { "epoch": 0.22166476624857467, "grad_norm": 0.14593251049518585, "learning_rate": 5.5551765427713884e-05, "loss": 1.3337, "step": 243 }, { "epoch": 0.2225769669327252, "grad_norm": 0.15233193337917328, "learning_rate": 5.522642316338268e-05, "loss": 1.1762, "step": 244 }, { "epoch": 0.2234891676168757, "grad_norm": 0.16144950687885284, "learning_rate": 5.490085701647805e-05, "loss": 1.2968, "step": 245 }, { "epoch": 0.22440136830102622, "grad_norm": 0.16887733340263367, "learning_rate": 5.457508093317013e-05, "loss": 1.312, "step": 246 }, { "epoch": 0.22531356898517674, "grad_norm": 0.24274863302707672, "learning_rate": 5.4249108868622086e-05, "loss": 1.4066, "step": 247 }, { "epoch": 0.22622576966932725, "grad_norm": 0.24290831387043, "learning_rate": 5.392295478639225e-05, "loss": 1.3587, "step": 248 }, { "epoch": 0.22713797035347777, "grad_norm": 0.35980361700057983, "learning_rate": 5.359663265783598e-05, "loss": 1.545, "step": 249 }, { "epoch": 0.22805017103762829, "grad_norm": 1.2798198461532593, "learning_rate": 5.327015646150716e-05, "loss": 1.8089, "step": 250 }, { "epoch": 0.2289623717217788, "grad_norm": 0.15012221038341522, "learning_rate": 5.294354018255945e-05, "loss": 1.1607, "step": 251 }, { "epoch": 0.22987457240592932, "grad_norm": 0.14231781661510468, "learning_rate": 5.26167978121472e-05, "loss": 1.0741, "step": 252 }, { "epoch": 0.2307867730900798, "grad_norm": 0.1374557614326477, "learning_rate": 5.228994334682604e-05, "loss": 1.1327, "step": 253 }, { "epoch": 0.23169897377423032, "grad_norm": 0.16241897642612457, "learning_rate": 5.196299078795344e-05, "loss": 1.1431, "step": 254 }, { "epoch": 0.23261117445838084, "grad_norm": 0.19717371463775635, "learning_rate": 5.1635954141088813e-05, "loss": 1.2522, "step": 255 }, { "epoch": 0.23352337514253135, "grad_norm": 0.16350847482681274, "learning_rate": 5.1308847415393666e-05, "loss": 1.1139, "step": 256 }, { "epoch": 0.23443557582668187, "grad_norm": 0.2092381715774536, "learning_rate": 5.0981684623031415e-05, "loss": 1.4112, "step": 257 }, { "epoch": 0.23534777651083238, "grad_norm": 0.1716463714838028, "learning_rate": 5.0654479778567223e-05, "loss": 1.0679, "step": 258 }, { "epoch": 0.2362599771949829, "grad_norm": 0.2410898059606552, "learning_rate": 5.0327246898367597e-05, "loss": 1.5, "step": 259 }, { "epoch": 0.23717217787913342, "grad_norm": 0.21982981264591217, "learning_rate": 5e-05, "loss": 1.1005, "step": 260 }, { "epoch": 0.23808437856328393, "grad_norm": 0.3865964412689209, "learning_rate": 4.9672753101632415e-05, "loss": 1.3476, "step": 261 }, { "epoch": 0.23899657924743445, "grad_norm": 0.2032884657382965, "learning_rate": 4.934552022143279e-05, "loss": 0.9498, "step": 262 }, { "epoch": 0.23990877993158494, "grad_norm": 0.1798202246427536, "learning_rate": 4.901831537696859e-05, "loss": 1.153, "step": 263 }, { "epoch": 0.24082098061573545, "grad_norm": 0.16176198422908783, "learning_rate": 4.869115258460635e-05, "loss": 0.9052, "step": 264 }, { "epoch": 0.24173318129988597, "grad_norm": 0.18311551213264465, "learning_rate": 4.83640458589112e-05, "loss": 1.1653, "step": 265 }, { "epoch": 0.24264538198403648, "grad_norm": 0.1691948026418686, "learning_rate": 4.8037009212046586e-05, "loss": 0.9367, "step": 266 }, { "epoch": 0.243557582668187, "grad_norm": 0.144092857837677, "learning_rate": 4.7710056653173976e-05, "loss": 0.9161, "step": 267 }, { "epoch": 0.24446978335233752, "grad_norm": 0.16229337453842163, "learning_rate": 4.738320218785281e-05, "loss": 0.9686, "step": 268 }, { "epoch": 0.24538198403648803, "grad_norm": 0.14550195634365082, "learning_rate": 4.7056459817440544e-05, "loss": 0.7886, "step": 269 }, { "epoch": 0.24629418472063855, "grad_norm": 0.15839822590351105, "learning_rate": 4.6729843538492847e-05, "loss": 1.0488, "step": 270 }, { "epoch": 0.24720638540478906, "grad_norm": 0.16177891194820404, "learning_rate": 4.640336734216403e-05, "loss": 1.0537, "step": 271 }, { "epoch": 0.24811858608893958, "grad_norm": 0.15026786923408508, "learning_rate": 4.607704521360776e-05, "loss": 0.9876, "step": 272 }, { "epoch": 0.24903078677309007, "grad_norm": 0.16026102006435394, "learning_rate": 4.575089113137792e-05, "loss": 0.9325, "step": 273 }, { "epoch": 0.24994298745724058, "grad_norm": 0.14951908588409424, "learning_rate": 4.542491906682989e-05, "loss": 0.9553, "step": 274 }, { "epoch": 0.2508551881413911, "grad_norm": 0.1985773742198944, "learning_rate": 4.509914298352197e-05, "loss": 1.078, "step": 275 }, { "epoch": 0.2517673888255416, "grad_norm": 0.15566202998161316, "learning_rate": 4.477357683661734e-05, "loss": 0.9121, "step": 276 }, { "epoch": 0.25267958950969216, "grad_norm": 0.15593530237674713, "learning_rate": 4.444823457228612e-05, "loss": 0.8146, "step": 277 }, { "epoch": 0.25359179019384265, "grad_norm": 0.16978751122951508, "learning_rate": 4.412313012710813e-05, "loss": 0.8579, "step": 278 }, { "epoch": 0.25450399087799314, "grad_norm": 0.22708088159561157, "learning_rate": 4.379827742747575e-05, "loss": 0.9367, "step": 279 }, { "epoch": 0.2554161915621437, "grad_norm": 0.18289873003959656, "learning_rate": 4.347369038899744e-05, "loss": 1.1297, "step": 280 }, { "epoch": 0.25632839224629417, "grad_norm": 0.19509181380271912, "learning_rate": 4.3149382915901606e-05, "loss": 0.9119, "step": 281 }, { "epoch": 0.2572405929304447, "grad_norm": 0.19992905855178833, "learning_rate": 4.282536890044104e-05, "loss": 1.1763, "step": 282 }, { "epoch": 0.2581527936145952, "grad_norm": 0.25321832299232483, "learning_rate": 4.250166222229774e-05, "loss": 1.1429, "step": 283 }, { "epoch": 0.25906499429874574, "grad_norm": 0.26218289136886597, "learning_rate": 4.2178276747988446e-05, "loss": 1.2399, "step": 284 }, { "epoch": 0.25997719498289623, "grad_norm": 0.2333952635526657, "learning_rate": 4.185522633027057e-05, "loss": 0.8556, "step": 285 }, { "epoch": 0.2608893956670468, "grad_norm": 0.1518288403749466, "learning_rate": 4.153252480754877e-05, "loss": 1.2989, "step": 286 }, { "epoch": 0.26180159635119726, "grad_norm": 0.1034812331199646, "learning_rate": 4.1210186003282275e-05, "loss": 1.168, "step": 287 }, { "epoch": 0.26271379703534775, "grad_norm": 0.10131867974996567, "learning_rate": 4.088822372539263e-05, "loss": 1.0937, "step": 288 }, { "epoch": 0.2636259977194983, "grad_norm": 0.11592473089694977, "learning_rate": 4.0566651765672246e-05, "loss": 1.1637, "step": 289 }, { "epoch": 0.2645381984036488, "grad_norm": 0.1236988753080368, "learning_rate": 4.0245483899193595e-05, "loss": 1.2724, "step": 290 }, { "epoch": 0.2654503990877993, "grad_norm": 0.13103049993515015, "learning_rate": 3.992473388371915e-05, "loss": 1.1731, "step": 291 }, { "epoch": 0.2663625997719498, "grad_norm": 0.13575758039951324, "learning_rate": 3.960441545911204e-05, "loss": 1.256, "step": 292 }, { "epoch": 0.26727480045610036, "grad_norm": 0.13911360502243042, "learning_rate": 3.928454234674747e-05, "loss": 1.329, "step": 293 }, { "epoch": 0.26818700114025085, "grad_norm": 0.1553173065185547, "learning_rate": 3.896512824892495e-05, "loss": 1.2601, "step": 294 }, { "epoch": 0.2690992018244014, "grad_norm": 0.22820927202701569, "learning_rate": 3.864618684828134e-05, "loss": 1.2263, "step": 295 }, { "epoch": 0.2700114025085519, "grad_norm": 0.251654714345932, "learning_rate": 3.832773180720475e-05, "loss": 1.3326, "step": 296 }, { "epoch": 0.27092360319270237, "grad_norm": 0.28499627113342285, "learning_rate": 3.800977676724919e-05, "loss": 1.673, "step": 297 }, { "epoch": 0.2718358038768529, "grad_norm": 0.2628072500228882, "learning_rate": 3.769233534855035e-05, "loss": 1.4846, "step": 298 }, { "epoch": 0.2727480045610034, "grad_norm": 0.3170837461948395, "learning_rate": 3.73754211492421e-05, "loss": 1.3636, "step": 299 }, { "epoch": 0.27366020524515394, "grad_norm": 0.8047922253608704, "learning_rate": 3.705904774487396e-05, "loss": 1.1731, "step": 300 }, { "epoch": 0.27366020524515394, "eval_loss": 1.1972854137420654, "eval_runtime": 113.2028, "eval_samples_per_second": 16.307, "eval_steps_per_second": 4.081, "step": 300 }, { "epoch": 0.27457240592930443, "grad_norm": 0.10662523657083511, "learning_rate": 3.6743228687829595e-05, "loss": 1.0097, "step": 301 }, { "epoch": 0.275484606613455, "grad_norm": 0.12424086779356003, "learning_rate": 3.642797750674629e-05, "loss": 1.0579, "step": 302 }, { "epoch": 0.27639680729760546, "grad_norm": 0.1385432481765747, "learning_rate": 3.6113307705935396e-05, "loss": 1.344, "step": 303 }, { "epoch": 0.277309007981756, "grad_norm": 0.16241592168807983, "learning_rate": 3.579923276480387e-05, "loss": 1.4409, "step": 304 }, { "epoch": 0.2782212086659065, "grad_norm": 0.15419059991836548, "learning_rate": 3.5485766137276894e-05, "loss": 1.3336, "step": 305 }, { "epoch": 0.27913340935005704, "grad_norm": 0.18154077231884003, "learning_rate": 3.5172921251221455e-05, "loss": 1.3504, "step": 306 }, { "epoch": 0.2800456100342075, "grad_norm": 0.1785464584827423, "learning_rate": 3.486071150787128e-05, "loss": 1.2279, "step": 307 }, { "epoch": 0.280957810718358, "grad_norm": 0.20182940363883972, "learning_rate": 3.4549150281252636e-05, "loss": 1.4907, "step": 308 }, { "epoch": 0.28187001140250856, "grad_norm": 0.22802571952342987, "learning_rate": 3.423825091761153e-05, "loss": 1.4304, "step": 309 }, { "epoch": 0.28278221208665905, "grad_norm": 0.41071566939353943, "learning_rate": 3.392802673484193e-05, "loss": 1.5391, "step": 310 }, { "epoch": 0.2836944127708096, "grad_norm": 0.2370689958333969, "learning_rate": 3.361849102191533e-05, "loss": 1.1847, "step": 311 }, { "epoch": 0.2846066134549601, "grad_norm": 0.18752293288707733, "learning_rate": 3.330965703831146e-05, "loss": 1.1175, "step": 312 }, { "epoch": 0.2855188141391106, "grad_norm": 0.1348540335893631, "learning_rate": 3.300153801345028e-05, "loss": 0.9726, "step": 313 }, { "epoch": 0.2864310148232611, "grad_norm": 0.14316678047180176, "learning_rate": 3.2694147146125345e-05, "loss": 1.0138, "step": 314 }, { "epoch": 0.28734321550741165, "grad_norm": 0.13163286447525024, "learning_rate": 3.2387497603938326e-05, "loss": 0.8848, "step": 315 }, { "epoch": 0.28825541619156214, "grad_norm": 0.19286373257637024, "learning_rate": 3.2081602522734986e-05, "loss": 1.105, "step": 316 }, { "epoch": 0.28916761687571263, "grad_norm": 0.12224922329187393, "learning_rate": 3.177647500604252e-05, "loss": 0.6901, "step": 317 }, { "epoch": 0.2900798175598632, "grad_norm": 0.1634187549352646, "learning_rate": 3.147212812450819e-05, "loss": 0.9563, "step": 318 }, { "epoch": 0.29099201824401366, "grad_norm": 0.15377987921237946, "learning_rate": 3.116857491533947e-05, "loss": 0.8784, "step": 319 }, { "epoch": 0.2919042189281642, "grad_norm": 0.13165193796157837, "learning_rate": 3.086582838174551e-05, "loss": 0.8488, "step": 320 }, { "epoch": 0.2928164196123147, "grad_norm": 0.16833990812301636, "learning_rate": 3.056390149238022e-05, "loss": 1.1677, "step": 321 }, { "epoch": 0.29372862029646524, "grad_norm": 0.15428762137889862, "learning_rate": 3.0262807180786647e-05, "loss": 0.934, "step": 322 }, { "epoch": 0.2946408209806157, "grad_norm": 0.15865963697433472, "learning_rate": 2.996255834484296e-05, "loss": 1.0805, "step": 323 }, { "epoch": 0.29555302166476627, "grad_norm": 0.14876452088356018, "learning_rate": 2.9663167846209998e-05, "loss": 0.9789, "step": 324 }, { "epoch": 0.29646522234891676, "grad_norm": 0.1823773831129074, "learning_rate": 2.936464850978027e-05, "loss": 1.0265, "step": 325 }, { "epoch": 0.2973774230330673, "grad_norm": 0.17371368408203125, "learning_rate": 2.9067013123128613e-05, "loss": 0.9221, "step": 326 }, { "epoch": 0.2982896237172178, "grad_norm": 0.19091551005840302, "learning_rate": 2.8770274435964355e-05, "loss": 0.9469, "step": 327 }, { "epoch": 0.2992018244013683, "grad_norm": 0.18162409961223602, "learning_rate": 2.8474445159585235e-05, "loss": 1.0272, "step": 328 }, { "epoch": 0.3001140250855188, "grad_norm": 0.16869843006134033, "learning_rate": 2.8179537966332887e-05, "loss": 1.1221, "step": 329 }, { "epoch": 0.3010262257696693, "grad_norm": 0.1988871544599533, "learning_rate": 2.7885565489049946e-05, "loss": 0.9206, "step": 330 }, { "epoch": 0.30193842645381985, "grad_norm": 0.17574124038219452, "learning_rate": 2.759254032053888e-05, "loss": 1.067, "step": 331 }, { "epoch": 0.30285062713797034, "grad_norm": 0.20466162264347076, "learning_rate": 2.7300475013022663e-05, "loss": 1.0049, "step": 332 }, { "epoch": 0.3037628278221209, "grad_norm": 0.30088433623313904, "learning_rate": 2.700938207760701e-05, "loss": 1.1633, "step": 333 }, { "epoch": 0.3046750285062714, "grad_norm": 0.31139159202575684, "learning_rate": 2.671927398374443e-05, "loss": 1.1374, "step": 334 }, { "epoch": 0.3055872291904219, "grad_norm": 0.20724207162857056, "learning_rate": 2.6430163158700115e-05, "loss": 1.379, "step": 335 }, { "epoch": 0.3064994298745724, "grad_norm": 0.12832605838775635, "learning_rate": 2.6142061987019577e-05, "loss": 1.1881, "step": 336 }, { "epoch": 0.3074116305587229, "grad_norm": 0.10497277975082397, "learning_rate": 2.5854982809998153e-05, "loss": 1.0799, "step": 337 }, { "epoch": 0.30832383124287344, "grad_norm": 0.1079997569322586, "learning_rate": 2.556893792515227e-05, "loss": 1.1005, "step": 338 }, { "epoch": 0.3092360319270239, "grad_norm": 0.10871066898107529, "learning_rate": 2.5283939585692783e-05, "loss": 1.0749, "step": 339 }, { "epoch": 0.31014823261117447, "grad_norm": 0.11225587129592896, "learning_rate": 2.500000000000001e-05, "loss": 1.2332, "step": 340 }, { "epoch": 0.31106043329532496, "grad_norm": 0.12466397881507874, "learning_rate": 2.471713133110078e-05, "loss": 1.2856, "step": 341 }, { "epoch": 0.3119726339794755, "grad_norm": 0.14498235285282135, "learning_rate": 2.4435345696147403e-05, "loss": 1.243, "step": 342 }, { "epoch": 0.312884834663626, "grad_norm": 0.1273733228445053, "learning_rate": 2.4154655165898627e-05, "loss": 1.199, "step": 343 }, { "epoch": 0.31379703534777653, "grad_norm": 0.17925268411636353, "learning_rate": 2.3875071764202563e-05, "loss": 1.1705, "step": 344 }, { "epoch": 0.314709236031927, "grad_norm": 0.18367154896259308, "learning_rate": 2.3596607467481603e-05, "loss": 1.412, "step": 345 }, { "epoch": 0.31562143671607756, "grad_norm": 0.18743032217025757, "learning_rate": 2.3319274204219428e-05, "loss": 1.1635, "step": 346 }, { "epoch": 0.31653363740022805, "grad_norm": 0.197452574968338, "learning_rate": 2.3043083854449988e-05, "loss": 1.3102, "step": 347 }, { "epoch": 0.31744583808437854, "grad_norm": 0.26595744490623474, "learning_rate": 2.2768048249248648e-05, "loss": 1.279, "step": 348 }, { "epoch": 0.3183580387685291, "grad_norm": 0.4088784158229828, "learning_rate": 2.2494179170225333e-05, "loss": 1.5259, "step": 349 }, { "epoch": 0.31927023945267957, "grad_norm": 0.5466248393058777, "learning_rate": 2.2221488349019903e-05, "loss": 1.5536, "step": 350 }, { "epoch": 0.3201824401368301, "grad_norm": 0.13032464683055878, "learning_rate": 2.194998746679952e-05, "loss": 1.1141, "step": 351 }, { "epoch": 0.3210946408209806, "grad_norm": 0.1586756855249405, "learning_rate": 2.167968815375837e-05, "loss": 1.1245, "step": 352 }, { "epoch": 0.32200684150513115, "grad_norm": 0.1475306749343872, "learning_rate": 2.1410601988619394e-05, "loss": 1.0502, "step": 353 }, { "epoch": 0.32291904218928164, "grad_norm": 0.16401484608650208, "learning_rate": 2.1142740498138324e-05, "loss": 1.2346, "step": 354 }, { "epoch": 0.3238312428734322, "grad_norm": 0.17974554002285004, "learning_rate": 2.08761151566099e-05, "loss": 1.1683, "step": 355 }, { "epoch": 0.32474344355758267, "grad_norm": 0.17584644258022308, "learning_rate": 2.061073738537635e-05, "loss": 1.4097, "step": 356 }, { "epoch": 0.32565564424173316, "grad_norm": 0.17970524728298187, "learning_rate": 2.034661855233815e-05, "loss": 1.2119, "step": 357 }, { "epoch": 0.3265678449258837, "grad_norm": 0.1974216103553772, "learning_rate": 2.008376997146705e-05, "loss": 1.1943, "step": 358 }, { "epoch": 0.3274800456100342, "grad_norm": 0.1703203320503235, "learning_rate": 1.982220290232143e-05, "loss": 0.9671, "step": 359 }, { "epoch": 0.32839224629418473, "grad_norm": 0.2753969132900238, "learning_rate": 1.9561928549563968e-05, "loss": 1.3897, "step": 360 }, { "epoch": 0.3293044469783352, "grad_norm": 0.3241826593875885, "learning_rate": 1.9302958062481673e-05, "loss": 1.1562, "step": 361 }, { "epoch": 0.33021664766248576, "grad_norm": 0.2132183462381363, "learning_rate": 1.9045302534508297e-05, "loss": 0.7516, "step": 362 }, { "epoch": 0.33112884834663625, "grad_norm": 0.14898429811000824, "learning_rate": 1.8788973002749112e-05, "loss": 1.1512, "step": 363 }, { "epoch": 0.3320410490307868, "grad_norm": 0.1288880854845047, "learning_rate": 1.8533980447508137e-05, "loss": 0.8832, "step": 364 }, { "epoch": 0.3329532497149373, "grad_norm": 0.15893982350826263, "learning_rate": 1.8280335791817733e-05, "loss": 1.1057, "step": 365 }, { "epoch": 0.3338654503990878, "grad_norm": 0.15993352234363556, "learning_rate": 1.8028049900970767e-05, "loss": 0.9927, "step": 366 }, { "epoch": 0.3347776510832383, "grad_norm": 0.14739666879177094, "learning_rate": 1.777713358205514e-05, "loss": 0.8977, "step": 367 }, { "epoch": 0.3356898517673888, "grad_norm": 0.1271563321352005, "learning_rate": 1.7527597583490822e-05, "loss": 0.8456, "step": 368 }, { "epoch": 0.33660205245153935, "grad_norm": 0.1543639749288559, "learning_rate": 1.7279452594569483e-05, "loss": 1.0841, "step": 369 }, { "epoch": 0.33751425313568983, "grad_norm": 0.13747648894786835, "learning_rate": 1.703270924499656e-05, "loss": 0.8802, "step": 370 }, { "epoch": 0.3384264538198404, "grad_norm": 0.1359814554452896, "learning_rate": 1.678737810443593e-05, "loss": 0.8192, "step": 371 }, { "epoch": 0.33933865450399087, "grad_norm": 0.16268639266490936, "learning_rate": 1.6543469682057106e-05, "loss": 0.8709, "step": 372 }, { "epoch": 0.3402508551881414, "grad_norm": 0.1843179315328598, "learning_rate": 1.6300994426085103e-05, "loss": 0.8591, "step": 373 }, { "epoch": 0.3411630558722919, "grad_norm": 0.19724664092063904, "learning_rate": 1.605996272335291e-05, "loss": 1.1374, "step": 374 }, { "epoch": 0.34207525655644244, "grad_norm": 0.2107182741165161, "learning_rate": 1.5820384898856434e-05, "loss": 1.0738, "step": 375 }, { "epoch": 0.34298745724059293, "grad_norm": 0.18194270133972168, "learning_rate": 1.5582271215312294e-05, "loss": 1.1537, "step": 376 }, { "epoch": 0.3438996579247434, "grad_norm": 0.14367586374282837, "learning_rate": 1.5345631872718214e-05, "loss": 0.981, "step": 377 }, { "epoch": 0.34481185860889396, "grad_norm": 0.1595829576253891, "learning_rate": 1.5110477007916001e-05, "loss": 0.9495, "step": 378 }, { "epoch": 0.34572405929304445, "grad_norm": 0.16781240701675415, "learning_rate": 1.4876816694157419e-05, "loss": 0.8829, "step": 379 }, { "epoch": 0.346636259977195, "grad_norm": 0.18402887880802155, "learning_rate": 1.4644660940672627e-05, "loss": 1.0337, "step": 380 }, { "epoch": 0.3475484606613455, "grad_norm": 0.21096819639205933, "learning_rate": 1.4414019692241437e-05, "loss": 1.235, "step": 381 }, { "epoch": 0.348460661345496, "grad_norm": 0.20575056970119476, "learning_rate": 1.4184902828767287e-05, "loss": 1.1308, "step": 382 }, { "epoch": 0.3493728620296465, "grad_norm": 0.23069782555103302, "learning_rate": 1.3957320164854059e-05, "loss": 0.7216, "step": 383 }, { "epoch": 0.35028506271379706, "grad_norm": 0.2631959617137909, "learning_rate": 1.373128144938563e-05, "loss": 1.051, "step": 384 }, { "epoch": 0.35119726339794755, "grad_norm": 0.2909131646156311, "learning_rate": 1.3506796365108232e-05, "loss": 1.4182, "step": 385 }, { "epoch": 0.3521094640820981, "grad_norm": 0.1559658795595169, "learning_rate": 1.3283874528215733e-05, "loss": 1.2366, "step": 386 }, { "epoch": 0.3530216647662486, "grad_norm": 0.09982962161302567, "learning_rate": 1.3062525487937699e-05, "loss": 1.1475, "step": 387 }, { "epoch": 0.35393386545039907, "grad_norm": 0.10416360944509506, "learning_rate": 1.2842758726130283e-05, "loss": 1.0829, "step": 388 }, { "epoch": 0.3548460661345496, "grad_norm": 0.10417391359806061, "learning_rate": 1.2624583656870154e-05, "loss": 1.0939, "step": 389 }, { "epoch": 0.3557582668187001, "grad_norm": 0.12138167768716812, "learning_rate": 1.2408009626051137e-05, "loss": 1.3434, "step": 390 }, { "epoch": 0.35667046750285064, "grad_norm": 0.13164609670639038, "learning_rate": 1.2193045910983863e-05, "loss": 1.2702, "step": 391 }, { "epoch": 0.35758266818700113, "grad_norm": 0.1307135969400406, "learning_rate": 1.1979701719998453e-05, "loss": 1.2093, "step": 392 }, { "epoch": 0.3584948688711517, "grad_norm": 0.1524544358253479, "learning_rate": 1.1767986192049984e-05, "loss": 1.179, "step": 393 }, { "epoch": 0.35940706955530216, "grad_norm": 0.16346591711044312, "learning_rate": 1.1557908396327028e-05, "loss": 1.2838, "step": 394 }, { "epoch": 0.3603192702394527, "grad_norm": 0.18474753201007843, "learning_rate": 1.134947733186315e-05, "loss": 1.2491, "step": 395 }, { "epoch": 0.3612314709236032, "grad_norm": 0.19478006660938263, "learning_rate": 1.1142701927151456e-05, "loss": 1.4218, "step": 396 }, { "epoch": 0.3621436716077537, "grad_norm": 0.22533515095710754, "learning_rate": 1.0937591039762085e-05, "loss": 1.4451, "step": 397 }, { "epoch": 0.3630558722919042, "grad_norm": 0.27334368228912354, "learning_rate": 1.0734153455962765e-05, "loss": 1.4695, "step": 398 }, { "epoch": 0.3639680729760547, "grad_norm": 0.31324154138565063, "learning_rate": 1.0532397890342505e-05, "loss": 1.4504, "step": 399 }, { "epoch": 0.36488027366020526, "grad_norm": 0.9359192848205566, "learning_rate": 1.0332332985438248e-05, "loss": 1.5549, "step": 400 }, { "epoch": 0.36488027366020526, "eval_loss": 1.1766536235809326, "eval_runtime": 112.3781, "eval_samples_per_second": 16.427, "eval_steps_per_second": 4.111, "step": 400 }, { "epoch": 0.36579247434435574, "grad_norm": 0.12196308374404907, "learning_rate": 1.013396731136465e-05, "loss": 1.0011, "step": 401 }, { "epoch": 0.3667046750285063, "grad_norm": 0.13182106614112854, "learning_rate": 9.937309365446973e-06, "loss": 1.1752, "step": 402 }, { "epoch": 0.3676168757126568, "grad_norm": 0.14486649632453918, "learning_rate": 9.742367571857091e-06, "loss": 1.1526, "step": 403 }, { "epoch": 0.3685290763968073, "grad_norm": 0.14698417484760284, "learning_rate": 9.549150281252633e-06, "loss": 1.2639, "step": 404 }, { "epoch": 0.3694412770809578, "grad_norm": 0.16532011330127716, "learning_rate": 9.357665770419244e-06, "loss": 1.3237, "step": 405 }, { "epoch": 0.3703534777651083, "grad_norm": 0.17021730542182922, "learning_rate": 9.167922241916055e-06, "loss": 1.3305, "step": 406 }, { "epoch": 0.37126567844925884, "grad_norm": 0.1707400232553482, "learning_rate": 8.97992782372432e-06, "loss": 1.1334, "step": 407 }, { "epoch": 0.37217787913340933, "grad_norm": 0.1890312284231186, "learning_rate": 8.793690568899216e-06, "loss": 1.2203, "step": 408 }, { "epoch": 0.37309007981755987, "grad_norm": 0.2785188555717468, "learning_rate": 8.609218455224893e-06, "loss": 1.3662, "step": 409 }, { "epoch": 0.37400228050171036, "grad_norm": 0.3217642307281494, "learning_rate": 8.426519384872733e-06, "loss": 1.5266, "step": 410 }, { "epoch": 0.3749144811858609, "grad_norm": 0.637722909450531, "learning_rate": 8.245601184062852e-06, "loss": 1.7853, "step": 411 }, { "epoch": 0.3758266818700114, "grad_norm": 0.15114322304725647, "learning_rate": 8.066471602728803e-06, "loss": 0.8234, "step": 412 }, { "epoch": 0.37673888255416194, "grad_norm": 0.15969091653823853, "learning_rate": 7.889138314185678e-06, "loss": 1.0191, "step": 413 }, { "epoch": 0.3776510832383124, "grad_norm": 0.16418735682964325, "learning_rate": 7.71360891480134e-06, "loss": 1.1389, "step": 414 }, { "epoch": 0.37856328392246297, "grad_norm": 0.15018145740032196, "learning_rate": 7.539890923671062e-06, "loss": 1.0788, "step": 415 }, { "epoch": 0.37947548460661346, "grad_norm": 0.14109903573989868, "learning_rate": 7.367991782295391e-06, "loss": 0.9719, "step": 416 }, { "epoch": 0.38038768529076394, "grad_norm": 0.1540416181087494, "learning_rate": 7.197918854261432e-06, "loss": 1.0884, "step": 417 }, { "epoch": 0.3812998859749145, "grad_norm": 0.14207249879837036, "learning_rate": 7.029679424927365e-06, "loss": 1.1356, "step": 418 }, { "epoch": 0.382212086659065, "grad_norm": 0.14175021648406982, "learning_rate": 6.863280701110408e-06, "loss": 0.8289, "step": 419 }, { "epoch": 0.3831242873432155, "grad_norm": 0.1370278298854828, "learning_rate": 6.698729810778065e-06, "loss": 0.9297, "step": 420 }, { "epoch": 0.384036488027366, "grad_norm": 0.2763122618198395, "learning_rate": 6.536033802742813e-06, "loss": 1.2414, "step": 421 }, { "epoch": 0.38494868871151655, "grad_norm": 0.15884122252464294, "learning_rate": 6.375199646360142e-06, "loss": 1.0608, "step": 422 }, { "epoch": 0.38586088939566704, "grad_norm": 0.14970403909683228, "learning_rate": 6.216234231230012e-06, "loss": 0.7528, "step": 423 }, { "epoch": 0.3867730900798176, "grad_norm": 0.15861846506595612, "learning_rate": 6.059144366901736e-06, "loss": 0.924, "step": 424 }, { "epoch": 0.38768529076396807, "grad_norm": 0.18967342376708984, "learning_rate": 5.903936782582253e-06, "loss": 1.0738, "step": 425 }, { "epoch": 0.38859749144811856, "grad_norm": 0.1890667825937271, "learning_rate": 5.750618126847912e-06, "loss": 1.0767, "step": 426 }, { "epoch": 0.3895096921322691, "grad_norm": 0.18360556662082672, "learning_rate": 5.599194967359639e-06, "loss": 1.137, "step": 427 }, { "epoch": 0.3904218928164196, "grad_norm": 0.2094428986310959, "learning_rate": 5.449673790581611e-06, "loss": 1.0618, "step": 428 }, { "epoch": 0.39133409350057013, "grad_norm": 0.17591425776481628, "learning_rate": 5.302061001503394e-06, "loss": 0.9315, "step": 429 }, { "epoch": 0.3922462941847206, "grad_norm": 0.19204486906528473, "learning_rate": 5.156362923365588e-06, "loss": 1.1151, "step": 430 }, { "epoch": 0.39315849486887117, "grad_norm": 0.19108223915100098, "learning_rate": 5.012585797388936e-06, "loss": 0.9007, "step": 431 }, { "epoch": 0.39407069555302165, "grad_norm": 0.24487371742725372, "learning_rate": 4.87073578250698e-06, "loss": 1.0956, "step": 432 }, { "epoch": 0.3949828962371722, "grad_norm": 0.36626699566841125, "learning_rate": 4.730818955102234e-06, "loss": 1.1817, "step": 433 }, { "epoch": 0.3958950969213227, "grad_norm": 0.20722968876361847, "learning_rate": 4.592841308745932e-06, "loss": 1.1693, "step": 434 }, { "epoch": 0.39680729760547323, "grad_norm": 0.18895329535007477, "learning_rate": 4.456808753941205e-06, "loss": 1.2137, "step": 435 }, { "epoch": 0.3977194982896237, "grad_norm": 0.11202345043420792, "learning_rate": 4.322727117869951e-06, "loss": 1.1175, "step": 436 }, { "epoch": 0.3986316989737742, "grad_norm": 0.10380493104457855, "learning_rate": 4.190602144143207e-06, "loss": 0.9969, "step": 437 }, { "epoch": 0.39954389965792475, "grad_norm": 0.10471412539482117, "learning_rate": 4.06043949255509e-06, "loss": 1.1972, "step": 438 }, { "epoch": 0.40045610034207524, "grad_norm": 0.10842544585466385, "learning_rate": 3.932244738840379e-06, "loss": 1.233, "step": 439 }, { "epoch": 0.4013683010262258, "grad_norm": 0.1057286411523819, "learning_rate": 3.8060233744356633e-06, "loss": 1.1387, "step": 440 }, { "epoch": 0.40228050171037627, "grad_norm": 0.13041551411151886, "learning_rate": 3.681780806244095e-06, "loss": 1.3469, "step": 441 }, { "epoch": 0.4031927023945268, "grad_norm": 0.14460065960884094, "learning_rate": 3.5595223564037884e-06, "loss": 1.2273, "step": 442 }, { "epoch": 0.4041049030786773, "grad_norm": 0.1310551017522812, "learning_rate": 3.4392532620598216e-06, "loss": 1.2728, "step": 443 }, { "epoch": 0.40501710376282785, "grad_norm": 0.15493756532669067, "learning_rate": 3.3209786751399187e-06, "loss": 1.2911, "step": 444 }, { "epoch": 0.40592930444697833, "grad_norm": 0.17276257276535034, "learning_rate": 3.2047036621337236e-06, "loss": 1.2539, "step": 445 }, { "epoch": 0.4068415051311288, "grad_norm": 0.19491133093833923, "learning_rate": 3.0904332038757977e-06, "loss": 1.3042, "step": 446 }, { "epoch": 0.40775370581527937, "grad_norm": 0.2031175196170807, "learning_rate": 2.978172195332263e-06, "loss": 1.3051, "step": 447 }, { "epoch": 0.40866590649942985, "grad_norm": 0.24767906963825226, "learning_rate": 2.8679254453910785e-06, "loss": 1.3451, "step": 448 }, { "epoch": 0.4095781071835804, "grad_norm": 0.3026193380355835, "learning_rate": 2.759697676656098e-06, "loss": 1.4116, "step": 449 }, { "epoch": 0.4104903078677309, "grad_norm": 0.8892145156860352, "learning_rate": 2.653493525244721e-06, "loss": 1.6982, "step": 450 }, { "epoch": 0.41140250855188143, "grad_norm": 0.13387838006019592, "learning_rate": 2.549317540589308e-06, "loss": 1.0503, "step": 451 }, { "epoch": 0.4123147092360319, "grad_norm": 0.12145627290010452, "learning_rate": 2.4471741852423237e-06, "loss": 0.9952, "step": 452 }, { "epoch": 0.41322690992018246, "grad_norm": 0.14256002008914948, "learning_rate": 2.3470678346851518e-06, "loss": 1.0721, "step": 453 }, { "epoch": 0.41413911060433295, "grad_norm": 0.13277588784694672, "learning_rate": 2.2490027771406687e-06, "loss": 1.0023, "step": 454 }, { "epoch": 0.4150513112884835, "grad_norm": 0.14663711190223694, "learning_rate": 2.152983213389559e-06, "loss": 1.335, "step": 455 }, { "epoch": 0.415963511972634, "grad_norm": 0.1930977702140808, "learning_rate": 2.0590132565903476e-06, "loss": 1.212, "step": 456 }, { "epoch": 0.41687571265678447, "grad_norm": 0.19310270249843597, "learning_rate": 1.9670969321032407e-06, "loss": 1.281, "step": 457 }, { "epoch": 0.417787913340935, "grad_norm": 0.20942994952201843, "learning_rate": 1.8772381773176417e-06, "loss": 1.5858, "step": 458 }, { "epoch": 0.4187001140250855, "grad_norm": 0.1741369217634201, "learning_rate": 1.7894408414835362e-06, "loss": 1.1413, "step": 459 }, { "epoch": 0.41961231470923605, "grad_norm": 0.2890353500843048, "learning_rate": 1.70370868554659e-06, "loss": 1.5824, "step": 460 }, { "epoch": 0.42052451539338653, "grad_norm": 0.38743922114372253, "learning_rate": 1.620045381987012e-06, "loss": 1.2338, "step": 461 }, { "epoch": 0.4214367160775371, "grad_norm": 0.4166823923587799, "learning_rate": 1.5384545146622852e-06, "loss": 1.4114, "step": 462 }, { "epoch": 0.42234891676168757, "grad_norm": 0.16984394192695618, "learning_rate": 1.4589395786535953e-06, "loss": 1.033, "step": 463 }, { "epoch": 0.4232611174458381, "grad_norm": 0.1662890464067459, "learning_rate": 1.3815039801161721e-06, "loss": 1.1955, "step": 464 }, { "epoch": 0.4241733181299886, "grad_norm": 0.15266531705856323, "learning_rate": 1.3061510361333185e-06, "loss": 1.0023, "step": 465 }, { "epoch": 0.4250855188141391, "grad_norm": 0.13052251935005188, "learning_rate": 1.232883974574367e-06, "loss": 0.7769, "step": 466 }, { "epoch": 0.42599771949828963, "grad_norm": 0.15434938669204712, "learning_rate": 1.1617059339563807e-06, "loss": 0.9081, "step": 467 }, { "epoch": 0.4269099201824401, "grad_norm": 0.1454930305480957, "learning_rate": 1.0926199633097157e-06, "loss": 0.8577, "step": 468 }, { "epoch": 0.42782212086659066, "grad_norm": 0.14353904128074646, "learning_rate": 1.0256290220474307e-06, "loss": 0.9633, "step": 469 }, { "epoch": 0.42873432155074115, "grad_norm": 0.136052668094635, "learning_rate": 9.607359798384785e-07, "loss": 0.8172, "step": 470 }, { "epoch": 0.4296465222348917, "grad_norm": 0.1676524579524994, "learning_rate": 8.979436164848088e-07, "loss": 1.0447, "step": 471 }, { "epoch": 0.4305587229190422, "grad_norm": 0.2036374807357788, "learning_rate": 8.372546218022747e-07, "loss": 0.9286, "step": 472 }, { "epoch": 0.4314709236031927, "grad_norm": 0.14977683126926422, "learning_rate": 7.786715955054203e-07, "loss": 0.9408, "step": 473 }, { "epoch": 0.4323831242873432, "grad_norm": 0.1696736216545105, "learning_rate": 7.221970470961125e-07, "loss": 0.9405, "step": 474 }, { "epoch": 0.43329532497149376, "grad_norm": 0.1618494838476181, "learning_rate": 6.678333957560512e-07, "loss": 1.0186, "step": 475 }, { "epoch": 0.43420752565564424, "grad_norm": 0.19542407989501953, "learning_rate": 6.15582970243117e-07, "loss": 0.9891, "step": 476 }, { "epoch": 0.43511972633979473, "grad_norm": 0.1792437881231308, "learning_rate": 5.654480087916303e-07, "loss": 0.786, "step": 477 }, { "epoch": 0.4360319270239453, "grad_norm": 0.19154495000839233, "learning_rate": 5.174306590164879e-07, "loss": 1.026, "step": 478 }, { "epoch": 0.43694412770809576, "grad_norm": 0.19262973964214325, "learning_rate": 4.715329778211375e-07, "loss": 1.1393, "step": 479 }, { "epoch": 0.4378563283922463, "grad_norm": 0.19536390900611877, "learning_rate": 4.277569313094809e-07, "loss": 1.0381, "step": 480 }, { "epoch": 0.4387685290763968, "grad_norm": 0.19628259539604187, "learning_rate": 3.8610439470164737e-07, "loss": 0.9981, "step": 481 }, { "epoch": 0.43968072976054734, "grad_norm": 0.19776402413845062, "learning_rate": 3.465771522536854e-07, "loss": 1.0142, "step": 482 }, { "epoch": 0.44059293044469783, "grad_norm": 0.21788012981414795, "learning_rate": 3.09176897181096e-07, "loss": 0.9793, "step": 483 }, { "epoch": 0.44150513112884837, "grad_norm": 0.2997768521308899, "learning_rate": 2.7390523158633554e-07, "loss": 1.0695, "step": 484 }, { "epoch": 0.44241733181299886, "grad_norm": 0.27953919768333435, "learning_rate": 2.407636663901591e-07, "loss": 1.3182, "step": 485 }, { "epoch": 0.44332953249714935, "grad_norm": 0.15639732778072357, "learning_rate": 2.0975362126691712e-07, "loss": 1.0899, "step": 486 }, { "epoch": 0.4442417331812999, "grad_norm": 0.12085293978452682, "learning_rate": 1.8087642458373134e-07, "loss": 1.3242, "step": 487 }, { "epoch": 0.4451539338654504, "grad_norm": 0.105677030980587, "learning_rate": 1.5413331334360182e-07, "loss": 1.2878, "step": 488 }, { "epoch": 0.4460661345496009, "grad_norm": 0.12019108980894089, "learning_rate": 1.2952543313240472e-07, "loss": 1.076, "step": 489 }, { "epoch": 0.4469783352337514, "grad_norm": 0.12178989499807358, "learning_rate": 1.0705383806982606e-07, "loss": 1.2949, "step": 490 }, { "epoch": 0.44789053591790196, "grad_norm": 0.13249683380126953, "learning_rate": 8.671949076420882e-08, "loss": 1.4278, "step": 491 }, { "epoch": 0.44880273660205244, "grad_norm": 0.13811670243740082, "learning_rate": 6.852326227130834e-08, "loss": 1.3262, "step": 492 }, { "epoch": 0.449714937286203, "grad_norm": 0.14182856678962708, "learning_rate": 5.246593205699424e-08, "loss": 1.3477, "step": 493 }, { "epoch": 0.4506271379703535, "grad_norm": 0.1499001681804657, "learning_rate": 3.8548187963854956e-08, "loss": 1.2677, "step": 494 }, { "epoch": 0.45153933865450396, "grad_norm": 0.16916415095329285, "learning_rate": 2.6770626181715773e-08, "loss": 1.3193, "step": 495 }, { "epoch": 0.4524515393386545, "grad_norm": 0.20377494394779205, "learning_rate": 1.7133751222137007e-08, "loss": 1.357, "step": 496 }, { "epoch": 0.453363740022805, "grad_norm": 0.2306216061115265, "learning_rate": 9.637975896759077e-09, "loss": 1.7053, "step": 497 }, { "epoch": 0.45427594070695554, "grad_norm": 0.2522677481174469, "learning_rate": 4.2836212996499865e-09, "loss": 1.3653, "step": 498 }, { "epoch": 0.455188141391106, "grad_norm": 0.33547061681747437, "learning_rate": 1.0709167935385455e-09, "loss": 1.4946, "step": 499 }, { "epoch": 0.45610034207525657, "grad_norm": 0.8273468017578125, "learning_rate": 0.0, "loss": 1.7554, "step": 500 }, { "epoch": 0.45610034207525657, "eval_loss": 1.195978045463562, "eval_runtime": 113.8887, "eval_samples_per_second": 16.209, "eval_steps_per_second": 4.057, "step": 500 } ], "logging_steps": 1, "max_steps": 500, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 5, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 1 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.202037058732032e+16, "train_batch_size": 8, "trial_name": null, "trial_params": null }