diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,6581 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.0, + "eval_steps": 500, + "global_step": 4674, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0021399529210357373, + "grad_norm": 2.7134010791778564, + "learning_rate": 3.5460992907801423e-06, + "loss": 12.9166, + "step": 5 + }, + { + "epoch": 0.004279905842071475, + "grad_norm": 2.857581853866577, + "learning_rate": 7.092198581560285e-06, + "loss": 12.8988, + "step": 10 + }, + { + "epoch": 0.006419858763107212, + "grad_norm": 3.6460659503936768, + "learning_rate": 1.0638297872340426e-05, + "loss": 12.596, + "step": 15 + }, + { + "epoch": 0.00855981168414295, + "grad_norm": 4.4102678298950195, + "learning_rate": 1.418439716312057e-05, + "loss": 12.0487, + "step": 20 + }, + { + "epoch": 0.010699764605178685, + "grad_norm": 4.709506034851074, + "learning_rate": 1.773049645390071e-05, + "loss": 11.2542, + "step": 25 + }, + { + "epoch": 0.012839717526214423, + "grad_norm": 3.634777069091797, + "learning_rate": 2.1276595744680852e-05, + "loss": 10.3125, + "step": 30 + }, + { + "epoch": 0.014979670447250161, + "grad_norm": 3.215787649154663, + "learning_rate": 2.4822695035460995e-05, + "loss": 9.4876, + "step": 35 + }, + { + "epoch": 0.0171196233682859, + "grad_norm": 2.2986671924591064, + "learning_rate": 2.836879432624114e-05, + "loss": 8.8462, + "step": 40 + }, + { + "epoch": 0.019259576289321637, + "grad_norm": 3.012507677078247, + "learning_rate": 3.191489361702128e-05, + "loss": 8.3983, + "step": 45 + }, + { + "epoch": 0.02139952921035737, + "grad_norm": 1.9252734184265137, + "learning_rate": 3.546099290780142e-05, + "loss": 7.9777, + "step": 50 + }, + { + "epoch": 0.02353948213139311, + "grad_norm": 4.199497699737549, + "learning_rate": 3.900709219858156e-05, + "loss": 7.6574, + "step": 55 + }, + { + "epoch": 0.025679435052428846, + "grad_norm": 6.262833118438721, + "learning_rate": 4.2553191489361704e-05, + "loss": 7.4121, + "step": 60 + }, + { + "epoch": 0.027819387973464584, + "grad_norm": 6.515402317047119, + "learning_rate": 4.609929078014185e-05, + "loss": 7.1827, + "step": 65 + }, + { + "epoch": 0.029959340894500322, + "grad_norm": 2.35896372795105, + "learning_rate": 4.964539007092199e-05, + "loss": 6.9596, + "step": 70 + }, + { + "epoch": 0.032099293815536056, + "grad_norm": 3.057620048522949, + "learning_rate": 5.319148936170213e-05, + "loss": 6.7407, + "step": 75 + }, + { + "epoch": 0.0342392467365718, + "grad_norm": 2.1466054916381836, + "learning_rate": 5.673758865248228e-05, + "loss": 6.4509, + "step": 80 + }, + { + "epoch": 0.03637919965760753, + "grad_norm": 4.194606781005859, + "learning_rate": 6.0283687943262414e-05, + "loss": 6.3802, + "step": 85 + }, + { + "epoch": 0.03851915257864327, + "grad_norm": 2.6796622276306152, + "learning_rate": 6.382978723404256e-05, + "loss": 6.1995, + "step": 90 + }, + { + "epoch": 0.04065910549967901, + "grad_norm": 4.852718353271484, + "learning_rate": 6.737588652482269e-05, + "loss": 5.9464, + "step": 95 + }, + { + "epoch": 0.04279905842071474, + "grad_norm": 4.550829887390137, + "learning_rate": 7.092198581560284e-05, + "loss": 5.7867, + "step": 100 + }, + { + "epoch": 0.04493901134175048, + "grad_norm": 5.130778789520264, + "learning_rate": 7.446808510638298e-05, + "loss": 5.6702, + "step": 105 + }, + { + "epoch": 0.04707896426278622, + "grad_norm": 4.426118850708008, + "learning_rate": 7.801418439716312e-05, + "loss": 5.5532, + "step": 110 + }, + { + "epoch": 0.04921891718382196, + "grad_norm": 3.7589707374572754, + "learning_rate": 8.156028368794327e-05, + "loss": 5.2532, + "step": 115 + }, + { + "epoch": 0.05135887010485769, + "grad_norm": 4.644890785217285, + "learning_rate": 8.510638297872341e-05, + "loss": 5.1286, + "step": 120 + }, + { + "epoch": 0.05349882302589343, + "grad_norm": 3.4871771335601807, + "learning_rate": 8.865248226950354e-05, + "loss": 4.9816, + "step": 125 + }, + { + "epoch": 0.05563877594692917, + "grad_norm": 6.377041339874268, + "learning_rate": 9.21985815602837e-05, + "loss": 4.855, + "step": 130 + }, + { + "epoch": 0.0577787288679649, + "grad_norm": 4.984015941619873, + "learning_rate": 9.574468085106384e-05, + "loss": 4.6588, + "step": 135 + }, + { + "epoch": 0.059918681789000644, + "grad_norm": 5.524806976318359, + "learning_rate": 9.929078014184398e-05, + "loss": 4.5006, + "step": 140 + }, + { + "epoch": 0.06205863471003638, + "grad_norm": 6.326210021972656, + "learning_rate": 9.999980787316854e-05, + "loss": 4.5593, + "step": 145 + }, + { + "epoch": 0.06419858763107211, + "grad_norm": 6.40717887878418, + "learning_rate": 9.999902736044627e-05, + "loss": 4.2919, + "step": 150 + }, + { + "epoch": 0.06633854055210785, + "grad_norm": 5.881182670593262, + "learning_rate": 9.999764646327135e-05, + "loss": 4.2351, + "step": 155 + }, + { + "epoch": 0.0684784934731436, + "grad_norm": 7.092776298522949, + "learning_rate": 9.99956651982255e-05, + "loss": 4.1024, + "step": 160 + }, + { + "epoch": 0.07061844639417933, + "grad_norm": 6.636219501495361, + "learning_rate": 9.999308358909955e-05, + "loss": 4.0457, + "step": 165 + }, + { + "epoch": 0.07275839931521506, + "grad_norm": 6.4725141525268555, + "learning_rate": 9.998990166689332e-05, + "loss": 3.9062, + "step": 170 + }, + { + "epoch": 0.0748983522362508, + "grad_norm": 9.732144355773926, + "learning_rate": 9.998611946981506e-05, + "loss": 3.9227, + "step": 175 + }, + { + "epoch": 0.07703830515728655, + "grad_norm": 8.540387153625488, + "learning_rate": 9.998173704328112e-05, + "loss": 3.7503, + "step": 180 + }, + { + "epoch": 0.07917825807832228, + "grad_norm": 7.108048915863037, + "learning_rate": 9.99767544399153e-05, + "loss": 3.6129, + "step": 185 + }, + { + "epoch": 0.08131821099935801, + "grad_norm": 7.507834434509277, + "learning_rate": 9.997117171954835e-05, + "loss": 3.6377, + "step": 190 + }, + { + "epoch": 0.08345816392039375, + "grad_norm": 7.859340667724609, + "learning_rate": 9.996498894921713e-05, + "loss": 3.6035, + "step": 195 + }, + { + "epoch": 0.08559811684142948, + "grad_norm": 7.093273162841797, + "learning_rate": 9.995820620316386e-05, + "loss": 3.5484, + "step": 200 + }, + { + "epoch": 0.08773806976246523, + "grad_norm": 8.989423751831055, + "learning_rate": 9.995082356283525e-05, + "loss": 3.5132, + "step": 205 + }, + { + "epoch": 0.08987802268350097, + "grad_norm": 8.387621879577637, + "learning_rate": 9.994284111688145e-05, + "loss": 3.3861, + "step": 210 + }, + { + "epoch": 0.0920179756045367, + "grad_norm": 6.809689044952393, + "learning_rate": 9.993425896115509e-05, + "loss": 3.37, + "step": 215 + }, + { + "epoch": 0.09415792852557243, + "grad_norm": 7.266931056976318, + "learning_rate": 9.992507719870998e-05, + "loss": 3.4136, + "step": 220 + }, + { + "epoch": 0.09629788144660817, + "grad_norm": 6.893076419830322, + "learning_rate": 9.991529593980006e-05, + "loss": 3.3817, + "step": 225 + }, + { + "epoch": 0.09843783436764392, + "grad_norm": 6.13765287399292, + "learning_rate": 9.990491530187791e-05, + "loss": 3.3894, + "step": 230 + }, + { + "epoch": 0.10057778728867965, + "grad_norm": 8.638238906860352, + "learning_rate": 9.989393540959343e-05, + "loss": 3.2748, + "step": 235 + }, + { + "epoch": 0.10271774020971539, + "grad_norm": 7.530318737030029, + "learning_rate": 9.98823563947923e-05, + "loss": 3.1735, + "step": 240 + }, + { + "epoch": 0.10485769313075112, + "grad_norm": 7.243378639221191, + "learning_rate": 9.987017839651447e-05, + "loss": 3.186, + "step": 245 + }, + { + "epoch": 0.10699764605178685, + "grad_norm": 8.053439140319824, + "learning_rate": 9.985740156099239e-05, + "loss": 3.2066, + "step": 250 + }, + { + "epoch": 0.1091375989728226, + "grad_norm": 7.392351150512695, + "learning_rate": 9.984402604164928e-05, + "loss": 3.132, + "step": 255 + }, + { + "epoch": 0.11127755189385834, + "grad_norm": 6.198112487792969, + "learning_rate": 9.983005199909738e-05, + "loss": 3.1682, + "step": 260 + }, + { + "epoch": 0.11341750481489407, + "grad_norm": 7.847325801849365, + "learning_rate": 9.981547960113591e-05, + "loss": 2.9939, + "step": 265 + }, + { + "epoch": 0.1155574577359298, + "grad_norm": 6.414429664611816, + "learning_rate": 9.980030902274907e-05, + "loss": 2.9657, + "step": 270 + }, + { + "epoch": 0.11769741065696554, + "grad_norm": 7.576863765716553, + "learning_rate": 9.9784540446104e-05, + "loss": 3.0261, + "step": 275 + }, + { + "epoch": 0.11983736357800129, + "grad_norm": 7.394157886505127, + "learning_rate": 9.976817406054856e-05, + "loss": 3.0425, + "step": 280 + }, + { + "epoch": 0.12197731649903702, + "grad_norm": 6.541194915771484, + "learning_rate": 9.975121006260905e-05, + "loss": 2.9808, + "step": 285 + }, + { + "epoch": 0.12411726942007276, + "grad_norm": 7.00990104675293, + "learning_rate": 9.973364865598783e-05, + "loss": 3.0134, + "step": 290 + }, + { + "epoch": 0.1262572223411085, + "grad_norm": 8.408513069152832, + "learning_rate": 9.97154900515609e-05, + "loss": 2.9945, + "step": 295 + }, + { + "epoch": 0.12839717526214423, + "grad_norm": 6.816072463989258, + "learning_rate": 9.96967344673754e-05, + "loss": 2.968, + "step": 300 + }, + { + "epoch": 0.13053712818317997, + "grad_norm": 7.750121593475342, + "learning_rate": 9.967738212864692e-05, + "loss": 2.9725, + "step": 305 + }, + { + "epoch": 0.1326770811042157, + "grad_norm": 7.111683368682861, + "learning_rate": 9.965743326775686e-05, + "loss": 2.9475, + "step": 310 + }, + { + "epoch": 0.13481703402525144, + "grad_norm": 9.491382598876953, + "learning_rate": 9.963688812424958e-05, + "loss": 2.9236, + "step": 315 + }, + { + "epoch": 0.1369569869462872, + "grad_norm": 9.146512031555176, + "learning_rate": 9.96157469448296e-05, + "loss": 2.8553, + "step": 320 + }, + { + "epoch": 0.1390969398673229, + "grad_norm": 10.622199058532715, + "learning_rate": 9.959400998335855e-05, + "loss": 2.7986, + "step": 325 + }, + { + "epoch": 0.14123689278835866, + "grad_norm": 9.17403793334961, + "learning_rate": 9.957167750085217e-05, + "loss": 2.8332, + "step": 330 + }, + { + "epoch": 0.1433768457093944, + "grad_norm": 8.492594718933105, + "learning_rate": 9.95487497654772e-05, + "loss": 2.8036, + "step": 335 + }, + { + "epoch": 0.14551679863043013, + "grad_norm": 7.058006763458252, + "learning_rate": 9.95252270525481e-05, + "loss": 2.8065, + "step": 340 + }, + { + "epoch": 0.14765675155146588, + "grad_norm": 8.662881851196289, + "learning_rate": 9.950110964452382e-05, + "loss": 2.829, + "step": 345 + }, + { + "epoch": 0.1497967044725016, + "grad_norm": 8.287664413452148, + "learning_rate": 9.947639783100429e-05, + "loss": 2.8611, + "step": 350 + }, + { + "epoch": 0.15193665739353734, + "grad_norm": 6.383536338806152, + "learning_rate": 9.945109190872706e-05, + "loss": 2.719, + "step": 355 + }, + { + "epoch": 0.1540766103145731, + "grad_norm": 9.04951000213623, + "learning_rate": 9.94251921815637e-05, + "loss": 2.7782, + "step": 360 + }, + { + "epoch": 0.1562165632356088, + "grad_norm": 9.873688697814941, + "learning_rate": 9.939869896051613e-05, + "loss": 2.7622, + "step": 365 + }, + { + "epoch": 0.15835651615664456, + "grad_norm": 8.37444019317627, + "learning_rate": 9.93716125637129e-05, + "loss": 2.7327, + "step": 370 + }, + { + "epoch": 0.16049646907768028, + "grad_norm": 8.676753997802734, + "learning_rate": 9.934393331640536e-05, + "loss": 2.6887, + "step": 375 + }, + { + "epoch": 0.16263642199871603, + "grad_norm": 8.401046752929688, + "learning_rate": 9.931566155096378e-05, + "loss": 2.6337, + "step": 380 + }, + { + "epoch": 0.16477637491975178, + "grad_norm": 9.715431213378906, + "learning_rate": 9.928679760687333e-05, + "loss": 2.7182, + "step": 385 + }, + { + "epoch": 0.1669163278407875, + "grad_norm": 7.727181434631348, + "learning_rate": 9.925734183073001e-05, + "loss": 2.7088, + "step": 390 + }, + { + "epoch": 0.16905628076182325, + "grad_norm": 8.17239761352539, + "learning_rate": 9.92272945762365e-05, + "loss": 2.6574, + "step": 395 + }, + { + "epoch": 0.17119623368285897, + "grad_norm": 9.848095893859863, + "learning_rate": 9.919665620419792e-05, + "loss": 2.6521, + "step": 400 + }, + { + "epoch": 0.17333618660389472, + "grad_norm": 6.269837856292725, + "learning_rate": 9.916542708251745e-05, + "loss": 2.6658, + "step": 405 + }, + { + "epoch": 0.17547613952493046, + "grad_norm": 9.290328979492188, + "learning_rate": 9.913360758619199e-05, + "loss": 2.6467, + "step": 410 + }, + { + "epoch": 0.17761609244596618, + "grad_norm": 8.005675315856934, + "learning_rate": 9.910119809730759e-05, + "loss": 2.7059, + "step": 415 + }, + { + "epoch": 0.17975604536700193, + "grad_norm": 7.254702091217041, + "learning_rate": 9.906819900503486e-05, + "loss": 2.6869, + "step": 420 + }, + { + "epoch": 0.18189599828803765, + "grad_norm": 8.958318710327148, + "learning_rate": 9.903461070562436e-05, + "loss": 2.6647, + "step": 425 + }, + { + "epoch": 0.1840359512090734, + "grad_norm": 10.019506454467773, + "learning_rate": 9.900043360240181e-05, + "loss": 2.6647, + "step": 430 + }, + { + "epoch": 0.18617590413010915, + "grad_norm": 9.011824607849121, + "learning_rate": 9.89656681057632e-05, + "loss": 2.6462, + "step": 435 + }, + { + "epoch": 0.18831585705114487, + "grad_norm": 8.521088600158691, + "learning_rate": 9.893031463316996e-05, + "loss": 2.6276, + "step": 440 + }, + { + "epoch": 0.19045580997218062, + "grad_norm": 9.127740859985352, + "learning_rate": 9.889437360914379e-05, + "loss": 2.5787, + "step": 445 + }, + { + "epoch": 0.19259576289321634, + "grad_norm": 7.98406457901001, + "learning_rate": 9.885784546526177e-05, + "loss": 2.6082, + "step": 450 + }, + { + "epoch": 0.19473571581425209, + "grad_norm": 8.42713451385498, + "learning_rate": 9.882073064015102e-05, + "loss": 2.5825, + "step": 455 + }, + { + "epoch": 0.19687566873528783, + "grad_norm": 10.31495189666748, + "learning_rate": 9.87830295794835e-05, + "loss": 2.5767, + "step": 460 + }, + { + "epoch": 0.19901562165632355, + "grad_norm": 8.118849754333496, + "learning_rate": 9.874474273597059e-05, + "loss": 2.5494, + "step": 465 + }, + { + "epoch": 0.2011555745773593, + "grad_norm": 7.514265537261963, + "learning_rate": 9.870587056935777e-05, + "loss": 2.5824, + "step": 470 + }, + { + "epoch": 0.20329552749839502, + "grad_norm": 9.503680229187012, + "learning_rate": 9.866641354641901e-05, + "loss": 2.5055, + "step": 475 + }, + { + "epoch": 0.20543548041943077, + "grad_norm": 6.921200275421143, + "learning_rate": 9.862637214095121e-05, + "loss": 2.628, + "step": 480 + }, + { + "epoch": 0.20757543334046652, + "grad_norm": 6.561913013458252, + "learning_rate": 9.858574683376844e-05, + "loss": 2.5504, + "step": 485 + }, + { + "epoch": 0.20971538626150224, + "grad_norm": 8.711860656738281, + "learning_rate": 9.854453811269625e-05, + "loss": 2.6222, + "step": 490 + }, + { + "epoch": 0.211855339182538, + "grad_norm": 7.749279975891113, + "learning_rate": 9.85027464725658e-05, + "loss": 2.4887, + "step": 495 + }, + { + "epoch": 0.2139952921035737, + "grad_norm": 9.228084564208984, + "learning_rate": 9.846037241520782e-05, + "loss": 2.5844, + "step": 500 + }, + { + "epoch": 0.21613524502460946, + "grad_norm": 10.207733154296875, + "learning_rate": 9.841741644944675e-05, + "loss": 2.5684, + "step": 505 + }, + { + "epoch": 0.2182751979456452, + "grad_norm": 6.548727512359619, + "learning_rate": 9.837387909109452e-05, + "loss": 2.5237, + "step": 510 + }, + { + "epoch": 0.22041515086668093, + "grad_norm": 9.262266159057617, + "learning_rate": 9.832976086294432e-05, + "loss": 2.5004, + "step": 515 + }, + { + "epoch": 0.22255510378771667, + "grad_norm": 8.879256248474121, + "learning_rate": 9.828506229476444e-05, + "loss": 2.5217, + "step": 520 + }, + { + "epoch": 0.2246950567087524, + "grad_norm": 9.322160720825195, + "learning_rate": 9.823978392329183e-05, + "loss": 2.5006, + "step": 525 + }, + { + "epoch": 0.22683500962978814, + "grad_norm": 11.270200729370117, + "learning_rate": 9.819392629222568e-05, + "loss": 2.5517, + "step": 530 + }, + { + "epoch": 0.2289749625508239, + "grad_norm": 7.910391330718994, + "learning_rate": 9.814748995222085e-05, + "loss": 2.4775, + "step": 535 + }, + { + "epoch": 0.2311149154718596, + "grad_norm": 11.13558578491211, + "learning_rate": 9.810047546088133e-05, + "loss": 2.5364, + "step": 540 + }, + { + "epoch": 0.23325486839289536, + "grad_norm": 8.814207077026367, + "learning_rate": 9.805288338275352e-05, + "loss": 2.4367, + "step": 545 + }, + { + "epoch": 0.23539482131393108, + "grad_norm": 8.249216079711914, + "learning_rate": 9.800471428931939e-05, + "loss": 2.4926, + "step": 550 + }, + { + "epoch": 0.23753477423496683, + "grad_norm": 8.129894256591797, + "learning_rate": 9.795596875898967e-05, + "loss": 2.5597, + "step": 555 + }, + { + "epoch": 0.23967472715600258, + "grad_norm": 7.8779683113098145, + "learning_rate": 9.790664737709696e-05, + "loss": 2.4822, + "step": 560 + }, + { + "epoch": 0.2418146800770383, + "grad_norm": 9.75910472869873, + "learning_rate": 9.785675073588855e-05, + "loss": 2.459, + "step": 565 + }, + { + "epoch": 0.24395463299807404, + "grad_norm": 10.04244613647461, + "learning_rate": 9.78062794345195e-05, + "loss": 2.4718, + "step": 570 + }, + { + "epoch": 0.2460945859191098, + "grad_norm": 11.239801406860352, + "learning_rate": 9.775523407904525e-05, + "loss": 2.4605, + "step": 575 + }, + { + "epoch": 0.2482345388401455, + "grad_norm": 11.514714241027832, + "learning_rate": 9.770361528241452e-05, + "loss": 2.4807, + "step": 580 + }, + { + "epoch": 0.25037449176118126, + "grad_norm": 11.684077262878418, + "learning_rate": 9.765142366446178e-05, + "loss": 2.4928, + "step": 585 + }, + { + "epoch": 0.252514444682217, + "grad_norm": 9.043937683105469, + "learning_rate": 9.759865985189995e-05, + "loss": 2.3884, + "step": 590 + }, + { + "epoch": 0.2546543976032527, + "grad_norm": 8.806846618652344, + "learning_rate": 9.754532447831285e-05, + "loss": 2.4352, + "step": 595 + }, + { + "epoch": 0.25679435052428845, + "grad_norm": 11.088272094726562, + "learning_rate": 9.749141818414749e-05, + "loss": 2.4701, + "step": 600 + }, + { + "epoch": 0.2589343034453242, + "grad_norm": 10.800878524780273, + "learning_rate": 9.743694161670646e-05, + "loss": 2.4473, + "step": 605 + }, + { + "epoch": 0.26107425636635995, + "grad_norm": 8.226394653320312, + "learning_rate": 9.73818954301402e-05, + "loss": 2.3953, + "step": 610 + }, + { + "epoch": 0.2632142092873957, + "grad_norm": 7.874807834625244, + "learning_rate": 9.732628028543906e-05, + "loss": 2.4438, + "step": 615 + }, + { + "epoch": 0.2653541622084314, + "grad_norm": 9.975316047668457, + "learning_rate": 9.727009685042538e-05, + "loss": 2.407, + "step": 620 + }, + { + "epoch": 0.26749411512946714, + "grad_norm": 7.762071132659912, + "learning_rate": 9.72133457997455e-05, + "loss": 2.4803, + "step": 625 + }, + { + "epoch": 0.2696340680505029, + "grad_norm": 8.840380668640137, + "learning_rate": 9.715602781486166e-05, + "loss": 2.393, + "step": 630 + }, + { + "epoch": 0.27177402097153863, + "grad_norm": 9.467190742492676, + "learning_rate": 9.709814358404378e-05, + "loss": 2.3655, + "step": 635 + }, + { + "epoch": 0.2739139738925744, + "grad_norm": 7.103944778442383, + "learning_rate": 9.703969380236123e-05, + "loss": 2.3414, + "step": 640 + }, + { + "epoch": 0.2760539268136101, + "grad_norm": 7.247581958770752, + "learning_rate": 9.698067917167446e-05, + "loss": 2.4331, + "step": 645 + }, + { + "epoch": 0.2781938797346458, + "grad_norm": 8.669685363769531, + "learning_rate": 9.692110040062659e-05, + "loss": 2.4395, + "step": 650 + }, + { + "epoch": 0.28033383265568157, + "grad_norm": 8.114962577819824, + "learning_rate": 9.68609582046349e-05, + "loss": 2.3229, + "step": 655 + }, + { + "epoch": 0.2824737855767173, + "grad_norm": 7.252189636230469, + "learning_rate": 9.680025330588223e-05, + "loss": 2.3969, + "step": 660 + }, + { + "epoch": 0.28461373849775307, + "grad_norm": 9.382356643676758, + "learning_rate": 9.67389864333083e-05, + "loss": 2.3446, + "step": 665 + }, + { + "epoch": 0.2867536914187888, + "grad_norm": 7.623274803161621, + "learning_rate": 9.667715832260098e-05, + "loss": 2.3587, + "step": 670 + }, + { + "epoch": 0.2888936443398245, + "grad_norm": 7.448448657989502, + "learning_rate": 9.661476971618744e-05, + "loss": 2.3671, + "step": 675 + }, + { + "epoch": 0.29103359726086026, + "grad_norm": 8.833525657653809, + "learning_rate": 9.655182136322524e-05, + "loss": 2.3739, + "step": 680 + }, + { + "epoch": 0.293173550181896, + "grad_norm": 10.259654998779297, + "learning_rate": 9.648831401959333e-05, + "loss": 2.3668, + "step": 685 + }, + { + "epoch": 0.29531350310293175, + "grad_norm": 10.228246688842773, + "learning_rate": 9.642424844788298e-05, + "loss": 2.3793, + "step": 690 + }, + { + "epoch": 0.2974534560239675, + "grad_norm": 8.286998748779297, + "learning_rate": 9.635962541738862e-05, + "loss": 2.3696, + "step": 695 + }, + { + "epoch": 0.2995934089450032, + "grad_norm": 8.957596778869629, + "learning_rate": 9.62944457040986e-05, + "loss": 2.3829, + "step": 700 + }, + { + "epoch": 0.30173336186603894, + "grad_norm": 8.238673210144043, + "learning_rate": 9.622871009068588e-05, + "loss": 2.3366, + "step": 705 + }, + { + "epoch": 0.3038733147870747, + "grad_norm": 8.315441131591797, + "learning_rate": 9.616241936649862e-05, + "loss": 2.3475, + "step": 710 + }, + { + "epoch": 0.30601326770811044, + "grad_norm": 8.352023124694824, + "learning_rate": 9.609557432755068e-05, + "loss": 2.2942, + "step": 715 + }, + { + "epoch": 0.3081532206291462, + "grad_norm": 10.10500717163086, + "learning_rate": 9.602817577651217e-05, + "loss": 2.3405, + "step": 720 + }, + { + "epoch": 0.3102931735501819, + "grad_norm": 7.825136184692383, + "learning_rate": 9.596022452269962e-05, + "loss": 2.2845, + "step": 725 + }, + { + "epoch": 0.3124331264712176, + "grad_norm": 10.414852142333984, + "learning_rate": 9.589172138206648e-05, + "loss": 2.3508, + "step": 730 + }, + { + "epoch": 0.3145730793922534, + "grad_norm": 8.343073844909668, + "learning_rate": 9.582266717719314e-05, + "loss": 2.3429, + "step": 735 + }, + { + "epoch": 0.3167130323132891, + "grad_norm": 8.42280101776123, + "learning_rate": 9.575306273727713e-05, + "loss": 2.2881, + "step": 740 + }, + { + "epoch": 0.31885298523432487, + "grad_norm": 7.317922592163086, + "learning_rate": 9.568290889812322e-05, + "loss": 2.3279, + "step": 745 + }, + { + "epoch": 0.32099293815536056, + "grad_norm": 8.232301712036133, + "learning_rate": 9.561220650213326e-05, + "loss": 2.2888, + "step": 750 + }, + { + "epoch": 0.3231328910763963, + "grad_norm": 10.112414360046387, + "learning_rate": 9.554095639829615e-05, + "loss": 2.3637, + "step": 755 + }, + { + "epoch": 0.32527284399743206, + "grad_norm": 10.274619102478027, + "learning_rate": 9.546915944217764e-05, + "loss": 2.373, + "step": 760 + }, + { + "epoch": 0.3274127969184678, + "grad_norm": 10.280336380004883, + "learning_rate": 9.539681649591002e-05, + "loss": 2.3361, + "step": 765 + }, + { + "epoch": 0.32955274983950356, + "grad_norm": 8.281644821166992, + "learning_rate": 9.532392842818177e-05, + "loss": 2.3094, + "step": 770 + }, + { + "epoch": 0.33169270276053925, + "grad_norm": 9.085750579833984, + "learning_rate": 9.52504961142272e-05, + "loss": 2.3396, + "step": 775 + }, + { + "epoch": 0.333832655681575, + "grad_norm": 8.183152198791504, + "learning_rate": 9.517652043581583e-05, + "loss": 2.2929, + "step": 780 + }, + { + "epoch": 0.33597260860261074, + "grad_norm": 10.052477836608887, + "learning_rate": 9.510200228124191e-05, + "loss": 2.3608, + "step": 785 + }, + { + "epoch": 0.3381125615236465, + "grad_norm": 8.700220108032227, + "learning_rate": 9.502694254531364e-05, + "loss": 2.2485, + "step": 790 + }, + { + "epoch": 0.34025251444468224, + "grad_norm": 7.829571723937988, + "learning_rate": 9.495134212934256e-05, + "loss": 2.3197, + "step": 795 + }, + { + "epoch": 0.34239246736571793, + "grad_norm": 7.233550071716309, + "learning_rate": 9.48752019411326e-05, + "loss": 2.2776, + "step": 800 + }, + { + "epoch": 0.3445324202867537, + "grad_norm": 8.803876876831055, + "learning_rate": 9.479852289496925e-05, + "loss": 2.3447, + "step": 805 + }, + { + "epoch": 0.34667237320778943, + "grad_norm": 7.353137016296387, + "learning_rate": 9.472130591160855e-05, + "loss": 2.3355, + "step": 810 + }, + { + "epoch": 0.3488123261288252, + "grad_norm": 8.22417163848877, + "learning_rate": 9.464355191826608e-05, + "loss": 2.2921, + "step": 815 + }, + { + "epoch": 0.3509522790498609, + "grad_norm": 9.752290725708008, + "learning_rate": 9.456526184860579e-05, + "loss": 2.31, + "step": 820 + }, + { + "epoch": 0.3530922319708966, + "grad_norm": 9.317963600158691, + "learning_rate": 9.448643664272876e-05, + "loss": 2.2816, + "step": 825 + }, + { + "epoch": 0.35523218489193237, + "grad_norm": 7.882107734680176, + "learning_rate": 9.440707724716196e-05, + "loss": 2.2285, + "step": 830 + }, + { + "epoch": 0.3573721378129681, + "grad_norm": 7.263378620147705, + "learning_rate": 9.432718461484688e-05, + "loss": 2.2865, + "step": 835 + }, + { + "epoch": 0.35951209073400386, + "grad_norm": 8.264389038085938, + "learning_rate": 9.424675970512808e-05, + "loss": 2.2278, + "step": 840 + }, + { + "epoch": 0.3616520436550396, + "grad_norm": 7.658196926116943, + "learning_rate": 9.416580348374163e-05, + "loss": 2.2733, + "step": 845 + }, + { + "epoch": 0.3637919965760753, + "grad_norm": 7.531430721282959, + "learning_rate": 9.40843169228036e-05, + "loss": 2.2451, + "step": 850 + }, + { + "epoch": 0.36593194949711105, + "grad_norm": 6.57159423828125, + "learning_rate": 9.400230100079829e-05, + "loss": 2.2292, + "step": 855 + }, + { + "epoch": 0.3680719024181468, + "grad_norm": 7.315071105957031, + "learning_rate": 9.391975670256657e-05, + "loss": 2.269, + "step": 860 + }, + { + "epoch": 0.37021185533918255, + "grad_norm": 6.916749000549316, + "learning_rate": 9.383668501929395e-05, + "loss": 2.2789, + "step": 865 + }, + { + "epoch": 0.3723518082602183, + "grad_norm": 6.972721576690674, + "learning_rate": 9.37530869484988e-05, + "loss": 2.2388, + "step": 870 + }, + { + "epoch": 0.374491761181254, + "grad_norm": 7.823705673217773, + "learning_rate": 9.36689634940203e-05, + "loss": 2.2251, + "step": 875 + }, + { + "epoch": 0.37663171410228974, + "grad_norm": 10.453600883483887, + "learning_rate": 9.358431566600636e-05, + "loss": 2.2375, + "step": 880 + }, + { + "epoch": 0.3787716670233255, + "grad_norm": 10.327241897583008, + "learning_rate": 9.349914448090156e-05, + "loss": 2.2713, + "step": 885 + }, + { + "epoch": 0.38091161994436123, + "grad_norm": 9.396978378295898, + "learning_rate": 9.34134509614349e-05, + "loss": 2.2676, + "step": 890 + }, + { + "epoch": 0.383051572865397, + "grad_norm": 7.874500751495361, + "learning_rate": 9.332723613660754e-05, + "loss": 2.2645, + "step": 895 + }, + { + "epoch": 0.3851915257864327, + "grad_norm": 9.300860404968262, + "learning_rate": 9.32405010416804e-05, + "loss": 2.2894, + "step": 900 + }, + { + "epoch": 0.3873314787074684, + "grad_norm": 7.320137977600098, + "learning_rate": 9.315324671816183e-05, + "loss": 2.2072, + "step": 905 + }, + { + "epoch": 0.38947143162850417, + "grad_norm": 7.579836368560791, + "learning_rate": 9.306547421379497e-05, + "loss": 2.2244, + "step": 910 + }, + { + "epoch": 0.3916113845495399, + "grad_norm": 8.637717247009277, + "learning_rate": 9.297718458254528e-05, + "loss": 2.1833, + "step": 915 + }, + { + "epoch": 0.39375133747057567, + "grad_norm": 6.750846862792969, + "learning_rate": 9.288837888458782e-05, + "loss": 2.2157, + "step": 920 + }, + { + "epoch": 0.39589129039161136, + "grad_norm": 8.550662994384766, + "learning_rate": 9.27990581862945e-05, + "loss": 2.224, + "step": 925 + }, + { + "epoch": 0.3980312433126471, + "grad_norm": 7.079535961151123, + "learning_rate": 9.270922356022142e-05, + "loss": 2.1862, + "step": 930 + }, + { + "epoch": 0.40017119623368286, + "grad_norm": 6.939311981201172, + "learning_rate": 9.261887608509579e-05, + "loss": 2.2017, + "step": 935 + }, + { + "epoch": 0.4023111491547186, + "grad_norm": 8.532953262329102, + "learning_rate": 9.252801684580308e-05, + "loss": 2.1758, + "step": 940 + }, + { + "epoch": 0.40445110207575435, + "grad_norm": 7.4575018882751465, + "learning_rate": 9.243664693337404e-05, + "loss": 2.1896, + "step": 945 + }, + { + "epoch": 0.40659105499679005, + "grad_norm": 7.295337677001953, + "learning_rate": 9.234476744497149e-05, + "loss": 2.252, + "step": 950 + }, + { + "epoch": 0.4087310079178258, + "grad_norm": 6.507411479949951, + "learning_rate": 9.225237948387722e-05, + "loss": 2.2629, + "step": 955 + }, + { + "epoch": 0.41087096083886154, + "grad_norm": 8.774213790893555, + "learning_rate": 9.215948415947875e-05, + "loss": 2.1715, + "step": 960 + }, + { + "epoch": 0.4130109137598973, + "grad_norm": 9.428667068481445, + "learning_rate": 9.20660825872559e-05, + "loss": 2.1916, + "step": 965 + }, + { + "epoch": 0.41515086668093304, + "grad_norm": 7.89529275894165, + "learning_rate": 9.197217588876756e-05, + "loss": 2.2247, + "step": 970 + }, + { + "epoch": 0.41729081960196873, + "grad_norm": 7.525676727294922, + "learning_rate": 9.187776519163811e-05, + "loss": 2.1985, + "step": 975 + }, + { + "epoch": 0.4194307725230045, + "grad_norm": 7.482941627502441, + "learning_rate": 9.178285162954386e-05, + "loss": 2.1853, + "step": 980 + }, + { + "epoch": 0.42157072544404023, + "grad_norm": 6.92722225189209, + "learning_rate": 9.168743634219955e-05, + "loss": 2.1629, + "step": 985 + }, + { + "epoch": 0.423710678365076, + "grad_norm": 7.211160659790039, + "learning_rate": 9.159152047534454e-05, + "loss": 2.1817, + "step": 990 + }, + { + "epoch": 0.4258506312861117, + "grad_norm": 7.496613025665283, + "learning_rate": 9.149510518072916e-05, + "loss": 2.1736, + "step": 995 + }, + { + "epoch": 0.4279905842071474, + "grad_norm": 7.383609294891357, + "learning_rate": 9.139819161610082e-05, + "loss": 2.2043, + "step": 1000 + }, + { + "epoch": 0.43013053712818317, + "grad_norm": 7.53545618057251, + "learning_rate": 9.130078094519008e-05, + "loss": 2.232, + "step": 1005 + }, + { + "epoch": 0.4322704900492189, + "grad_norm": 6.151658535003662, + "learning_rate": 9.120287433769674e-05, + "loss": 2.1939, + "step": 1010 + }, + { + "epoch": 0.43441044297025466, + "grad_norm": 7.977633953094482, + "learning_rate": 9.11044729692758e-05, + "loss": 2.2096, + "step": 1015 + }, + { + "epoch": 0.4365503958912904, + "grad_norm": 8.515596389770508, + "learning_rate": 9.100557802152328e-05, + "loss": 2.2042, + "step": 1020 + }, + { + "epoch": 0.4386903488123261, + "grad_norm": 6.877818584442139, + "learning_rate": 9.090619068196203e-05, + "loss": 2.1006, + "step": 1025 + }, + { + "epoch": 0.44083030173336185, + "grad_norm": 7.46975040435791, + "learning_rate": 9.080631214402754e-05, + "loss": 2.0974, + "step": 1030 + }, + { + "epoch": 0.4429702546543976, + "grad_norm": 7.358278751373291, + "learning_rate": 9.070594360705358e-05, + "loss": 2.1188, + "step": 1035 + }, + { + "epoch": 0.44511020757543335, + "grad_norm": 7.501420974731445, + "learning_rate": 9.060508627625779e-05, + "loss": 2.134, + "step": 1040 + }, + { + "epoch": 0.4472501604964691, + "grad_norm": 7.990818500518799, + "learning_rate": 9.050374136272717e-05, + "loss": 2.1787, + "step": 1045 + }, + { + "epoch": 0.4493901134175048, + "grad_norm": 6.6185760498046875, + "learning_rate": 9.04019100834036e-05, + "loss": 2.14, + "step": 1050 + }, + { + "epoch": 0.45153006633854054, + "grad_norm": 9.137276649475098, + "learning_rate": 9.029959366106923e-05, + "loss": 2.1632, + "step": 1055 + }, + { + "epoch": 0.4536700192595763, + "grad_norm": 8.287654876708984, + "learning_rate": 9.019679332433173e-05, + "loss": 2.1734, + "step": 1060 + }, + { + "epoch": 0.45580997218061203, + "grad_norm": 8.167359352111816, + "learning_rate": 9.009351030760958e-05, + "loss": 2.116, + "step": 1065 + }, + { + "epoch": 0.4579499251016478, + "grad_norm": 7.944402694702148, + "learning_rate": 8.998974585111729e-05, + "loss": 2.1126, + "step": 1070 + }, + { + "epoch": 0.4600898780226835, + "grad_norm": 8.373970985412598, + "learning_rate": 8.988550120085038e-05, + "loss": 2.1179, + "step": 1075 + }, + { + "epoch": 0.4622298309437192, + "grad_norm": 7.862525463104248, + "learning_rate": 8.978077760857058e-05, + "loss": 2.1599, + "step": 1080 + }, + { + "epoch": 0.46436978386475497, + "grad_norm": 6.35279655456543, + "learning_rate": 8.967557633179067e-05, + "loss": 2.1058, + "step": 1085 + }, + { + "epoch": 0.4665097367857907, + "grad_norm": 6.809432506561279, + "learning_rate": 8.956989863375944e-05, + "loss": 2.1049, + "step": 1090 + }, + { + "epoch": 0.46864968970682647, + "grad_norm": 10.011709213256836, + "learning_rate": 8.946374578344653e-05, + "loss": 2.1654, + "step": 1095 + }, + { + "epoch": 0.47078964262786216, + "grad_norm": 8.779292106628418, + "learning_rate": 8.935711905552713e-05, + "loss": 2.089, + "step": 1100 + }, + { + "epoch": 0.4729295955488979, + "grad_norm": 8.868700981140137, + "learning_rate": 8.925001973036677e-05, + "loss": 2.16, + "step": 1105 + }, + { + "epoch": 0.47506954846993366, + "grad_norm": 8.026627540588379, + "learning_rate": 8.914244909400585e-05, + "loss": 2.1683, + "step": 1110 + }, + { + "epoch": 0.4772095013909694, + "grad_norm": 6.766802787780762, + "learning_rate": 8.903440843814423e-05, + "loss": 2.1475, + "step": 1115 + }, + { + "epoch": 0.47934945431200515, + "grad_norm": 6.383385181427002, + "learning_rate": 8.892589906012577e-05, + "loss": 2.151, + "step": 1120 + }, + { + "epoch": 0.4814894072330409, + "grad_norm": 7.344244956970215, + "learning_rate": 8.881692226292269e-05, + "loss": 2.1037, + "step": 1125 + }, + { + "epoch": 0.4836293601540766, + "grad_norm": 8.650083541870117, + "learning_rate": 8.870747935511992e-05, + "loss": 2.0868, + "step": 1130 + }, + { + "epoch": 0.48576931307511234, + "grad_norm": 7.451263427734375, + "learning_rate": 8.859757165089943e-05, + "loss": 2.1287, + "step": 1135 + }, + { + "epoch": 0.4879092659961481, + "grad_norm": 7.021806716918945, + "learning_rate": 8.848720047002446e-05, + "loss": 2.1256, + "step": 1140 + }, + { + "epoch": 0.49004921891718384, + "grad_norm": 7.580230712890625, + "learning_rate": 8.837636713782358e-05, + "loss": 2.058, + "step": 1145 + }, + { + "epoch": 0.4921891718382196, + "grad_norm": 6.413537979125977, + "learning_rate": 8.826507298517489e-05, + "loss": 2.0742, + "step": 1150 + }, + { + "epoch": 0.4943291247592553, + "grad_norm": 7.502536296844482, + "learning_rate": 8.815331934848996e-05, + "loss": 2.1497, + "step": 1155 + }, + { + "epoch": 0.496469077680291, + "grad_norm": 6.689949035644531, + "learning_rate": 8.804110756969781e-05, + "loss": 2.112, + "step": 1160 + }, + { + "epoch": 0.4986090306013268, + "grad_norm": 8.030281066894531, + "learning_rate": 8.792843899622879e-05, + "loss": 2.1844, + "step": 1165 + }, + { + "epoch": 0.5007489835223625, + "grad_norm": 6.825489044189453, + "learning_rate": 8.781531498099844e-05, + "loss": 2.1259, + "step": 1170 + }, + { + "epoch": 0.5028889364433983, + "grad_norm": 7.533500671386719, + "learning_rate": 8.770173688239116e-05, + "loss": 2.0923, + "step": 1175 + }, + { + "epoch": 0.505028889364434, + "grad_norm": 6.412694454193115, + "learning_rate": 8.758770606424398e-05, + "loss": 2.1432, + "step": 1180 + }, + { + "epoch": 0.5071688422854698, + "grad_norm": 6.011632442474365, + "learning_rate": 8.747322389583013e-05, + "loss": 2.0918, + "step": 1185 + }, + { + "epoch": 0.5093087952065054, + "grad_norm": 7.687342166900635, + "learning_rate": 8.735829175184267e-05, + "loss": 2.1275, + "step": 1190 + }, + { + "epoch": 0.5114487481275412, + "grad_norm": 7.902970314025879, + "learning_rate": 8.724291101237784e-05, + "loss": 2.0638, + "step": 1195 + }, + { + "epoch": 0.5135887010485769, + "grad_norm": 6.375851631164551, + "learning_rate": 8.71270830629187e-05, + "loss": 2.1268, + "step": 1200 + }, + { + "epoch": 0.5157286539696126, + "grad_norm": 11.488832473754883, + "learning_rate": 8.701080929431824e-05, + "loss": 2.1133, + "step": 1205 + }, + { + "epoch": 0.5178686068906484, + "grad_norm": 7.46975564956665, + "learning_rate": 8.689409110278292e-05, + "loss": 2.0586, + "step": 1210 + }, + { + "epoch": 0.5200085598116841, + "grad_norm": 10.648994445800781, + "learning_rate": 8.677692988985575e-05, + "loss": 2.0896, + "step": 1215 + }, + { + "epoch": 0.5221485127327199, + "grad_norm": 8.54841136932373, + "learning_rate": 8.665932706239949e-05, + "loss": 2.0725, + "step": 1220 + }, + { + "epoch": 0.5242884656537556, + "grad_norm": 6.593012809753418, + "learning_rate": 8.654128403257982e-05, + "loss": 2.0339, + "step": 1225 + }, + { + "epoch": 0.5264284185747914, + "grad_norm": 8.725300788879395, + "learning_rate": 8.642280221784828e-05, + "loss": 2.0586, + "step": 1230 + }, + { + "epoch": 0.5285683714958271, + "grad_norm": 8.611812591552734, + "learning_rate": 8.630388304092536e-05, + "loss": 2.049, + "step": 1235 + }, + { + "epoch": 0.5307083244168628, + "grad_norm": 8.650274276733398, + "learning_rate": 8.618452792978336e-05, + "loss": 2.0758, + "step": 1240 + }, + { + "epoch": 0.5328482773378985, + "grad_norm": 5.806057453155518, + "learning_rate": 8.606473831762916e-05, + "loss": 2.075, + "step": 1245 + }, + { + "epoch": 0.5349882302589343, + "grad_norm": 6.725940227508545, + "learning_rate": 8.59445156428872e-05, + "loss": 2.1084, + "step": 1250 + }, + { + "epoch": 0.53712818317997, + "grad_norm": 7.8081889152526855, + "learning_rate": 8.582386134918204e-05, + "loss": 2.076, + "step": 1255 + }, + { + "epoch": 0.5392681361010058, + "grad_norm": 7.350069046020508, + "learning_rate": 8.570277688532112e-05, + "loss": 2.0705, + "step": 1260 + }, + { + "epoch": 0.5414080890220415, + "grad_norm": 6.163577079772949, + "learning_rate": 8.55812637052773e-05, + "loss": 2.0214, + "step": 1265 + }, + { + "epoch": 0.5435480419430773, + "grad_norm": 8.294384956359863, + "learning_rate": 8.545932326817145e-05, + "loss": 2.0657, + "step": 1270 + }, + { + "epoch": 0.545687994864113, + "grad_norm": 7.981808185577393, + "learning_rate": 8.533695703825493e-05, + "loss": 2.0762, + "step": 1275 + }, + { + "epoch": 0.5478279477851488, + "grad_norm": 7.62240743637085, + "learning_rate": 8.521416648489193e-05, + "loss": 2.0571, + "step": 1280 + }, + { + "epoch": 0.5499679007061845, + "grad_norm": 7.337045192718506, + "learning_rate": 8.509095308254191e-05, + "loss": 2.0734, + "step": 1285 + }, + { + "epoch": 0.5521078536272201, + "grad_norm": 6.613148212432861, + "learning_rate": 8.496731831074189e-05, + "loss": 2.0913, + "step": 1290 + }, + { + "epoch": 0.5542478065482559, + "grad_norm": 6.551748275756836, + "learning_rate": 8.484326365408866e-05, + "loss": 2.0183, + "step": 1295 + }, + { + "epoch": 0.5563877594692916, + "grad_norm": 6.191518783569336, + "learning_rate": 8.471879060222094e-05, + "loss": 2.0514, + "step": 1300 + }, + { + "epoch": 0.5585277123903274, + "grad_norm": 6.579594135284424, + "learning_rate": 8.459390064980146e-05, + "loss": 2.083, + "step": 1305 + }, + { + "epoch": 0.5606676653113631, + "grad_norm": 8.821534156799316, + "learning_rate": 8.446859529649917e-05, + "loss": 2.0494, + "step": 1310 + }, + { + "epoch": 0.5628076182323989, + "grad_norm": 8.188867568969727, + "learning_rate": 8.434287604697101e-05, + "loss": 2.0842, + "step": 1315 + }, + { + "epoch": 0.5649475711534346, + "grad_norm": 6.688650608062744, + "learning_rate": 8.421674441084404e-05, + "loss": 2.059, + "step": 1320 + }, + { + "epoch": 0.5670875240744704, + "grad_norm": 9.34494400024414, + "learning_rate": 8.409020190269716e-05, + "loss": 2.0587, + "step": 1325 + }, + { + "epoch": 0.5692274769955061, + "grad_norm": 9.420138359069824, + "learning_rate": 8.396325004204303e-05, + "loss": 2.0278, + "step": 1330 + }, + { + "epoch": 0.5713674299165419, + "grad_norm": 8.35175609588623, + "learning_rate": 8.383589035330977e-05, + "loss": 1.9988, + "step": 1335 + }, + { + "epoch": 0.5735073828375776, + "grad_norm": 7.808613300323486, + "learning_rate": 8.370812436582267e-05, + "loss": 2.0572, + "step": 1340 + }, + { + "epoch": 0.5756473357586133, + "grad_norm": 6.553249359130859, + "learning_rate": 8.357995361378583e-05, + "loss": 2.0234, + "step": 1345 + }, + { + "epoch": 0.577787288679649, + "grad_norm": 6.828055381774902, + "learning_rate": 8.345137963626372e-05, + "loss": 2.0428, + "step": 1350 + }, + { + "epoch": 0.5799272416006848, + "grad_norm": 8.460224151611328, + "learning_rate": 8.332240397716272e-05, + "loss": 2.05, + "step": 1355 + }, + { + "epoch": 0.5820671945217205, + "grad_norm": 6.507852554321289, + "learning_rate": 8.319302818521255e-05, + "loss": 2.0947, + "step": 1360 + }, + { + "epoch": 0.5842071474427563, + "grad_norm": 6.2819085121154785, + "learning_rate": 8.306325381394774e-05, + "loss": 2.0366, + "step": 1365 + }, + { + "epoch": 0.586347100363792, + "grad_norm": 7.773951530456543, + "learning_rate": 8.293308242168889e-05, + "loss": 2.0078, + "step": 1370 + }, + { + "epoch": 0.5884870532848278, + "grad_norm": 6.938888072967529, + "learning_rate": 8.280251557152399e-05, + "loss": 2.0739, + "step": 1375 + }, + { + "epoch": 0.5906270062058635, + "grad_norm": 7.499892234802246, + "learning_rate": 8.26715548312897e-05, + "loss": 2.011, + "step": 1380 + }, + { + "epoch": 0.5927669591268993, + "grad_norm": 6.491375923156738, + "learning_rate": 8.254020177355243e-05, + "loss": 1.9787, + "step": 1385 + }, + { + "epoch": 0.594906912047935, + "grad_norm": 7.310795307159424, + "learning_rate": 8.240845797558958e-05, + "loss": 2.0306, + "step": 1390 + }, + { + "epoch": 0.5970468649689706, + "grad_norm": 6.266601085662842, + "learning_rate": 8.227632501937045e-05, + "loss": 1.9754, + "step": 1395 + }, + { + "epoch": 0.5991868178900064, + "grad_norm": 7.139545440673828, + "learning_rate": 8.214380449153735e-05, + "loss": 2.0119, + "step": 1400 + }, + { + "epoch": 0.6013267708110421, + "grad_norm": 7.089097499847412, + "learning_rate": 8.201089798338655e-05, + "loss": 2.0456, + "step": 1405 + }, + { + "epoch": 0.6034667237320779, + "grad_norm": 6.723423004150391, + "learning_rate": 8.187760709084911e-05, + "loss": 2.0233, + "step": 1410 + }, + { + "epoch": 0.6056066766531136, + "grad_norm": 6.559820175170898, + "learning_rate": 8.174393341447177e-05, + "loss": 1.995, + "step": 1415 + }, + { + "epoch": 0.6077466295741494, + "grad_norm": 7.716728687286377, + "learning_rate": 8.160987855939766e-05, + "loss": 2.0308, + "step": 1420 + }, + { + "epoch": 0.6098865824951851, + "grad_norm": 5.852180480957031, + "learning_rate": 8.147544413534714e-05, + "loss": 2.0049, + "step": 1425 + }, + { + "epoch": 0.6120265354162209, + "grad_norm": 9.875506401062012, + "learning_rate": 8.134063175659836e-05, + "loss": 2.0243, + "step": 1430 + }, + { + "epoch": 0.6141664883372566, + "grad_norm": 7.915645122528076, + "learning_rate": 8.120544304196793e-05, + "loss": 2.0137, + "step": 1435 + }, + { + "epoch": 0.6163064412582924, + "grad_norm": 6.7806077003479, + "learning_rate": 8.10698796147915e-05, + "loss": 2.0122, + "step": 1440 + }, + { + "epoch": 0.618446394179328, + "grad_norm": 6.318131446838379, + "learning_rate": 8.093394310290421e-05, + "loss": 2.0041, + "step": 1445 + }, + { + "epoch": 0.6205863471003638, + "grad_norm": 6.302967071533203, + "learning_rate": 8.079763513862116e-05, + "loss": 2.0402, + "step": 1450 + }, + { + "epoch": 0.6227263000213995, + "grad_norm": 8.546146392822266, + "learning_rate": 8.066095735871786e-05, + "loss": 2.036, + "step": 1455 + }, + { + "epoch": 0.6248662529424353, + "grad_norm": 7.018831729888916, + "learning_rate": 8.052391140441051e-05, + "loss": 1.9625, + "step": 1460 + }, + { + "epoch": 0.627006205863471, + "grad_norm": 6.960865497589111, + "learning_rate": 8.038649892133632e-05, + "loss": 2.013, + "step": 1465 + }, + { + "epoch": 0.6291461587845067, + "grad_norm": 7.643087863922119, + "learning_rate": 8.024872155953376e-05, + "loss": 2.0234, + "step": 1470 + }, + { + "epoch": 0.6312861117055425, + "grad_norm": 6.25241231918335, + "learning_rate": 8.011058097342275e-05, + "loss": 1.9427, + "step": 1475 + }, + { + "epoch": 0.6334260646265782, + "grad_norm": 7.6290283203125, + "learning_rate": 7.997207882178474e-05, + "loss": 1.9874, + "step": 1480 + }, + { + "epoch": 0.635566017547614, + "grad_norm": 10.151625633239746, + "learning_rate": 7.983321676774285e-05, + "loss": 1.9609, + "step": 1485 + }, + { + "epoch": 0.6377059704686497, + "grad_norm": 6.939437389373779, + "learning_rate": 7.969399647874191e-05, + "loss": 2.0095, + "step": 1490 + }, + { + "epoch": 0.6398459233896854, + "grad_norm": 7.24613618850708, + "learning_rate": 7.955441962652835e-05, + "loss": 1.9874, + "step": 1495 + }, + { + "epoch": 0.6419858763107211, + "grad_norm": 8.051011085510254, + "learning_rate": 7.941448788713024e-05, + "loss": 2.0215, + "step": 1500 + }, + { + "epoch": 0.6441258292317569, + "grad_norm": 7.338986873626709, + "learning_rate": 7.927420294083705e-05, + "loss": 1.941, + "step": 1505 + }, + { + "epoch": 0.6462657821527926, + "grad_norm": 7.442719459533691, + "learning_rate": 7.91335664721796e-05, + "loss": 2.0215, + "step": 1510 + }, + { + "epoch": 0.6484057350738284, + "grad_norm": 7.5543341636657715, + "learning_rate": 7.899258016990969e-05, + "loss": 1.9694, + "step": 1515 + }, + { + "epoch": 0.6505456879948641, + "grad_norm": 7.134289264678955, + "learning_rate": 7.885124572697998e-05, + "loss": 2.0004, + "step": 1520 + }, + { + "epoch": 0.6526856409158999, + "grad_norm": 6.567378520965576, + "learning_rate": 7.870956484052346e-05, + "loss": 1.9725, + "step": 1525 + }, + { + "epoch": 0.6548255938369356, + "grad_norm": 7.110774040222168, + "learning_rate": 7.856753921183331e-05, + "loss": 2.0402, + "step": 1530 + }, + { + "epoch": 0.6569655467579714, + "grad_norm": 6.285329818725586, + "learning_rate": 7.842517054634226e-05, + "loss": 1.974, + "step": 1535 + }, + { + "epoch": 0.6591054996790071, + "grad_norm": 5.5663676261901855, + "learning_rate": 7.828246055360226e-05, + "loss": 2.0189, + "step": 1540 + }, + { + "epoch": 0.6612454526000427, + "grad_norm": 6.485752582550049, + "learning_rate": 7.813941094726384e-05, + "loss": 1.9699, + "step": 1545 + }, + { + "epoch": 0.6633854055210785, + "grad_norm": 5.899412155151367, + "learning_rate": 7.79960234450556e-05, + "loss": 1.9966, + "step": 1550 + }, + { + "epoch": 0.6655253584421142, + "grad_norm": 7.0582051277160645, + "learning_rate": 7.78522997687636e-05, + "loss": 1.9811, + "step": 1555 + }, + { + "epoch": 0.66766531136315, + "grad_norm": 6.355223178863525, + "learning_rate": 7.770824164421062e-05, + "loss": 1.9238, + "step": 1560 + }, + { + "epoch": 0.6698052642841857, + "grad_norm": 7.558047771453857, + "learning_rate": 7.756385080123546e-05, + "loss": 1.9915, + "step": 1565 + }, + { + "epoch": 0.6719452172052215, + "grad_norm": 6.747286319732666, + "learning_rate": 7.741912897367221e-05, + "loss": 1.9746, + "step": 1570 + }, + { + "epoch": 0.6740851701262572, + "grad_norm": 6.1656293869018555, + "learning_rate": 7.727407789932935e-05, + "loss": 1.9615, + "step": 1575 + }, + { + "epoch": 0.676225123047293, + "grad_norm": 7.076664447784424, + "learning_rate": 7.7128699319969e-05, + "loss": 2.0052, + "step": 1580 + }, + { + "epoch": 0.6783650759683287, + "grad_norm": 7.908239841461182, + "learning_rate": 7.698299498128587e-05, + "loss": 2.0223, + "step": 1585 + }, + { + "epoch": 0.6805050288893645, + "grad_norm": 6.080158233642578, + "learning_rate": 7.68369666328864e-05, + "loss": 1.9561, + "step": 1590 + }, + { + "epoch": 0.6826449818104001, + "grad_norm": 6.817996025085449, + "learning_rate": 7.669061602826768e-05, + "loss": 1.9811, + "step": 1595 + }, + { + "epoch": 0.6847849347314359, + "grad_norm": 5.7741475105285645, + "learning_rate": 7.654394492479648e-05, + "loss": 1.9011, + "step": 1600 + }, + { + "epoch": 0.6869248876524716, + "grad_norm": 5.87952995300293, + "learning_rate": 7.639695508368803e-05, + "loss": 1.9695, + "step": 1605 + }, + { + "epoch": 0.6890648405735074, + "grad_norm": 6.916929721832275, + "learning_rate": 7.6249648269985e-05, + "loss": 1.9536, + "step": 1610 + }, + { + "epoch": 0.6912047934945431, + "grad_norm": 8.327000617980957, + "learning_rate": 7.61020262525362e-05, + "loss": 1.961, + "step": 1615 + }, + { + "epoch": 0.6933447464155789, + "grad_norm": 7.954501628875732, + "learning_rate": 7.59540908039754e-05, + "loss": 2.0029, + "step": 1620 + }, + { + "epoch": 0.6954846993366146, + "grad_norm": 6.015987873077393, + "learning_rate": 7.580584370070001e-05, + "loss": 1.9753, + "step": 1625 + }, + { + "epoch": 0.6976246522576504, + "grad_norm": 7.0654520988464355, + "learning_rate": 7.565728672284979e-05, + "loss": 1.9378, + "step": 1630 + }, + { + "epoch": 0.6997646051786861, + "grad_norm": 7.740458965301514, + "learning_rate": 7.550842165428543e-05, + "loss": 1.9464, + "step": 1635 + }, + { + "epoch": 0.7019045580997219, + "grad_norm": 7.246245861053467, + "learning_rate": 7.535925028256717e-05, + "loss": 2.0064, + "step": 1640 + }, + { + "epoch": 0.7040445110207575, + "grad_norm": 6.526742458343506, + "learning_rate": 7.520977439893329e-05, + "loss": 1.9545, + "step": 1645 + }, + { + "epoch": 0.7061844639417932, + "grad_norm": 7.05415153503418, + "learning_rate": 7.505999579827863e-05, + "loss": 1.979, + "step": 1650 + }, + { + "epoch": 0.708324416862829, + "grad_norm": 6.696120738983154, + "learning_rate": 7.490991627913306e-05, + "loss": 1.9026, + "step": 1655 + }, + { + "epoch": 0.7104643697838647, + "grad_norm": 7.282532215118408, + "learning_rate": 7.475953764363983e-05, + "loss": 1.9877, + "step": 1660 + }, + { + "epoch": 0.7126043227049005, + "grad_norm": 5.767569541931152, + "learning_rate": 7.460886169753397e-05, + "loss": 2.005, + "step": 1665 + }, + { + "epoch": 0.7147442756259362, + "grad_norm": 6.214768409729004, + "learning_rate": 7.445789025012055e-05, + "loss": 1.949, + "step": 1670 + }, + { + "epoch": 0.716884228546972, + "grad_norm": 6.871894836425781, + "learning_rate": 7.430662511425308e-05, + "loss": 1.9725, + "step": 1675 + }, + { + "epoch": 0.7190241814680077, + "grad_norm": 7.443453311920166, + "learning_rate": 7.415506810631155e-05, + "loss": 1.9316, + "step": 1680 + }, + { + "epoch": 0.7211641343890435, + "grad_norm": 6.8776140213012695, + "learning_rate": 7.400322104618085e-05, + "loss": 1.9556, + "step": 1685 + }, + { + "epoch": 0.7233040873100792, + "grad_norm": 6.675306797027588, + "learning_rate": 7.385108575722868e-05, + "loss": 1.9833, + "step": 1690 + }, + { + "epoch": 0.7254440402311149, + "grad_norm": 6.732053756713867, + "learning_rate": 7.369866406628385e-05, + "loss": 1.9036, + "step": 1695 + }, + { + "epoch": 0.7275839931521506, + "grad_norm": 7.516349792480469, + "learning_rate": 7.354595780361423e-05, + "loss": 1.9876, + "step": 1700 + }, + { + "epoch": 0.7297239460731864, + "grad_norm": 6.245758056640625, + "learning_rate": 7.339296880290481e-05, + "loss": 1.9403, + "step": 1705 + }, + { + "epoch": 0.7318638989942221, + "grad_norm": 6.886528015136719, + "learning_rate": 7.323969890123565e-05, + "loss": 1.9668, + "step": 1710 + }, + { + "epoch": 0.7340038519152579, + "grad_norm": 7.249610900878906, + "learning_rate": 7.308614993905992e-05, + "loss": 2.0097, + "step": 1715 + }, + { + "epoch": 0.7361438048362936, + "grad_norm": 6.901829719543457, + "learning_rate": 7.293232376018164e-05, + "loss": 1.9935, + "step": 1720 + }, + { + "epoch": 0.7382837577573294, + "grad_norm": 7.697319984436035, + "learning_rate": 7.277822221173367e-05, + "loss": 1.9295, + "step": 1725 + }, + { + "epoch": 0.7404237106783651, + "grad_norm": 6.5634565353393555, + "learning_rate": 7.262384714415551e-05, + "loss": 1.9149, + "step": 1730 + }, + { + "epoch": 0.7425636635994008, + "grad_norm": 6.889857769012451, + "learning_rate": 7.2469200411171e-05, + "loss": 1.947, + "step": 1735 + }, + { + "epoch": 0.7447036165204366, + "grad_norm": 7.622896194458008, + "learning_rate": 7.231428386976618e-05, + "loss": 1.8894, + "step": 1740 + }, + { + "epoch": 0.7468435694414722, + "grad_norm": 7.235098838806152, + "learning_rate": 7.215909938016684e-05, + "loss": 1.9483, + "step": 1745 + }, + { + "epoch": 0.748983522362508, + "grad_norm": 5.822272300720215, + "learning_rate": 7.200364880581637e-05, + "loss": 1.9301, + "step": 1750 + }, + { + "epoch": 0.7511234752835437, + "grad_norm": 6.663504123687744, + "learning_rate": 7.184793401335322e-05, + "loss": 1.9482, + "step": 1755 + }, + { + "epoch": 0.7532634282045795, + "grad_norm": 7.109114170074463, + "learning_rate": 7.169195687258859e-05, + "loss": 1.9659, + "step": 1760 + }, + { + "epoch": 0.7554033811256152, + "grad_norm": 7.012261867523193, + "learning_rate": 7.15357192564839e-05, + "loss": 1.9284, + "step": 1765 + }, + { + "epoch": 0.757543334046651, + "grad_norm": 7.140257835388184, + "learning_rate": 7.137922304112838e-05, + "loss": 1.9185, + "step": 1770 + }, + { + "epoch": 0.7596832869676867, + "grad_norm": 5.866714954376221, + "learning_rate": 7.122247010571647e-05, + "loss": 1.9535, + "step": 1775 + }, + { + "epoch": 0.7618232398887225, + "grad_norm": 6.297852516174316, + "learning_rate": 7.106546233252528e-05, + "loss": 1.9455, + "step": 1780 + }, + { + "epoch": 0.7639631928097582, + "grad_norm": 5.564944267272949, + "learning_rate": 7.090820160689201e-05, + "loss": 1.9153, + "step": 1785 + }, + { + "epoch": 0.766103145730794, + "grad_norm": 6.030523777008057, + "learning_rate": 7.07506898171913e-05, + "loss": 1.9386, + "step": 1790 + }, + { + "epoch": 0.7682430986518297, + "grad_norm": 7.070890426635742, + "learning_rate": 7.059292885481253e-05, + "loss": 1.9576, + "step": 1795 + }, + { + "epoch": 0.7703830515728654, + "grad_norm": 7.497870922088623, + "learning_rate": 7.04349206141371e-05, + "loss": 1.8669, + "step": 1800 + }, + { + "epoch": 0.7725230044939011, + "grad_norm": 5.917843818664551, + "learning_rate": 7.02766669925158e-05, + "loss": 1.9455, + "step": 1805 + }, + { + "epoch": 0.7746629574149368, + "grad_norm": 6.862101078033447, + "learning_rate": 7.011816989024583e-05, + "loss": 1.9464, + "step": 1810 + }, + { + "epoch": 0.7768029103359726, + "grad_norm": 6.385555267333984, + "learning_rate": 6.995943121054816e-05, + "loss": 1.9477, + "step": 1815 + }, + { + "epoch": 0.7789428632570083, + "grad_norm": 6.692836761474609, + "learning_rate": 6.980045285954456e-05, + "loss": 1.916, + "step": 1820 + }, + { + "epoch": 0.7810828161780441, + "grad_norm": 7.266310214996338, + "learning_rate": 6.964123674623475e-05, + "loss": 1.929, + "step": 1825 + }, + { + "epoch": 0.7832227690990798, + "grad_norm": 6.483401775360107, + "learning_rate": 6.948178478247355e-05, + "loss": 1.9205, + "step": 1830 + }, + { + "epoch": 0.7853627220201156, + "grad_norm": 7.336221694946289, + "learning_rate": 6.932209888294777e-05, + "loss": 1.9402, + "step": 1835 + }, + { + "epoch": 0.7875026749411513, + "grad_norm": 7.22684383392334, + "learning_rate": 6.916218096515336e-05, + "loss": 1.8924, + "step": 1840 + }, + { + "epoch": 0.7896426278621871, + "grad_norm": 6.436882972717285, + "learning_rate": 6.900203294937229e-05, + "loss": 1.9378, + "step": 1845 + }, + { + "epoch": 0.7917825807832227, + "grad_norm": 6.200204849243164, + "learning_rate": 6.884165675864954e-05, + "loss": 1.864, + "step": 1850 + }, + { + "epoch": 0.7939225337042585, + "grad_norm": 6.617162227630615, + "learning_rate": 6.868105431877002e-05, + "loss": 1.9279, + "step": 1855 + }, + { + "epoch": 0.7960624866252942, + "grad_norm": 5.673040866851807, + "learning_rate": 6.85202275582354e-05, + "loss": 1.8682, + "step": 1860 + }, + { + "epoch": 0.79820243954633, + "grad_norm": 6.891549110412598, + "learning_rate": 6.835917840824097e-05, + "loss": 1.8942, + "step": 1865 + }, + { + "epoch": 0.8003423924673657, + "grad_norm": 6.0316386222839355, + "learning_rate": 6.819790880265246e-05, + "loss": 1.906, + "step": 1870 + }, + { + "epoch": 0.8024823453884015, + "grad_norm": 7.065370559692383, + "learning_rate": 6.803642067798284e-05, + "loss": 1.9118, + "step": 1875 + }, + { + "epoch": 0.8046222983094372, + "grad_norm": 6.079885959625244, + "learning_rate": 6.7874715973369e-05, + "loss": 1.934, + "step": 1880 + }, + { + "epoch": 0.806762251230473, + "grad_norm": 6.481017112731934, + "learning_rate": 6.771279663054853e-05, + "loss": 1.9257, + "step": 1885 + }, + { + "epoch": 0.8089022041515087, + "grad_norm": 6.038636207580566, + "learning_rate": 6.755066459383637e-05, + "loss": 1.8815, + "step": 1890 + }, + { + "epoch": 0.8110421570725445, + "grad_norm": 6.196339130401611, + "learning_rate": 6.738832181010149e-05, + "loss": 1.9115, + "step": 1895 + }, + { + "epoch": 0.8131821099935801, + "grad_norm": 6.638113498687744, + "learning_rate": 6.722577022874345e-05, + "loss": 1.9637, + "step": 1900 + }, + { + "epoch": 0.8153220629146158, + "grad_norm": 6.1240973472595215, + "learning_rate": 6.706301180166909e-05, + "loss": 1.9453, + "step": 1905 + }, + { + "epoch": 0.8174620158356516, + "grad_norm": 6.026691436767578, + "learning_rate": 6.690004848326898e-05, + "loss": 1.8705, + "step": 1910 + }, + { + "epoch": 0.8196019687566873, + "grad_norm": 6.312042713165283, + "learning_rate": 6.673688223039406e-05, + "loss": 1.8869, + "step": 1915 + }, + { + "epoch": 0.8217419216777231, + "grad_norm": 6.738498687744141, + "learning_rate": 6.65735150023321e-05, + "loss": 1.9406, + "step": 1920 + }, + { + "epoch": 0.8238818745987588, + "grad_norm": 5.581632614135742, + "learning_rate": 6.64099487607841e-05, + "loss": 1.8928, + "step": 1925 + }, + { + "epoch": 0.8260218275197946, + "grad_norm": 6.734776973724365, + "learning_rate": 6.624618546984082e-05, + "loss": 1.8833, + "step": 1930 + }, + { + "epoch": 0.8281617804408303, + "grad_norm": 6.365813732147217, + "learning_rate": 6.608222709595925e-05, + "loss": 1.8952, + "step": 1935 + }, + { + "epoch": 0.8303017333618661, + "grad_norm": 6.569833278656006, + "learning_rate": 6.59180756079388e-05, + "loss": 1.9004, + "step": 1940 + }, + { + "epoch": 0.8324416862829018, + "grad_norm": 5.505297660827637, + "learning_rate": 6.575373297689786e-05, + "loss": 1.8726, + "step": 1945 + }, + { + "epoch": 0.8345816392039375, + "grad_norm": 6.422504901885986, + "learning_rate": 6.558920117625005e-05, + "loss": 1.9392, + "step": 1950 + }, + { + "epoch": 0.8367215921249732, + "grad_norm": 5.929808139801025, + "learning_rate": 6.542448218168049e-05, + "loss": 1.8829, + "step": 1955 + }, + { + "epoch": 0.838861545046009, + "grad_norm": 5.6545796394348145, + "learning_rate": 6.525957797112211e-05, + "loss": 1.8625, + "step": 1960 + }, + { + "epoch": 0.8410014979670447, + "grad_norm": 5.610157489776611, + "learning_rate": 6.509449052473193e-05, + "loss": 1.9184, + "step": 1965 + }, + { + "epoch": 0.8431414508880805, + "grad_norm": 6.630683898925781, + "learning_rate": 6.492922182486722e-05, + "loss": 1.9024, + "step": 1970 + }, + { + "epoch": 0.8452814038091162, + "grad_norm": 5.480852127075195, + "learning_rate": 6.476377385606175e-05, + "loss": 1.8924, + "step": 1975 + }, + { + "epoch": 0.847421356730152, + "grad_norm": 6.156518459320068, + "learning_rate": 6.459814860500194e-05, + "loss": 1.8499, + "step": 1980 + }, + { + "epoch": 0.8495613096511877, + "grad_norm": 6.882572174072266, + "learning_rate": 6.443234806050298e-05, + "loss": 1.8714, + "step": 1985 + }, + { + "epoch": 0.8517012625722234, + "grad_norm": 6.394567489624023, + "learning_rate": 6.4266374213485e-05, + "loss": 1.889, + "step": 1990 + }, + { + "epoch": 0.8538412154932592, + "grad_norm": 5.844759941101074, + "learning_rate": 6.41002290569491e-05, + "loss": 1.919, + "step": 1995 + }, + { + "epoch": 0.8559811684142948, + "grad_norm": 5.819794178009033, + "learning_rate": 6.393391458595345e-05, + "loss": 1.8827, + "step": 2000 + }, + { + "epoch": 0.8581211213353306, + "grad_norm": 7.27986478805542, + "learning_rate": 6.37674327975894e-05, + "loss": 1.8559, + "step": 2005 + }, + { + "epoch": 0.8602610742563663, + "grad_norm": 6.1043195724487305, + "learning_rate": 6.360078569095734e-05, + "loss": 1.8751, + "step": 2010 + }, + { + "epoch": 0.8624010271774021, + "grad_norm": 6.8420090675354, + "learning_rate": 6.343397526714284e-05, + "loss": 1.8557, + "step": 2015 + }, + { + "epoch": 0.8645409800984378, + "grad_norm": 6.02893590927124, + "learning_rate": 6.326700352919259e-05, + "loss": 1.9407, + "step": 2020 + }, + { + "epoch": 0.8666809330194736, + "grad_norm": 6.999791145324707, + "learning_rate": 6.309987248209029e-05, + "loss": 1.8652, + "step": 2025 + }, + { + "epoch": 0.8688208859405093, + "grad_norm": 6.686702728271484, + "learning_rate": 6.293258413273262e-05, + "loss": 1.9068, + "step": 2030 + }, + { + "epoch": 0.8709608388615451, + "grad_norm": 6.408459186553955, + "learning_rate": 6.276514048990515e-05, + "loss": 1.8797, + "step": 2035 + }, + { + "epoch": 0.8731007917825808, + "grad_norm": 6.277412414550781, + "learning_rate": 6.259754356425818e-05, + "loss": 1.8417, + "step": 2040 + }, + { + "epoch": 0.8752407447036166, + "grad_norm": 5.6100287437438965, + "learning_rate": 6.242979536828262e-05, + "loss": 1.87, + "step": 2045 + }, + { + "epoch": 0.8773806976246522, + "grad_norm": 6.132706165313721, + "learning_rate": 6.226189791628583e-05, + "loss": 1.9009, + "step": 2050 + }, + { + "epoch": 0.879520650545688, + "grad_norm": 5.885615825653076, + "learning_rate": 6.209385322436746e-05, + "loss": 1.8624, + "step": 2055 + }, + { + "epoch": 0.8816606034667237, + "grad_norm": 6.174984455108643, + "learning_rate": 6.192566331039514e-05, + "loss": 1.8723, + "step": 2060 + }, + { + "epoch": 0.8838005563877595, + "grad_norm": 5.804656982421875, + "learning_rate": 6.175733019398034e-05, + "loss": 1.8639, + "step": 2065 + }, + { + "epoch": 0.8859405093087952, + "grad_norm": 7.446183204650879, + "learning_rate": 6.15888558964541e-05, + "loss": 1.8703, + "step": 2070 + }, + { + "epoch": 0.888080462229831, + "grad_norm": 6.513533115386963, + "learning_rate": 6.142024244084278e-05, + "loss": 1.8334, + "step": 2075 + }, + { + "epoch": 0.8902204151508667, + "grad_norm": 6.254002571105957, + "learning_rate": 6.125149185184369e-05, + "loss": 1.8476, + "step": 2080 + }, + { + "epoch": 0.8923603680719024, + "grad_norm": 5.820918560028076, + "learning_rate": 6.108260615580086e-05, + "loss": 1.9445, + "step": 2085 + }, + { + "epoch": 0.8945003209929382, + "grad_norm": 7.240386962890625, + "learning_rate": 6.091358738068064e-05, + "loss": 1.8563, + "step": 2090 + }, + { + "epoch": 0.8966402739139739, + "grad_norm": 6.337396621704102, + "learning_rate": 6.074443755604744e-05, + "loss": 1.8851, + "step": 2095 + }, + { + "epoch": 0.8987802268350096, + "grad_norm": 7.422278881072998, + "learning_rate": 6.0575158713039234e-05, + "loss": 1.8112, + "step": 2100 + }, + { + "epoch": 0.9009201797560453, + "grad_norm": 6.84944486618042, + "learning_rate": 6.040575288434329e-05, + "loss": 1.865, + "step": 2105 + }, + { + "epoch": 0.9030601326770811, + "grad_norm": 6.545510768890381, + "learning_rate": 6.02362221041717e-05, + "loss": 1.8829, + "step": 2110 + }, + { + "epoch": 0.9052000855981168, + "grad_norm": 6.097507476806641, + "learning_rate": 6.006656840823696e-05, + "loss": 1.8363, + "step": 2115 + }, + { + "epoch": 0.9073400385191526, + "grad_norm": 7.34000301361084, + "learning_rate": 5.9896793833727496e-05, + "loss": 1.8738, + "step": 2120 + }, + { + "epoch": 0.9094799914401883, + "grad_norm": 5.860642433166504, + "learning_rate": 5.972690041928331e-05, + "loss": 1.8792, + "step": 2125 + }, + { + "epoch": 0.9116199443612241, + "grad_norm": 6.345127582550049, + "learning_rate": 5.9556890204971326e-05, + "loss": 1.838, + "step": 2130 + }, + { + "epoch": 0.9137598972822598, + "grad_norm": 6.226019382476807, + "learning_rate": 5.93867652322611e-05, + "loss": 1.881, + "step": 2135 + }, + { + "epoch": 0.9158998502032956, + "grad_norm": 6.527356147766113, + "learning_rate": 5.921652754400011e-05, + "loss": 1.8406, + "step": 2140 + }, + { + "epoch": 0.9180398031243313, + "grad_norm": 5.739786148071289, + "learning_rate": 5.904617918438936e-05, + "loss": 1.8496, + "step": 2145 + }, + { + "epoch": 0.920179756045367, + "grad_norm": 7.399641990661621, + "learning_rate": 5.887572219895877e-05, + "loss": 1.8212, + "step": 2150 + }, + { + "epoch": 0.9223197089664027, + "grad_norm": 5.731332302093506, + "learning_rate": 5.8705158634542644e-05, + "loss": 1.9046, + "step": 2155 + }, + { + "epoch": 0.9244596618874384, + "grad_norm": 7.024575710296631, + "learning_rate": 5.853449053925505e-05, + "loss": 1.8799, + "step": 2160 + }, + { + "epoch": 0.9265996148084742, + "grad_norm": 5.317823886871338, + "learning_rate": 5.836371996246528e-05, + "loss": 1.8679, + "step": 2165 + }, + { + "epoch": 0.9287395677295099, + "grad_norm": 5.5442795753479, + "learning_rate": 5.819284895477323e-05, + "loss": 1.8641, + "step": 2170 + }, + { + "epoch": 0.9308795206505457, + "grad_norm": 5.924790382385254, + "learning_rate": 5.80218795679847e-05, + "loss": 1.8456, + "step": 2175 + }, + { + "epoch": 0.9330194735715814, + "grad_norm": 5.661588668823242, + "learning_rate": 5.78508138550869e-05, + "loss": 1.848, + "step": 2180 + }, + { + "epoch": 0.9351594264926172, + "grad_norm": 5.443496227264404, + "learning_rate": 5.7679653870223673e-05, + "loss": 1.8579, + "step": 2185 + }, + { + "epoch": 0.9372993794136529, + "grad_norm": 5.579583644866943, + "learning_rate": 5.750840166867085e-05, + "loss": 1.8477, + "step": 2190 + }, + { + "epoch": 0.9394393323346887, + "grad_norm": 6.413700580596924, + "learning_rate": 5.733705930681165e-05, + "loss": 1.8889, + "step": 2195 + }, + { + "epoch": 0.9415792852557243, + "grad_norm": 6.018016338348389, + "learning_rate": 5.7165628842111866e-05, + "loss": 1.8509, + "step": 2200 + }, + { + "epoch": 0.9437192381767601, + "grad_norm": 6.163970470428467, + "learning_rate": 5.699411233309528e-05, + "loss": 1.8492, + "step": 2205 + }, + { + "epoch": 0.9458591910977958, + "grad_norm": 7.421061992645264, + "learning_rate": 5.682251183931886e-05, + "loss": 1.8493, + "step": 2210 + }, + { + "epoch": 0.9479991440188316, + "grad_norm": 6.061490058898926, + "learning_rate": 5.6650829421348065e-05, + "loss": 1.8128, + "step": 2215 + }, + { + "epoch": 0.9501390969398673, + "grad_norm": 6.558000087738037, + "learning_rate": 5.647906714073208e-05, + "loss": 1.8476, + "step": 2220 + }, + { + "epoch": 0.9522790498609031, + "grad_norm": 7.766129970550537, + "learning_rate": 5.630722705997908e-05, + "loss": 1.8357, + "step": 2225 + }, + { + "epoch": 0.9544190027819388, + "grad_norm": 6.785080909729004, + "learning_rate": 5.6135311242531473e-05, + "loss": 1.845, + "step": 2230 + }, + { + "epoch": 0.9565589557029746, + "grad_norm": 5.932000637054443, + "learning_rate": 5.59633217527411e-05, + "loss": 1.8277, + "step": 2235 + }, + { + "epoch": 0.9586989086240103, + "grad_norm": 6.414392948150635, + "learning_rate": 5.5791260655844424e-05, + "loss": 1.9083, + "step": 2240 + }, + { + "epoch": 0.960838861545046, + "grad_norm": 5.878129482269287, + "learning_rate": 5.5619130017937806e-05, + "loss": 1.8427, + "step": 2245 + }, + { + "epoch": 0.9629788144660818, + "grad_norm": 5.418539524078369, + "learning_rate": 5.5446931905952624e-05, + "loss": 1.8509, + "step": 2250 + }, + { + "epoch": 0.9651187673871174, + "grad_norm": 5.333776473999023, + "learning_rate": 5.527466838763049e-05, + "loss": 1.8308, + "step": 2255 + }, + { + "epoch": 0.9672587203081532, + "grad_norm": 5.526288032531738, + "learning_rate": 5.510234153149839e-05, + "loss": 1.8443, + "step": 2260 + }, + { + "epoch": 0.9693986732291889, + "grad_norm": 5.585949897766113, + "learning_rate": 5.4929953406843906e-05, + "loss": 1.8008, + "step": 2265 + }, + { + "epoch": 0.9715386261502247, + "grad_norm": 6.400721549987793, + "learning_rate": 5.475750608369029e-05, + "loss": 1.8238, + "step": 2270 + }, + { + "epoch": 0.9736785790712604, + "grad_norm": 6.028564453125, + "learning_rate": 5.4585001632771656e-05, + "loss": 1.8195, + "step": 2275 + }, + { + "epoch": 0.9758185319922962, + "grad_norm": 6.440245628356934, + "learning_rate": 5.4412442125508113e-05, + "loss": 1.8225, + "step": 2280 + }, + { + "epoch": 0.9779584849133319, + "grad_norm": 6.663422584533691, + "learning_rate": 5.4239829633980866e-05, + "loss": 1.8487, + "step": 2285 + }, + { + "epoch": 0.9800984378343677, + "grad_norm": 6.021454811096191, + "learning_rate": 5.4067166230907365e-05, + "loss": 1.8593, + "step": 2290 + }, + { + "epoch": 0.9822383907554034, + "grad_norm": 6.139098644256592, + "learning_rate": 5.389445398961639e-05, + "loss": 1.8337, + "step": 2295 + }, + { + "epoch": 0.9843783436764392, + "grad_norm": 5.165621280670166, + "learning_rate": 5.3721694984023194e-05, + "loss": 1.899, + "step": 2300 + }, + { + "epoch": 0.9865182965974748, + "grad_norm": 6.382976055145264, + "learning_rate": 5.354889128860454e-05, + "loss": 1.8559, + "step": 2305 + }, + { + "epoch": 0.9886582495185106, + "grad_norm": 6.370285511016846, + "learning_rate": 5.337604497837383e-05, + "loss": 1.7831, + "step": 2310 + }, + { + "epoch": 0.9907982024395463, + "grad_norm": 5.373252868652344, + "learning_rate": 5.320315812885618e-05, + "loss": 1.8073, + "step": 2315 + }, + { + "epoch": 0.992938155360582, + "grad_norm": 6.694858551025391, + "learning_rate": 5.3030232816063505e-05, + "loss": 1.7992, + "step": 2320 + }, + { + "epoch": 0.9950781082816178, + "grad_norm": 6.6283440589904785, + "learning_rate": 5.28572711164696e-05, + "loss": 1.8422, + "step": 2325 + }, + { + "epoch": 0.9972180612026535, + "grad_norm": 6.046857833862305, + "learning_rate": 5.268427510698517e-05, + "loss": 1.8237, + "step": 2330 + }, + { + "epoch": 0.9993580141236893, + "grad_norm": 6.323689937591553, + "learning_rate": 5.251124686493291e-05, + "loss": 1.8662, + "step": 2335 + }, + { + "epoch": 1.0012839717526214, + "grad_norm": 6.170103549957275, + "learning_rate": 5.233818846802255e-05, + "loss": 1.7427, + "step": 2340 + }, + { + "epoch": 1.0034239246736572, + "grad_norm": 5.78736686706543, + "learning_rate": 5.216510199432596e-05, + "loss": 1.83, + "step": 2345 + }, + { + "epoch": 1.005563877594693, + "grad_norm": 5.895803451538086, + "learning_rate": 5.199198952225212e-05, + "loss": 1.8002, + "step": 2350 + }, + { + "epoch": 1.0077038305157286, + "grad_norm": 6.435295581817627, + "learning_rate": 5.1818853130522184e-05, + "loss": 1.7659, + "step": 2355 + }, + { + "epoch": 1.0098437834367644, + "grad_norm": 6.010552406311035, + "learning_rate": 5.164569489814456e-05, + "loss": 1.7948, + "step": 2360 + }, + { + "epoch": 1.0119837363578001, + "grad_norm": 6.7105255126953125, + "learning_rate": 5.147251690438992e-05, + "loss": 1.797, + "step": 2365 + }, + { + "epoch": 1.014123689278836, + "grad_norm": 6.607806205749512, + "learning_rate": 5.1299321228766194e-05, + "loss": 1.7717, + "step": 2370 + }, + { + "epoch": 1.0162636421998716, + "grad_norm": 8.161765098571777, + "learning_rate": 5.112610995099368e-05, + "loss": 1.7855, + "step": 2375 + }, + { + "epoch": 1.0184035951209074, + "grad_norm": 6.3959150314331055, + "learning_rate": 5.095288515097999e-05, + "loss": 1.8212, + "step": 2380 + }, + { + "epoch": 1.0205435480419431, + "grad_norm": 5.896310806274414, + "learning_rate": 5.0779648908795116e-05, + "loss": 1.7695, + "step": 2385 + }, + { + "epoch": 1.0226835009629789, + "grad_norm": 5.285399913787842, + "learning_rate": 5.060640330464646e-05, + "loss": 1.7961, + "step": 2390 + }, + { + "epoch": 1.0248234538840146, + "grad_norm": 6.583270072937012, + "learning_rate": 5.043315041885383e-05, + "loss": 1.8068, + "step": 2395 + }, + { + "epoch": 1.0269634068050504, + "grad_norm": 6.104979991912842, + "learning_rate": 5.0259892331824474e-05, + "loss": 1.7806, + "step": 2400 + }, + { + "epoch": 1.0291033597260861, + "grad_norm": 5.723150730133057, + "learning_rate": 5.008663112402811e-05, + "loss": 1.7961, + "step": 2405 + }, + { + "epoch": 1.0312433126471217, + "grad_norm": 5.977392673492432, + "learning_rate": 4.991336887597192e-05, + "loss": 1.7362, + "step": 2410 + }, + { + "epoch": 1.0333832655681574, + "grad_norm": 6.383233070373535, + "learning_rate": 4.974010766817555e-05, + "loss": 1.7895, + "step": 2415 + }, + { + "epoch": 1.0355232184891932, + "grad_norm": 6.362678527832031, + "learning_rate": 4.95668495811462e-05, + "loss": 1.7526, + "step": 2420 + }, + { + "epoch": 1.037663171410229, + "grad_norm": 5.289202690124512, + "learning_rate": 4.939359669535357e-05, + "loss": 1.8182, + "step": 2425 + }, + { + "epoch": 1.0398031243312646, + "grad_norm": 6.294672012329102, + "learning_rate": 4.922035109120491e-05, + "loss": 1.7909, + "step": 2430 + }, + { + "epoch": 1.0419430772523004, + "grad_norm": 6.827216625213623, + "learning_rate": 4.904711484902003e-05, + "loss": 1.8041, + "step": 2435 + }, + { + "epoch": 1.0440830301733361, + "grad_norm": 6.308746337890625, + "learning_rate": 4.887389004900633e-05, + "loss": 1.8017, + "step": 2440 + }, + { + "epoch": 1.046222983094372, + "grad_norm": 5.453704357147217, + "learning_rate": 4.870067877123382e-05, + "loss": 1.791, + "step": 2445 + }, + { + "epoch": 1.0483629360154076, + "grad_norm": 5.51895809173584, + "learning_rate": 4.852748309561009e-05, + "loss": 1.7959, + "step": 2450 + }, + { + "epoch": 1.0505028889364434, + "grad_norm": 5.434797286987305, + "learning_rate": 4.835430510185545e-05, + "loss": 1.8243, + "step": 2455 + }, + { + "epoch": 1.0526428418574791, + "grad_norm": 5.3560590744018555, + "learning_rate": 4.818114686947783e-05, + "loss": 1.8104, + "step": 2460 + }, + { + "epoch": 1.0547827947785149, + "grad_norm": 6.535407543182373, + "learning_rate": 4.800801047774789e-05, + "loss": 1.7754, + "step": 2465 + }, + { + "epoch": 1.0569227476995506, + "grad_norm": 5.90179967880249, + "learning_rate": 4.783489800567405e-05, + "loss": 1.8041, + "step": 2470 + }, + { + "epoch": 1.0590627006205864, + "grad_norm": 6.035419464111328, + "learning_rate": 4.766181153197746e-05, + "loss": 1.8154, + "step": 2475 + }, + { + "epoch": 1.0612026535416221, + "grad_norm": 5.992060661315918, + "learning_rate": 4.748875313506711e-05, + "loss": 1.8101, + "step": 2480 + }, + { + "epoch": 1.0633426064626579, + "grad_norm": 6.405609130859375, + "learning_rate": 4.7315724893014846e-05, + "loss": 1.7881, + "step": 2485 + }, + { + "epoch": 1.0654825593836936, + "grad_norm": 5.741700649261475, + "learning_rate": 4.714272888353041e-05, + "loss": 1.7986, + "step": 2490 + }, + { + "epoch": 1.0676225123047294, + "grad_norm": 5.35413122177124, + "learning_rate": 4.69697671839365e-05, + "loss": 1.7832, + "step": 2495 + }, + { + "epoch": 1.0697624652257651, + "grad_norm": 5.857509613037109, + "learning_rate": 4.679684187114384e-05, + "loss": 1.7965, + "step": 2500 + }, + { + "epoch": 1.0719024181468009, + "grad_norm": 6.7070698738098145, + "learning_rate": 4.6623955021626184e-05, + "loss": 1.7924, + "step": 2505 + }, + { + "epoch": 1.0740423710678364, + "grad_norm": 6.540157318115234, + "learning_rate": 4.645110871139547e-05, + "loss": 1.7951, + "step": 2510 + }, + { + "epoch": 1.0761823239888721, + "grad_norm": 6.704575538635254, + "learning_rate": 4.627830501597681e-05, + "loss": 1.7831, + "step": 2515 + }, + { + "epoch": 1.078322276909908, + "grad_norm": 6.622875213623047, + "learning_rate": 4.610554601038361e-05, + "loss": 1.7869, + "step": 2520 + }, + { + "epoch": 1.0804622298309436, + "grad_norm": 5.487006187438965, + "learning_rate": 4.593283376909264e-05, + "loss": 1.7893, + "step": 2525 + }, + { + "epoch": 1.0826021827519794, + "grad_norm": 6.443029403686523, + "learning_rate": 4.576017036601914e-05, + "loss": 1.7392, + "step": 2530 + }, + { + "epoch": 1.0847421356730151, + "grad_norm": 5.588938236236572, + "learning_rate": 4.558755787449189e-05, + "loss": 1.7468, + "step": 2535 + }, + { + "epoch": 1.0868820885940509, + "grad_norm": 5.544746398925781, + "learning_rate": 4.541499836722835e-05, + "loss": 1.7849, + "step": 2540 + }, + { + "epoch": 1.0890220415150866, + "grad_norm": 5.714095592498779, + "learning_rate": 4.5242493916309705e-05, + "loss": 1.7508, + "step": 2545 + }, + { + "epoch": 1.0911619944361224, + "grad_norm": 5.488661289215088, + "learning_rate": 4.507004659315611e-05, + "loss": 1.8118, + "step": 2550 + }, + { + "epoch": 1.0933019473571581, + "grad_norm": 6.51017951965332, + "learning_rate": 4.489765846850162e-05, + "loss": 1.7632, + "step": 2555 + }, + { + "epoch": 1.0954419002781939, + "grad_norm": 5.363729476928711, + "learning_rate": 4.472533161236954e-05, + "loss": 1.7984, + "step": 2560 + }, + { + "epoch": 1.0975818531992296, + "grad_norm": 5.888598918914795, + "learning_rate": 4.4553068094047394e-05, + "loss": 1.7983, + "step": 2565 + }, + { + "epoch": 1.0997218061202654, + "grad_norm": 6.100613594055176, + "learning_rate": 4.438086998206221e-05, + "loss": 1.7707, + "step": 2570 + }, + { + "epoch": 1.1018617590413011, + "grad_norm": 5.885534286499023, + "learning_rate": 4.4208739344155594e-05, + "loss": 1.8299, + "step": 2575 + }, + { + "epoch": 1.1040017119623369, + "grad_norm": 6.399609088897705, + "learning_rate": 4.4036678247258924e-05, + "loss": 1.7365, + "step": 2580 + }, + { + "epoch": 1.1061416648833726, + "grad_norm": 7.783679008483887, + "learning_rate": 4.386468875746854e-05, + "loss": 1.8312, + "step": 2585 + }, + { + "epoch": 1.1082816178044084, + "grad_norm": 5.783539772033691, + "learning_rate": 4.369277294002093e-05, + "loss": 1.7481, + "step": 2590 + }, + { + "epoch": 1.1104215707254441, + "grad_norm": 5.326051235198975, + "learning_rate": 4.352093285926793e-05, + "loss": 1.7541, + "step": 2595 + }, + { + "epoch": 1.1125615236464799, + "grad_norm": 5.783180236816406, + "learning_rate": 4.334917057865194e-05, + "loss": 1.7801, + "step": 2600 + }, + { + "epoch": 1.1147014765675156, + "grad_norm": 5.667355537414551, + "learning_rate": 4.3177488160681146e-05, + "loss": 1.7967, + "step": 2605 + }, + { + "epoch": 1.1168414294885514, + "grad_norm": 5.82685661315918, + "learning_rate": 4.300588766690473e-05, + "loss": 1.7543, + "step": 2610 + }, + { + "epoch": 1.1189813824095869, + "grad_norm": 6.610193252563477, + "learning_rate": 4.283437115788814e-05, + "loss": 1.7785, + "step": 2615 + }, + { + "epoch": 1.1211213353306226, + "grad_norm": 7.57377815246582, + "learning_rate": 4.266294069318837e-05, + "loss": 1.7679, + "step": 2620 + }, + { + "epoch": 1.1232612882516584, + "grad_norm": 6.202375411987305, + "learning_rate": 4.2491598331329154e-05, + "loss": 1.8045, + "step": 2625 + }, + { + "epoch": 1.1254012411726941, + "grad_norm": 6.378594875335693, + "learning_rate": 4.232034612977634e-05, + "loss": 1.7454, + "step": 2630 + }, + { + "epoch": 1.1275411940937299, + "grad_norm": 5.158632755279541, + "learning_rate": 4.21491861449131e-05, + "loss": 1.7993, + "step": 2635 + }, + { + "epoch": 1.1296811470147656, + "grad_norm": 6.776663780212402, + "learning_rate": 4.19781204320153e-05, + "loss": 1.7253, + "step": 2640 + }, + { + "epoch": 1.1318210999358014, + "grad_norm": 5.978639125823975, + "learning_rate": 4.180715104522679e-05, + "loss": 1.7127, + "step": 2645 + }, + { + "epoch": 1.1339610528568371, + "grad_norm": 5.210892200469971, + "learning_rate": 4.1636280037534725e-05, + "loss": 1.7764, + "step": 2650 + }, + { + "epoch": 1.1361010057778729, + "grad_norm": 6.2656121253967285, + "learning_rate": 4.1465509460744963e-05, + "loss": 1.7425, + "step": 2655 + }, + { + "epoch": 1.1382409586989086, + "grad_norm": 5.737490177154541, + "learning_rate": 4.129484136545737e-05, + "loss": 1.7626, + "step": 2660 + }, + { + "epoch": 1.1403809116199444, + "grad_norm": 5.1411261558532715, + "learning_rate": 4.112427780104124e-05, + "loss": 1.7971, + "step": 2665 + }, + { + "epoch": 1.1425208645409801, + "grad_norm": 5.617253303527832, + "learning_rate": 4.0953820815610636e-05, + "loss": 1.7431, + "step": 2670 + }, + { + "epoch": 1.1446608174620159, + "grad_norm": 6.250400543212891, + "learning_rate": 4.0783472455999886e-05, + "loss": 1.7756, + "step": 2675 + }, + { + "epoch": 1.1468007703830516, + "grad_norm": 7.598665714263916, + "learning_rate": 4.06132347677389e-05, + "loss": 1.7462, + "step": 2680 + }, + { + "epoch": 1.1489407233040874, + "grad_norm": 5.751972675323486, + "learning_rate": 4.0443109795028665e-05, + "loss": 1.7599, + "step": 2685 + }, + { + "epoch": 1.1510806762251231, + "grad_norm": 6.453872203826904, + "learning_rate": 4.0273099580716725e-05, + "loss": 1.7637, + "step": 2690 + }, + { + "epoch": 1.1532206291461589, + "grad_norm": 6.619513034820557, + "learning_rate": 4.010320616627252e-05, + "loss": 1.7396, + "step": 2695 + }, + { + "epoch": 1.1553605820671946, + "grad_norm": 6.701980113983154, + "learning_rate": 3.993343159176307e-05, + "loss": 1.6959, + "step": 2700 + }, + { + "epoch": 1.1575005349882304, + "grad_norm": 5.974419116973877, + "learning_rate": 3.976377789582832e-05, + "loss": 1.8088, + "step": 2705 + }, + { + "epoch": 1.1596404879092659, + "grad_norm": 5.028608798980713, + "learning_rate": 3.959424711565672e-05, + "loss": 1.7439, + "step": 2710 + }, + { + "epoch": 1.1617804408303019, + "grad_norm": 5.743880748748779, + "learning_rate": 3.9424841286960784e-05, + "loss": 1.748, + "step": 2715 + }, + { + "epoch": 1.1639203937513374, + "grad_norm": 5.568974018096924, + "learning_rate": 3.925556244395259e-05, + "loss": 1.7883, + "step": 2720 + }, + { + "epoch": 1.1660603466723731, + "grad_norm": 5.259049892425537, + "learning_rate": 3.908641261931937e-05, + "loss": 1.7304, + "step": 2725 + }, + { + "epoch": 1.1682002995934089, + "grad_norm": 5.7519850730896, + "learning_rate": 3.8917393844199156e-05, + "loss": 1.7241, + "step": 2730 + }, + { + "epoch": 1.1703402525144446, + "grad_norm": 5.750739097595215, + "learning_rate": 3.874850814815632e-05, + "loss": 1.6955, + "step": 2735 + }, + { + "epoch": 1.1724802054354804, + "grad_norm": 6.248133182525635, + "learning_rate": 3.857975755915723e-05, + "loss": 1.7689, + "step": 2740 + }, + { + "epoch": 1.1746201583565161, + "grad_norm": 5.397186756134033, + "learning_rate": 3.8411144103545904e-05, + "loss": 1.7588, + "step": 2745 + }, + { + "epoch": 1.1767601112775519, + "grad_norm": 5.462392330169678, + "learning_rate": 3.824266980601968e-05, + "loss": 1.7879, + "step": 2750 + }, + { + "epoch": 1.1789000641985876, + "grad_norm": 6.177926063537598, + "learning_rate": 3.807433668960488e-05, + "loss": 1.7516, + "step": 2755 + }, + { + "epoch": 1.1810400171196234, + "grad_norm": 5.820185661315918, + "learning_rate": 3.7906146775632554e-05, + "loss": 1.7227, + "step": 2760 + }, + { + "epoch": 1.1831799700406591, + "grad_norm": 5.526188850402832, + "learning_rate": 3.773810208371417e-05, + "loss": 1.7781, + "step": 2765 + }, + { + "epoch": 1.1853199229616949, + "grad_norm": 5.152137756347656, + "learning_rate": 3.7570204631717395e-05, + "loss": 1.775, + "step": 2770 + }, + { + "epoch": 1.1874598758827306, + "grad_norm": 6.030661106109619, + "learning_rate": 3.740245643574184e-05, + "loss": 1.8056, + "step": 2775 + }, + { + "epoch": 1.1895998288037664, + "grad_norm": 5.847061634063721, + "learning_rate": 3.723485951009486e-05, + "loss": 1.7345, + "step": 2780 + }, + { + "epoch": 1.191739781724802, + "grad_norm": 5.8102126121521, + "learning_rate": 3.706741586726738e-05, + "loss": 1.7519, + "step": 2785 + }, + { + "epoch": 1.1938797346458379, + "grad_norm": 6.639953136444092, + "learning_rate": 3.690012751790972e-05, + "loss": 1.7485, + "step": 2790 + }, + { + "epoch": 1.1960196875668736, + "grad_norm": 5.996955871582031, + "learning_rate": 3.673299647080742e-05, + "loss": 1.7072, + "step": 2795 + }, + { + "epoch": 1.1981596404879093, + "grad_norm": 5.289943218231201, + "learning_rate": 3.656602473285717e-05, + "loss": 1.7831, + "step": 2800 + }, + { + "epoch": 1.200299593408945, + "grad_norm": 5.406320571899414, + "learning_rate": 3.639921430904268e-05, + "loss": 1.7397, + "step": 2805 + }, + { + "epoch": 1.2024395463299808, + "grad_norm": 6.86167573928833, + "learning_rate": 3.6232567202410624e-05, + "loss": 1.7408, + "step": 2810 + }, + { + "epoch": 1.2045794992510164, + "grad_norm": 5.778091907501221, + "learning_rate": 3.606608541404656e-05, + "loss": 1.765, + "step": 2815 + }, + { + "epoch": 1.2067194521720521, + "grad_norm": 6.24983024597168, + "learning_rate": 3.5899770943050924e-05, + "loss": 1.7543, + "step": 2820 + }, + { + "epoch": 1.2088594050930879, + "grad_norm": 6.323977470397949, + "learning_rate": 3.573362578651501e-05, + "loss": 1.7168, + "step": 2825 + }, + { + "epoch": 1.2109993580141236, + "grad_norm": 6.535098075866699, + "learning_rate": 3.556765193949704e-05, + "loss": 1.7583, + "step": 2830 + }, + { + "epoch": 1.2131393109351594, + "grad_norm": 5.957965850830078, + "learning_rate": 3.5401851394998084e-05, + "loss": 1.7272, + "step": 2835 + }, + { + "epoch": 1.2152792638561951, + "grad_norm": 5.54447078704834, + "learning_rate": 3.523622614393827e-05, + "loss": 1.8009, + "step": 2840 + }, + { + "epoch": 1.2174192167772309, + "grad_norm": 5.613691329956055, + "learning_rate": 3.5070778175132806e-05, + "loss": 1.7559, + "step": 2845 + }, + { + "epoch": 1.2195591696982666, + "grad_norm": 5.480210781097412, + "learning_rate": 3.4905509475268104e-05, + "loss": 1.74, + "step": 2850 + }, + { + "epoch": 1.2216991226193024, + "grad_norm": 6.216781139373779, + "learning_rate": 3.474042202887792e-05, + "loss": 1.7272, + "step": 2855 + }, + { + "epoch": 1.223839075540338, + "grad_norm": 6.038024425506592, + "learning_rate": 3.4575517818319534e-05, + "loss": 1.7408, + "step": 2860 + }, + { + "epoch": 1.2259790284613739, + "grad_norm": 5.762340068817139, + "learning_rate": 3.4410798823749964e-05, + "loss": 1.7576, + "step": 2865 + }, + { + "epoch": 1.2281189813824096, + "grad_norm": 6.565494060516357, + "learning_rate": 3.424626702310214e-05, + "loss": 1.7008, + "step": 2870 + }, + { + "epoch": 1.2302589343034454, + "grad_norm": 6.1579508781433105, + "learning_rate": 3.408192439206121e-05, + "loss": 1.7218, + "step": 2875 + }, + { + "epoch": 1.232398887224481, + "grad_norm": 5.965446949005127, + "learning_rate": 3.391777290404077e-05, + "loss": 1.6989, + "step": 2880 + }, + { + "epoch": 1.2345388401455168, + "grad_norm": 6.2155070304870605, + "learning_rate": 3.3753814530159176e-05, + "loss": 1.7461, + "step": 2885 + }, + { + "epoch": 1.2366787930665526, + "grad_norm": 6.247200965881348, + "learning_rate": 3.3590051239215916e-05, + "loss": 1.7542, + "step": 2890 + }, + { + "epoch": 1.2388187459875883, + "grad_norm": 4.975124359130859, + "learning_rate": 3.342648499766791e-05, + "loss": 1.7299, + "step": 2895 + }, + { + "epoch": 1.240958698908624, + "grad_norm": 5.148104190826416, + "learning_rate": 3.326311776960593e-05, + "loss": 1.7252, + "step": 2900 + }, + { + "epoch": 1.2430986518296598, + "grad_norm": 6.518229961395264, + "learning_rate": 3.309995151673103e-05, + "loss": 1.754, + "step": 2905 + }, + { + "epoch": 1.2452386047506954, + "grad_norm": 5.220431804656982, + "learning_rate": 3.293698819833093e-05, + "loss": 1.7349, + "step": 2910 + }, + { + "epoch": 1.2473785576717313, + "grad_norm": 5.202417373657227, + "learning_rate": 3.277422977125656e-05, + "loss": 1.7521, + "step": 2915 + }, + { + "epoch": 1.2495185105927669, + "grad_norm": 5.129791736602783, + "learning_rate": 3.2611678189898523e-05, + "loss": 1.7653, + "step": 2920 + }, + { + "epoch": 1.2516584635138028, + "grad_norm": 6.944059371948242, + "learning_rate": 3.244933540616363e-05, + "loss": 1.7411, + "step": 2925 + }, + { + "epoch": 1.2537984164348384, + "grad_norm": 5.993988513946533, + "learning_rate": 3.228720336945148e-05, + "loss": 1.7544, + "step": 2930 + }, + { + "epoch": 1.255938369355874, + "grad_norm": 5.735143184661865, + "learning_rate": 3.2125284026631006e-05, + "loss": 1.7374, + "step": 2935 + }, + { + "epoch": 1.2580783222769099, + "grad_norm": 5.272975444793701, + "learning_rate": 3.196357932201717e-05, + "loss": 1.7362, + "step": 2940 + }, + { + "epoch": 1.2602182751979456, + "grad_norm": 6.259987831115723, + "learning_rate": 3.1802091197347544e-05, + "loss": 1.745, + "step": 2945 + }, + { + "epoch": 1.2623582281189814, + "grad_norm": 5.6015400886535645, + "learning_rate": 3.164082159175904e-05, + "loss": 1.7478, + "step": 2950 + }, + { + "epoch": 1.264498181040017, + "grad_norm": 6.1617536544799805, + "learning_rate": 3.147977244176461e-05, + "loss": 1.757, + "step": 2955 + }, + { + "epoch": 1.2666381339610528, + "grad_norm": 4.88234806060791, + "learning_rate": 3.131894568122999e-05, + "loss": 1.73, + "step": 2960 + }, + { + "epoch": 1.2687780868820886, + "grad_norm": 5.751287460327148, + "learning_rate": 3.115834324135047e-05, + "loss": 1.7807, + "step": 2965 + }, + { + "epoch": 1.2709180398031243, + "grad_norm": 5.901682376861572, + "learning_rate": 3.099796705062773e-05, + "loss": 1.7535, + "step": 2970 + }, + { + "epoch": 1.27305799272416, + "grad_norm": 5.5405144691467285, + "learning_rate": 3.083781903484667e-05, + "loss": 1.7742, + "step": 2975 + }, + { + "epoch": 1.2751979456451958, + "grad_norm": 5.621067523956299, + "learning_rate": 3.067790111705225e-05, + "loss": 1.6996, + "step": 2980 + }, + { + "epoch": 1.2773378985662316, + "grad_norm": 5.193610191345215, + "learning_rate": 3.051821521752647e-05, + "loss": 1.7568, + "step": 2985 + }, + { + "epoch": 1.2794778514872673, + "grad_norm": 5.218043804168701, + "learning_rate": 3.0358763253765264e-05, + "loss": 1.7299, + "step": 2990 + }, + { + "epoch": 1.281617804408303, + "grad_norm": 6.297114849090576, + "learning_rate": 3.0199547140455474e-05, + "loss": 1.7814, + "step": 2995 + }, + { + "epoch": 1.2837577573293388, + "grad_norm": 6.044188976287842, + "learning_rate": 3.0040568789451862e-05, + "loss": 1.7381, + "step": 3000 + }, + { + "epoch": 1.2858977102503744, + "grad_norm": 5.990639686584473, + "learning_rate": 2.9881830109754184e-05, + "loss": 1.7458, + "step": 3005 + }, + { + "epoch": 1.2880376631714103, + "grad_norm": 5.563923358917236, + "learning_rate": 2.9723333007484218e-05, + "loss": 1.7403, + "step": 3010 + }, + { + "epoch": 1.2901776160924459, + "grad_norm": 5.493687152862549, + "learning_rate": 2.9565079385862903e-05, + "loss": 1.7141, + "step": 3015 + }, + { + "epoch": 1.2923175690134818, + "grad_norm": 5.307217121124268, + "learning_rate": 2.9407071145187494e-05, + "loss": 1.7546, + "step": 3020 + }, + { + "epoch": 1.2944575219345174, + "grad_norm": 5.565853118896484, + "learning_rate": 2.924931018280871e-05, + "loss": 1.7392, + "step": 3025 + }, + { + "epoch": 1.296597474855553, + "grad_norm": 5.918313503265381, + "learning_rate": 2.9091798393107994e-05, + "loss": 1.7598, + "step": 3030 + }, + { + "epoch": 1.2987374277765888, + "grad_norm": 5.450323104858398, + "learning_rate": 2.8934537667474732e-05, + "loss": 1.7204, + "step": 3035 + }, + { + "epoch": 1.3008773806976246, + "grad_norm": 5.59390115737915, + "learning_rate": 2.8777529894283538e-05, + "loss": 1.7345, + "step": 3040 + }, + { + "epoch": 1.3030173336186603, + "grad_norm": 5.590770721435547, + "learning_rate": 2.8620776958871627e-05, + "loss": 1.7814, + "step": 3045 + }, + { + "epoch": 1.305157286539696, + "grad_norm": 5.129935264587402, + "learning_rate": 2.8464280743516102e-05, + "loss": 1.7527, + "step": 3050 + }, + { + "epoch": 1.3072972394607318, + "grad_norm": 6.99540376663208, + "learning_rate": 2.8308043127411423e-05, + "loss": 1.7197, + "step": 3055 + }, + { + "epoch": 1.3094371923817676, + "grad_norm": 6.13936710357666, + "learning_rate": 2.8152065986646788e-05, + "loss": 1.7153, + "step": 3060 + }, + { + "epoch": 1.3115771453028033, + "grad_norm": 5.489097595214844, + "learning_rate": 2.7996351194183646e-05, + "loss": 1.7562, + "step": 3065 + }, + { + "epoch": 1.313717098223839, + "grad_norm": 6.8664398193359375, + "learning_rate": 2.784090061983317e-05, + "loss": 1.7034, + "step": 3070 + }, + { + "epoch": 1.3158570511448748, + "grad_norm": 5.367057800292969, + "learning_rate": 2.7685716130233842e-05, + "loss": 1.6691, + "step": 3075 + }, + { + "epoch": 1.3179970040659106, + "grad_norm": 6.975956916809082, + "learning_rate": 2.7530799588829005e-05, + "loss": 1.7107, + "step": 3080 + }, + { + "epoch": 1.3201369569869463, + "grad_norm": 6.1998443603515625, + "learning_rate": 2.7376152855844495e-05, + "loss": 1.7955, + "step": 3085 + }, + { + "epoch": 1.322276909907982, + "grad_norm": 6.424160957336426, + "learning_rate": 2.7221777788266324e-05, + "loss": 1.7312, + "step": 3090 + }, + { + "epoch": 1.3244168628290178, + "grad_norm": 5.118837356567383, + "learning_rate": 2.7067676239818364e-05, + "loss": 1.7105, + "step": 3095 + }, + { + "epoch": 1.3265568157500536, + "grad_norm": 5.895681858062744, + "learning_rate": 2.6913850060940083e-05, + "loss": 1.7077, + "step": 3100 + }, + { + "epoch": 1.3286967686710893, + "grad_norm": 6.501992702484131, + "learning_rate": 2.676030109876434e-05, + "loss": 1.728, + "step": 3105 + }, + { + "epoch": 1.3308367215921248, + "grad_norm": 5.911489486694336, + "learning_rate": 2.66070311970952e-05, + "loss": 1.7214, + "step": 3110 + }, + { + "epoch": 1.3329766745131608, + "grad_norm": 5.91454553604126, + "learning_rate": 2.6454042196385798e-05, + "loss": 1.6895, + "step": 3115 + }, + { + "epoch": 1.3351166274341963, + "grad_norm": 5.649311065673828, + "learning_rate": 2.6301335933716176e-05, + "loss": 1.7488, + "step": 3120 + }, + { + "epoch": 1.3372565803552323, + "grad_norm": 7.105709075927734, + "learning_rate": 2.614891424277135e-05, + "loss": 1.6687, + "step": 3125 + }, + { + "epoch": 1.3393965332762678, + "grad_norm": 5.769228458404541, + "learning_rate": 2.599677895381919e-05, + "loss": 1.7175, + "step": 3130 + }, + { + "epoch": 1.3415364861973036, + "grad_norm": 7.222846984863281, + "learning_rate": 2.5844931893688473e-05, + "loss": 1.721, + "step": 3135 + }, + { + "epoch": 1.3436764391183393, + "grad_norm": 5.122998237609863, + "learning_rate": 2.5693374885746957e-05, + "loss": 1.7325, + "step": 3140 + }, + { + "epoch": 1.345816392039375, + "grad_norm": 5.9646148681640625, + "learning_rate": 2.5542109749879473e-05, + "loss": 1.7016, + "step": 3145 + }, + { + "epoch": 1.3479563449604108, + "grad_norm": 5.405865669250488, + "learning_rate": 2.5391138302466062e-05, + "loss": 1.7666, + "step": 3150 + }, + { + "epoch": 1.3500962978814466, + "grad_norm": 6.047122478485107, + "learning_rate": 2.5240462356360184e-05, + "loss": 1.7438, + "step": 3155 + }, + { + "epoch": 1.3522362508024823, + "grad_norm": 6.545495986938477, + "learning_rate": 2.5090083720866952e-05, + "loss": 1.7322, + "step": 3160 + }, + { + "epoch": 1.354376203723518, + "grad_norm": 6.075828552246094, + "learning_rate": 2.4940004201721384e-05, + "loss": 1.7089, + "step": 3165 + }, + { + "epoch": 1.3565161566445538, + "grad_norm": 5.580131530761719, + "learning_rate": 2.479022560106673e-05, + "loss": 1.7456, + "step": 3170 + }, + { + "epoch": 1.3586561095655896, + "grad_norm": 5.123158931732178, + "learning_rate": 2.4640749717432854e-05, + "loss": 1.7118, + "step": 3175 + }, + { + "epoch": 1.3607960624866253, + "grad_norm": 5.250161170959473, + "learning_rate": 2.4491578345714587e-05, + "loss": 1.7271, + "step": 3180 + }, + { + "epoch": 1.362936015407661, + "grad_norm": 5.588467121124268, + "learning_rate": 2.434271327715023e-05, + "loss": 1.7026, + "step": 3185 + }, + { + "epoch": 1.3650759683286968, + "grad_norm": 5.4725518226623535, + "learning_rate": 2.419415629930001e-05, + "loss": 1.7128, + "step": 3190 + }, + { + "epoch": 1.3672159212497326, + "grad_norm": 5.664406776428223, + "learning_rate": 2.4045909196024624e-05, + "loss": 1.7244, + "step": 3195 + }, + { + "epoch": 1.3693558741707683, + "grad_norm": 4.935891151428223, + "learning_rate": 2.389797374746382e-05, + "loss": 1.7123, + "step": 3200 + }, + { + "epoch": 1.3714958270918038, + "grad_norm": 5.0147881507873535, + "learning_rate": 2.3750351730015015e-05, + "loss": 1.6921, + "step": 3205 + }, + { + "epoch": 1.3736357800128398, + "grad_norm": 5.382151126861572, + "learning_rate": 2.3603044916311963e-05, + "loss": 1.78, + "step": 3210 + }, + { + "epoch": 1.3757757329338753, + "grad_norm": 5.000919818878174, + "learning_rate": 2.345605507520352e-05, + "loss": 1.699, + "step": 3215 + }, + { + "epoch": 1.3779156858549113, + "grad_norm": 4.845434665679932, + "learning_rate": 2.3309383971732312e-05, + "loss": 1.7508, + "step": 3220 + }, + { + "epoch": 1.3800556387759468, + "grad_norm": 4.90060567855835, + "learning_rate": 2.3163033367113602e-05, + "loss": 1.6904, + "step": 3225 + }, + { + "epoch": 1.3821955916969826, + "grad_norm": 5.377330303192139, + "learning_rate": 2.3017005018714126e-05, + "loss": 1.7236, + "step": 3230 + }, + { + "epoch": 1.3843355446180183, + "grad_norm": 5.329347133636475, + "learning_rate": 2.2871300680030995e-05, + "loss": 1.6982, + "step": 3235 + }, + { + "epoch": 1.386475497539054, + "grad_norm": 4.9429931640625, + "learning_rate": 2.2725922100670644e-05, + "loss": 1.6973, + "step": 3240 + }, + { + "epoch": 1.3886154504600898, + "grad_norm": 5.456453323364258, + "learning_rate": 2.25808710263278e-05, + "loss": 1.7033, + "step": 3245 + }, + { + "epoch": 1.3907554033811256, + "grad_norm": 5.3230791091918945, + "learning_rate": 2.243614919876454e-05, + "loss": 1.7447, + "step": 3250 + }, + { + "epoch": 1.3928953563021613, + "grad_norm": 5.888358116149902, + "learning_rate": 2.22917583557894e-05, + "loss": 1.7347, + "step": 3255 + }, + { + "epoch": 1.395035309223197, + "grad_norm": 5.590538024902344, + "learning_rate": 2.214770023123641e-05, + "loss": 1.7566, + "step": 3260 + }, + { + "epoch": 1.3971752621442328, + "grad_norm": 5.970736503601074, + "learning_rate": 2.2003976554944405e-05, + "loss": 1.6941, + "step": 3265 + }, + { + "epoch": 1.3993152150652686, + "grad_norm": 5.476358413696289, + "learning_rate": 2.186058905273618e-05, + "loss": 1.7016, + "step": 3270 + }, + { + "epoch": 1.4014551679863043, + "grad_norm": 5.488466262817383, + "learning_rate": 2.1717539446397754e-05, + "loss": 1.7063, + "step": 3275 + }, + { + "epoch": 1.40359512090734, + "grad_norm": 5.273731708526611, + "learning_rate": 2.1574829453657746e-05, + "loss": 1.7218, + "step": 3280 + }, + { + "epoch": 1.4057350738283758, + "grad_norm": 5.290144920349121, + "learning_rate": 2.1432460788166704e-05, + "loss": 1.6931, + "step": 3285 + }, + { + "epoch": 1.4078750267494116, + "grad_norm": 6.16803503036499, + "learning_rate": 2.1290435159476545e-05, + "loss": 1.6872, + "step": 3290 + }, + { + "epoch": 1.4100149796704473, + "grad_norm": 5.351411819458008, + "learning_rate": 2.1148754273020038e-05, + "loss": 1.678, + "step": 3295 + }, + { + "epoch": 1.412154932591483, + "grad_norm": 6.350516319274902, + "learning_rate": 2.1007419830090306e-05, + "loss": 1.728, + "step": 3300 + }, + { + "epoch": 1.4142948855125188, + "grad_norm": 5.602326393127441, + "learning_rate": 2.0866433527820406e-05, + "loss": 1.6969, + "step": 3305 + }, + { + "epoch": 1.4164348384335543, + "grad_norm": 7.531297206878662, + "learning_rate": 2.072579705916295e-05, + "loss": 1.738, + "step": 3310 + }, + { + "epoch": 1.4185747913545903, + "grad_norm": 5.831987380981445, + "learning_rate": 2.058551211286977e-05, + "loss": 1.7518, + "step": 3315 + }, + { + "epoch": 1.4207147442756258, + "grad_norm": 6.611841678619385, + "learning_rate": 2.0445580373471658e-05, + "loss": 1.6811, + "step": 3320 + }, + { + "epoch": 1.4228546971966618, + "grad_norm": 4.91347599029541, + "learning_rate": 2.03060035212581e-05, + "loss": 1.6897, + "step": 3325 + }, + { + "epoch": 1.4249946501176973, + "grad_norm": 6.171419620513916, + "learning_rate": 2.0166783232257154e-05, + "loss": 1.739, + "step": 3330 + }, + { + "epoch": 1.427134603038733, + "grad_norm": 5.1369524002075195, + "learning_rate": 2.0027921178215274e-05, + "loss": 1.7164, + "step": 3335 + }, + { + "epoch": 1.4292745559597688, + "grad_norm": 6.778637409210205, + "learning_rate": 1.9889419026577266e-05, + "loss": 1.7224, + "step": 3340 + }, + { + "epoch": 1.4314145088808046, + "grad_norm": 4.898647785186768, + "learning_rate": 1.9751278440466248e-05, + "loss": 1.6844, + "step": 3345 + }, + { + "epoch": 1.4335544618018403, + "grad_norm": 5.825972080230713, + "learning_rate": 1.9613501078663693e-05, + "loss": 1.7451, + "step": 3350 + }, + { + "epoch": 1.435694414722876, + "grad_norm": 5.397356986999512, + "learning_rate": 1.9476088595589504e-05, + "loss": 1.715, + "step": 3355 + }, + { + "epoch": 1.4378343676439118, + "grad_norm": 5.8002519607543945, + "learning_rate": 1.9339042641282146e-05, + "loss": 1.7006, + "step": 3360 + }, + { + "epoch": 1.4399743205649476, + "grad_norm": 5.10746431350708, + "learning_rate": 1.9202364861378842e-05, + "loss": 1.6758, + "step": 3365 + }, + { + "epoch": 1.4421142734859833, + "grad_norm": 5.267380237579346, + "learning_rate": 1.9066056897095796e-05, + "loss": 1.7395, + "step": 3370 + }, + { + "epoch": 1.444254226407019, + "grad_norm": 5.395501136779785, + "learning_rate": 1.8930120385208495e-05, + "loss": 1.7082, + "step": 3375 + }, + { + "epoch": 1.4463941793280548, + "grad_norm": 5.383699893951416, + "learning_rate": 1.8794556958032062e-05, + "loss": 1.7182, + "step": 3380 + }, + { + "epoch": 1.4485341322490906, + "grad_norm": 5.4255290031433105, + "learning_rate": 1.865936824340164e-05, + "loss": 1.6796, + "step": 3385 + }, + { + "epoch": 1.4506740851701263, + "grad_norm": 5.540951728820801, + "learning_rate": 1.8524555864652865e-05, + "loss": 1.7312, + "step": 3390 + }, + { + "epoch": 1.452814038091162, + "grad_norm": 6.032477855682373, + "learning_rate": 1.839012144060236e-05, + "loss": 1.6458, + "step": 3395 + }, + { + "epoch": 1.4549539910121978, + "grad_norm": 5.043034553527832, + "learning_rate": 1.825606658552826e-05, + "loss": 1.6969, + "step": 3400 + }, + { + "epoch": 1.4570939439332335, + "grad_norm": 5.323840618133545, + "learning_rate": 1.8122392909150904e-05, + "loss": 1.7225, + "step": 3405 + }, + { + "epoch": 1.4592338968542693, + "grad_norm": 5.682520389556885, + "learning_rate": 1.7989102016613463e-05, + "loss": 1.6943, + "step": 3410 + }, + { + "epoch": 1.4613738497753048, + "grad_norm": 5.428680419921875, + "learning_rate": 1.7856195508462663e-05, + "loss": 1.6992, + "step": 3415 + }, + { + "epoch": 1.4635138026963408, + "grad_norm": 5.1605024337768555, + "learning_rate": 1.7723674980629572e-05, + "loss": 1.6569, + "step": 3420 + }, + { + "epoch": 1.4656537556173763, + "grad_norm": 5.788964748382568, + "learning_rate": 1.759154202441044e-05, + "loss": 1.7081, + "step": 3425 + }, + { + "epoch": 1.4677937085384123, + "grad_norm": 5.63525390625, + "learning_rate": 1.7459798226447577e-05, + "loss": 1.7133, + "step": 3430 + }, + { + "epoch": 1.4699336614594478, + "grad_norm": 5.6762003898620605, + "learning_rate": 1.7328445168710323e-05, + "loss": 1.6922, + "step": 3435 + }, + { + "epoch": 1.4720736143804836, + "grad_norm": 5.393453121185303, + "learning_rate": 1.7197484428476023e-05, + "loss": 1.6721, + "step": 3440 + }, + { + "epoch": 1.4742135673015193, + "grad_norm": 4.811567783355713, + "learning_rate": 1.7066917578311132e-05, + "loss": 1.7071, + "step": 3445 + }, + { + "epoch": 1.476353520222555, + "grad_norm": 5.325284004211426, + "learning_rate": 1.693674618605227e-05, + "loss": 1.7109, + "step": 3450 + }, + { + "epoch": 1.4784934731435908, + "grad_norm": 5.2245774269104, + "learning_rate": 1.6806971814787458e-05, + "loss": 1.7103, + "step": 3455 + }, + { + "epoch": 1.4806334260646266, + "grad_norm": 5.08640718460083, + "learning_rate": 1.66775960228373e-05, + "loss": 1.7094, + "step": 3460 + }, + { + "epoch": 1.4827733789856623, + "grad_norm": 4.874035835266113, + "learning_rate": 1.6548620363736294e-05, + "loss": 1.7177, + "step": 3465 + }, + { + "epoch": 1.484913331906698, + "grad_norm": 5.959578037261963, + "learning_rate": 1.6420046386214184e-05, + "loss": 1.6586, + "step": 3470 + }, + { + "epoch": 1.4870532848277338, + "grad_norm": 5.682189464569092, + "learning_rate": 1.6291875634177346e-05, + "loss": 1.6731, + "step": 3475 + }, + { + "epoch": 1.4891932377487695, + "grad_norm": 5.864963054656982, + "learning_rate": 1.616410964669025e-05, + "loss": 1.6708, + "step": 3480 + }, + { + "epoch": 1.4913331906698053, + "grad_norm": 5.3527140617370605, + "learning_rate": 1.6036749957956993e-05, + "loss": 1.6596, + "step": 3485 + }, + { + "epoch": 1.493473143590841, + "grad_norm": 4.715349197387695, + "learning_rate": 1.5909798097302865e-05, + "loss": 1.6889, + "step": 3490 + }, + { + "epoch": 1.4956130965118768, + "grad_norm": 5.459346771240234, + "learning_rate": 1.578325558915598e-05, + "loss": 1.6716, + "step": 3495 + }, + { + "epoch": 1.4977530494329125, + "grad_norm": 5.306508541107178, + "learning_rate": 1.5657123953029003e-05, + "loss": 1.673, + "step": 3500 + }, + { + "epoch": 1.4998930023539483, + "grad_norm": 5.33494234085083, + "learning_rate": 1.5531404703500845e-05, + "loss": 1.7047, + "step": 3505 + }, + { + "epoch": 1.5020329552749838, + "grad_norm": 5.494200229644775, + "learning_rate": 1.5406099350198544e-05, + "loss": 1.7061, + "step": 3510 + }, + { + "epoch": 1.5041729081960198, + "grad_norm": 4.832003116607666, + "learning_rate": 1.528120939777908e-05, + "loss": 1.6623, + "step": 3515 + }, + { + "epoch": 1.5063128611170553, + "grad_norm": 5.2969970703125, + "learning_rate": 1.5156736345911342e-05, + "loss": 1.6838, + "step": 3520 + }, + { + "epoch": 1.5084528140380913, + "grad_norm": 5.263791084289551, + "learning_rate": 1.5032681689258105e-05, + "loss": 1.691, + "step": 3525 + }, + { + "epoch": 1.5105927669591268, + "grad_norm": 5.589069843292236, + "learning_rate": 1.4909046917458097e-05, + "loss": 1.7456, + "step": 3530 + }, + { + "epoch": 1.5127327198801628, + "grad_norm": 5.666341781616211, + "learning_rate": 1.4785833515108088e-05, + "loss": 1.7284, + "step": 3535 + }, + { + "epoch": 1.5148726728011983, + "grad_norm": 5.460354328155518, + "learning_rate": 1.4663042961745083e-05, + "loss": 1.6767, + "step": 3540 + }, + { + "epoch": 1.5170126257222343, + "grad_norm": 5.311362266540527, + "learning_rate": 1.4540676731828546e-05, + "loss": 1.662, + "step": 3545 + }, + { + "epoch": 1.5191525786432698, + "grad_norm": 5.237598419189453, + "learning_rate": 1.4418736294722701e-05, + "loss": 1.6841, + "step": 3550 + }, + { + "epoch": 1.5212925315643056, + "grad_norm": 5.252871513366699, + "learning_rate": 1.4297223114678887e-05, + "loss": 1.7057, + "step": 3555 + }, + { + "epoch": 1.5234324844853413, + "grad_norm": 4.881004333496094, + "learning_rate": 1.4176138650817967e-05, + "loss": 1.699, + "step": 3560 + }, + { + "epoch": 1.525572437406377, + "grad_norm": 5.0006103515625, + "learning_rate": 1.4055484357112808e-05, + "loss": 1.7438, + "step": 3565 + }, + { + "epoch": 1.5277123903274128, + "grad_norm": 4.589049339294434, + "learning_rate": 1.3935261682370849e-05, + "loss": 1.7316, + "step": 3570 + }, + { + "epoch": 1.5298523432484485, + "grad_norm": 5.060586929321289, + "learning_rate": 1.3815472070216656e-05, + "loss": 1.7279, + "step": 3575 + }, + { + "epoch": 1.5319922961694843, + "grad_norm": 5.7035322189331055, + "learning_rate": 1.3696116959074635e-05, + "loss": 1.7232, + "step": 3580 + }, + { + "epoch": 1.53413224909052, + "grad_norm": 5.462157726287842, + "learning_rate": 1.3577197782151724e-05, + "loss": 1.674, + "step": 3585 + }, + { + "epoch": 1.5362722020115558, + "grad_norm": 5.079245567321777, + "learning_rate": 1.3458715967420193e-05, + "loss": 1.7135, + "step": 3590 + }, + { + "epoch": 1.5384121549325915, + "grad_norm": 5.461935043334961, + "learning_rate": 1.3340672937600518e-05, + "loss": 1.6261, + "step": 3595 + }, + { + "epoch": 1.5405521078536273, + "grad_norm": 5.3971076011657715, + "learning_rate": 1.3223070110144265e-05, + "loss": 1.7194, + "step": 3600 + }, + { + "epoch": 1.5426920607746628, + "grad_norm": 5.163456916809082, + "learning_rate": 1.3105908897217084e-05, + "loss": 1.6877, + "step": 3605 + }, + { + "epoch": 1.5448320136956988, + "grad_norm": 6.106560707092285, + "learning_rate": 1.2989190705681758e-05, + "loss": 1.6893, + "step": 3610 + }, + { + "epoch": 1.5469719666167343, + "grad_norm": 5.627121925354004, + "learning_rate": 1.2872916937081308e-05, + "loss": 1.6939, + "step": 3615 + }, + { + "epoch": 1.5491119195377703, + "grad_norm": 5.5462117195129395, + "learning_rate": 1.2757088987622152e-05, + "loss": 1.6919, + "step": 3620 + }, + { + "epoch": 1.5512518724588058, + "grad_norm": 5.333662986755371, + "learning_rate": 1.2641708248157341e-05, + "loss": 1.6444, + "step": 3625 + }, + { + "epoch": 1.5533918253798418, + "grad_norm": 6.698999881744385, + "learning_rate": 1.2526776104169868e-05, + "loss": 1.7085, + "step": 3630 + }, + { + "epoch": 1.5555317783008773, + "grad_norm": 5.591616630554199, + "learning_rate": 1.241229393575603e-05, + "loss": 1.6643, + "step": 3635 + }, + { + "epoch": 1.5576717312219133, + "grad_norm": 5.279683589935303, + "learning_rate": 1.2298263117608855e-05, + "loss": 1.624, + "step": 3640 + }, + { + "epoch": 1.5598116841429488, + "grad_norm": 5.046063423156738, + "learning_rate": 1.2184685019001574e-05, + "loss": 1.6992, + "step": 3645 + }, + { + "epoch": 1.5619516370639845, + "grad_norm": 5.270909786224365, + "learning_rate": 1.2071561003771214e-05, + "loss": 1.6567, + "step": 3650 + }, + { + "epoch": 1.5640915899850203, + "grad_norm": 5.26682186126709, + "learning_rate": 1.1958892430302198e-05, + "loss": 1.7055, + "step": 3655 + }, + { + "epoch": 1.566231542906056, + "grad_norm": 5.099318504333496, + "learning_rate": 1.184668065151005e-05, + "loss": 1.7185, + "step": 3660 + }, + { + "epoch": 1.5683714958270918, + "grad_norm": 5.019136905670166, + "learning_rate": 1.1734927014825115e-05, + "loss": 1.7033, + "step": 3665 + }, + { + "epoch": 1.5705114487481275, + "grad_norm": 5.4806437492370605, + "learning_rate": 1.162363286217642e-05, + "loss": 1.7044, + "step": 3670 + }, + { + "epoch": 1.5726514016691633, + "grad_norm": 4.8851118087768555, + "learning_rate": 1.151279952997556e-05, + "loss": 1.6831, + "step": 3675 + }, + { + "epoch": 1.574791354590199, + "grad_norm": 5.282357692718506, + "learning_rate": 1.1402428349100585e-05, + "loss": 1.6689, + "step": 3680 + }, + { + "epoch": 1.5769313075112348, + "grad_norm": 5.409157752990723, + "learning_rate": 1.1292520644880105e-05, + "loss": 1.6455, + "step": 3685 + }, + { + "epoch": 1.5790712604322705, + "grad_norm": 5.711396217346191, + "learning_rate": 1.1183077737077336e-05, + "loss": 1.6363, + "step": 3690 + }, + { + "epoch": 1.5812112133533063, + "grad_norm": 5.7200541496276855, + "learning_rate": 1.107410093987425e-05, + "loss": 1.6737, + "step": 3695 + }, + { + "epoch": 1.5833511662743418, + "grad_norm": 5.475973606109619, + "learning_rate": 1.0965591561855788e-05, + "loss": 1.6393, + "step": 3700 + }, + { + "epoch": 1.5854911191953778, + "grad_norm": 5.419764995574951, + "learning_rate": 1.0857550905994175e-05, + "loss": 1.6862, + "step": 3705 + }, + { + "epoch": 1.5876310721164133, + "grad_norm": 4.915116786956787, + "learning_rate": 1.0749980269633243e-05, + "loss": 1.6878, + "step": 3710 + }, + { + "epoch": 1.5897710250374493, + "grad_norm": 5.841743469238281, + "learning_rate": 1.0642880944472878e-05, + "loss": 1.6474, + "step": 3715 + }, + { + "epoch": 1.5919109779584848, + "grad_norm": 5.388934135437012, + "learning_rate": 1.0536254216553487e-05, + "loss": 1.6828, + "step": 3720 + }, + { + "epoch": 1.5940509308795208, + "grad_norm": 4.492440223693848, + "learning_rate": 1.0430101366240575e-05, + "loss": 1.6947, + "step": 3725 + }, + { + "epoch": 1.5961908838005563, + "grad_norm": 4.8876495361328125, + "learning_rate": 1.0324423668209349e-05, + "loss": 1.7048, + "step": 3730 + }, + { + "epoch": 1.5983308367215923, + "grad_norm": 5.363291263580322, + "learning_rate": 1.021922239142944e-05, + "loss": 1.6837, + "step": 3735 + }, + { + "epoch": 1.6004707896426278, + "grad_norm": 5.814448356628418, + "learning_rate": 1.0114498799149635e-05, + "loss": 1.6784, + "step": 3740 + }, + { + "epoch": 1.6026107425636638, + "grad_norm": 5.522982120513916, + "learning_rate": 1.0010254148882731e-05, + "loss": 1.6786, + "step": 3745 + }, + { + "epoch": 1.6047506954846993, + "grad_norm": 5.147754192352295, + "learning_rate": 9.906489692390426e-06, + "loss": 1.7548, + "step": 3750 + }, + { + "epoch": 1.606890648405735, + "grad_norm": 5.689370632171631, + "learning_rate": 9.803206675668286e-06, + "loss": 1.6861, + "step": 3755 + }, + { + "epoch": 1.6090306013267708, + "grad_norm": 5.778447151184082, + "learning_rate": 9.700406338930778e-06, + "loss": 1.7062, + "step": 3760 + }, + { + "epoch": 1.6111705542478065, + "grad_norm": 4.873246669769287, + "learning_rate": 9.59808991659641e-06, + "loss": 1.6678, + "step": 3765 + }, + { + "epoch": 1.6133105071688423, + "grad_norm": 5.669099807739258, + "learning_rate": 9.496258637272849e-06, + "loss": 1.6874, + "step": 3770 + }, + { + "epoch": 1.615450460089878, + "grad_norm": 5.129676818847656, + "learning_rate": 9.394913723742227e-06, + "loss": 1.6908, + "step": 3775 + }, + { + "epoch": 1.6175904130109138, + "grad_norm": 5.257515907287598, + "learning_rate": 9.294056392946427e-06, + "loss": 1.673, + "step": 3780 + }, + { + "epoch": 1.6197303659319495, + "grad_norm": 4.824296474456787, + "learning_rate": 9.193687855972466e-06, + "loss": 1.673, + "step": 3785 + }, + { + "epoch": 1.6218703188529853, + "grad_norm": 5.123586654663086, + "learning_rate": 9.093809318037989e-06, + "loss": 1.7033, + "step": 3790 + }, + { + "epoch": 1.624010271774021, + "grad_norm": 4.633101463317871, + "learning_rate": 8.994421978476735e-06, + "loss": 1.693, + "step": 3795 + }, + { + "epoch": 1.6261502246950568, + "grad_norm": 5.188746452331543, + "learning_rate": 8.8955270307242e-06, + "loss": 1.6694, + "step": 3800 + }, + { + "epoch": 1.6282901776160923, + "grad_norm": 5.478397369384766, + "learning_rate": 8.797125662303257e-06, + "loss": 1.6574, + "step": 3805 + }, + { + "epoch": 1.6304301305371283, + "grad_norm": 5.444027900695801, + "learning_rate": 8.699219054809937e-06, + "loss": 1.6467, + "step": 3810 + }, + { + "epoch": 1.6325700834581638, + "grad_norm": 4.998203754425049, + "learning_rate": 8.6018083838992e-06, + "loss": 1.6715, + "step": 3815 + }, + { + "epoch": 1.6347100363791998, + "grad_norm": 5.098468780517578, + "learning_rate": 8.504894819270854e-06, + "loss": 1.6857, + "step": 3820 + }, + { + "epoch": 1.6368499893002353, + "grad_norm": 5.074070453643799, + "learning_rate": 8.408479524655477e-06, + "loss": 1.728, + "step": 3825 + }, + { + "epoch": 1.6389899422212713, + "grad_norm": 5.682558536529541, + "learning_rate": 8.312563657800475e-06, + "loss": 1.6509, + "step": 3830 + }, + { + "epoch": 1.6411298951423068, + "grad_norm": 5.497644424438477, + "learning_rate": 8.217148370456152e-06, + "loss": 1.6649, + "step": 3835 + }, + { + "epoch": 1.6432698480633428, + "grad_norm": 4.901436805725098, + "learning_rate": 8.122234808361907e-06, + "loss": 1.6662, + "step": 3840 + }, + { + "epoch": 1.6454098009843783, + "grad_norm": 4.8296732902526855, + "learning_rate": 8.027824111232435e-06, + "loss": 1.6546, + "step": 3845 + }, + { + "epoch": 1.647549753905414, + "grad_norm": 5.374551773071289, + "learning_rate": 7.933917412744097e-06, + "loss": 1.6182, + "step": 3850 + }, + { + "epoch": 1.6496897068264498, + "grad_norm": 4.431797504425049, + "learning_rate": 7.840515840521263e-06, + "loss": 1.6917, + "step": 3855 + }, + { + "epoch": 1.6518296597474855, + "grad_norm": 4.973811626434326, + "learning_rate": 7.747620516122777e-06, + "loss": 1.625, + "step": 3860 + }, + { + "epoch": 1.6539696126685213, + "grad_norm": 5.412280082702637, + "learning_rate": 7.655232555028518e-06, + "loss": 1.6722, + "step": 3865 + }, + { + "epoch": 1.656109565589557, + "grad_norm": 5.874164581298828, + "learning_rate": 7.563353066625972e-06, + "loss": 1.6448, + "step": 3870 + }, + { + "epoch": 1.6582495185105928, + "grad_norm": 5.387955188751221, + "learning_rate": 7.471983154196932e-06, + "loss": 1.6627, + "step": 3875 + }, + { + "epoch": 1.6603894714316285, + "grad_norm": 4.935237407684326, + "learning_rate": 7.381123914904231e-06, + "loss": 1.6915, + "step": 3880 + }, + { + "epoch": 1.6625294243526643, + "grad_norm": 4.828623294830322, + "learning_rate": 7.2907764397785845e-06, + "loss": 1.6893, + "step": 3885 + }, + { + "epoch": 1.6646693772737, + "grad_norm": 5.0376434326171875, + "learning_rate": 7.200941813705497e-06, + "loss": 1.6789, + "step": 3890 + }, + { + "epoch": 1.6668093301947358, + "grad_norm": 5.704905986785889, + "learning_rate": 7.111621115412193e-06, + "loss": 1.6262, + "step": 3895 + }, + { + "epoch": 1.6689492831157713, + "grad_norm": 5.178145885467529, + "learning_rate": 7.02281541745472e-06, + "loss": 1.7011, + "step": 3900 + }, + { + "epoch": 1.6710892360368073, + "grad_norm": 5.51984167098999, + "learning_rate": 6.9345257862050264e-06, + "loss": 1.6325, + "step": 3905 + }, + { + "epoch": 1.6732291889578428, + "grad_norm": 5.094444274902344, + "learning_rate": 6.846753281838169e-06, + "loss": 1.687, + "step": 3910 + }, + { + "epoch": 1.6753691418788788, + "grad_norm": 4.961489200592041, + "learning_rate": 6.759498958319599e-06, + "loss": 1.7111, + "step": 3915 + }, + { + "epoch": 1.6775090947999143, + "grad_norm": 5.067066669464111, + "learning_rate": 6.6727638633924725e-06, + "loss": 1.6547, + "step": 3920 + }, + { + "epoch": 1.6796490477209503, + "grad_norm": 4.8792033195495605, + "learning_rate": 6.58654903856511e-06, + "loss": 1.645, + "step": 3925 + }, + { + "epoch": 1.6817890006419858, + "grad_norm": 5.298194408416748, + "learning_rate": 6.500855519098448e-06, + "loss": 1.6642, + "step": 3930 + }, + { + "epoch": 1.6839289535630217, + "grad_norm": 4.876554489135742, + "learning_rate": 6.415684333993649e-06, + "loss": 1.7346, + "step": 3935 + }, + { + "epoch": 1.6860689064840573, + "grad_norm": 5.350904941558838, + "learning_rate": 6.3310365059797094e-06, + "loss": 1.7127, + "step": 3940 + }, + { + "epoch": 1.6882088594050932, + "grad_norm": 5.282578945159912, + "learning_rate": 6.246913051501202e-06, + "loss": 1.6912, + "step": 3945 + }, + { + "epoch": 1.6903488123261288, + "grad_norm": 5.434889793395996, + "learning_rate": 6.163314980706058e-06, + "loss": 1.6402, + "step": 3950 + }, + { + "epoch": 1.6924887652471645, + "grad_norm": 5.1130571365356445, + "learning_rate": 6.080243297433447e-06, + "loss": 1.6301, + "step": 3955 + }, + { + "epoch": 1.6946287181682003, + "grad_norm": 5.187545299530029, + "learning_rate": 5.997698999201723e-06, + "loss": 1.6238, + "step": 3960 + }, + { + "epoch": 1.696768671089236, + "grad_norm": 4.982557773590088, + "learning_rate": 5.915683077196415e-06, + "loss": 1.6592, + "step": 3965 + }, + { + "epoch": 1.6989086240102718, + "grad_norm": 4.824088096618652, + "learning_rate": 5.834196516258378e-06, + "loss": 1.6593, + "step": 3970 + }, + { + "epoch": 1.7010485769313075, + "grad_norm": 4.828253746032715, + "learning_rate": 5.753240294871937e-06, + "loss": 1.6426, + "step": 3975 + }, + { + "epoch": 1.7031885298523433, + "grad_norm": 5.136155605316162, + "learning_rate": 5.6728153851531295e-06, + "loss": 1.6446, + "step": 3980 + }, + { + "epoch": 1.705328482773379, + "grad_norm": 5.751850128173828, + "learning_rate": 5.592922752838053e-06, + "loss": 1.6836, + "step": 3985 + }, + { + "epoch": 1.7074684356944148, + "grad_norm": 5.147610664367676, + "learning_rate": 5.513563357271256e-06, + "loss": 1.6492, + "step": 3990 + }, + { + "epoch": 1.7096083886154505, + "grad_norm": 5.698842525482178, + "learning_rate": 5.43473815139422e-06, + "loss": 1.643, + "step": 3995 + }, + { + "epoch": 1.7117483415364863, + "grad_norm": 4.993765354156494, + "learning_rate": 5.356448081733922e-06, + "loss": 1.6808, + "step": 4000 + }, + { + "epoch": 1.7138882944575218, + "grad_norm": 5.209753036499023, + "learning_rate": 5.278694088391462e-06, + "loss": 1.6976, + "step": 4005 + }, + { + "epoch": 1.7160282473785577, + "grad_norm": 4.791884422302246, + "learning_rate": 5.201477105030766e-06, + "loss": 1.6394, + "step": 4010 + }, + { + "epoch": 1.7181682002995933, + "grad_norm": 5.030315399169922, + "learning_rate": 5.124798058867414e-06, + "loss": 1.5982, + "step": 4015 + }, + { + "epoch": 1.7203081532206292, + "grad_norm": 5.747689723968506, + "learning_rate": 5.048657870657447e-06, + "loss": 1.6619, + "step": 4020 + }, + { + "epoch": 1.7224481061416648, + "grad_norm": 4.98396110534668, + "learning_rate": 4.973057454686364e-06, + "loss": 1.7076, + "step": 4025 + }, + { + "epoch": 1.7245880590627007, + "grad_norm": 4.828854560852051, + "learning_rate": 4.897997718758107e-06, + "loss": 1.6687, + "step": 4030 + }, + { + "epoch": 1.7267280119837363, + "grad_norm": 4.917362689971924, + "learning_rate": 4.82347956418418e-06, + "loss": 1.6901, + "step": 4035 + }, + { + "epoch": 1.7288679649047722, + "grad_norm": 5.416051387786865, + "learning_rate": 4.7495038857728155e-06, + "loss": 1.6381, + "step": 4040 + }, + { + "epoch": 1.7310079178258078, + "grad_norm": 5.832434177398682, + "learning_rate": 4.676071571818236e-06, + "loss": 1.6695, + "step": 4045 + }, + { + "epoch": 1.7331478707468437, + "grad_norm": 5.534262657165527, + "learning_rate": 4.603183504089997e-06, + "loss": 1.7054, + "step": 4050 + }, + { + "epoch": 1.7352878236678793, + "grad_norm": 5.344274997711182, + "learning_rate": 4.5308405578223635e-06, + "loss": 1.6645, + "step": 4055 + }, + { + "epoch": 1.737427776588915, + "grad_norm": 5.08225679397583, + "learning_rate": 4.45904360170385e-06, + "loss": 1.6623, + "step": 4060 + }, + { + "epoch": 1.7395677295099508, + "grad_norm": 5.531522274017334, + "learning_rate": 4.387793497866744e-06, + "loss": 1.6633, + "step": 4065 + }, + { + "epoch": 1.7417076824309865, + "grad_norm": 4.792389869689941, + "learning_rate": 4.31709110187678e-06, + "loss": 1.6761, + "step": 4070 + }, + { + "epoch": 1.7438476353520223, + "grad_norm": 5.555240631103516, + "learning_rate": 4.246937262722866e-06, + "loss": 1.6628, + "step": 4075 + }, + { + "epoch": 1.745987588273058, + "grad_norm": 5.5509796142578125, + "learning_rate": 4.177332822806873e-06, + "loss": 1.6529, + "step": 4080 + }, + { + "epoch": 1.7481275411940937, + "grad_norm": 4.76316499710083, + "learning_rate": 4.108278617933525e-06, + "loss": 1.6898, + "step": 4085 + }, + { + "epoch": 1.7502674941151295, + "grad_norm": 4.821498870849609, + "learning_rate": 4.039775477300378e-06, + "loss": 1.6823, + "step": 4090 + }, + { + "epoch": 1.7524074470361652, + "grad_norm": 5.2127299308776855, + "learning_rate": 3.971824223487841e-06, + "loss": 1.7028, + "step": 4095 + }, + { + "epoch": 1.754547399957201, + "grad_norm": 4.9137701988220215, + "learning_rate": 3.90442567244933e-06, + "loss": 1.6858, + "step": 4100 + }, + { + "epoch": 1.7566873528782367, + "grad_norm": 5.513342380523682, + "learning_rate": 3.837580633501409e-06, + "loss": 1.6466, + "step": 4105 + }, + { + "epoch": 1.7588273057992723, + "grad_norm": 4.806028842926025, + "learning_rate": 3.7712899093141407e-06, + "loss": 1.7288, + "step": 4110 + }, + { + "epoch": 1.7609672587203082, + "grad_norm": 5.0646796226501465, + "learning_rate": 3.7055542959014087e-06, + "loss": 1.6827, + "step": 4115 + }, + { + "epoch": 1.7631072116413438, + "grad_norm": 5.064296245574951, + "learning_rate": 3.640374582611389e-06, + "loss": 1.6403, + "step": 4120 + }, + { + "epoch": 1.7652471645623797, + "grad_norm": 6.094568252563477, + "learning_rate": 3.575751552117029e-06, + "loss": 1.625, + "step": 4125 + }, + { + "epoch": 1.7673871174834153, + "grad_norm": 4.940478324890137, + "learning_rate": 3.511685980406676e-06, + "loss": 1.6847, + "step": 4130 + }, + { + "epoch": 1.7695270704044512, + "grad_norm": 4.668292999267578, + "learning_rate": 3.4481786367747627e-06, + "loss": 1.6886, + "step": 4135 + }, + { + "epoch": 1.7716670233254868, + "grad_norm": 5.22495174407959, + "learning_rate": 3.3852302838125626e-06, + "loss": 1.6492, + "step": 4140 + }, + { + "epoch": 1.7738069762465227, + "grad_norm": 5.06997013092041, + "learning_rate": 3.322841677399019e-06, + "loss": 1.6441, + "step": 4145 + }, + { + "epoch": 1.7759469291675583, + "grad_norm": 4.730724811553955, + "learning_rate": 3.2610135666917007e-06, + "loss": 1.6843, + "step": 4150 + }, + { + "epoch": 1.778086882088594, + "grad_norm": 5.150337219238281, + "learning_rate": 3.1997466941177666e-06, + "loss": 1.6854, + "step": 4155 + }, + { + "epoch": 1.7802268350096297, + "grad_norm": 4.789018630981445, + "learning_rate": 3.139041795365094e-06, + "loss": 1.634, + "step": 4160 + }, + { + "epoch": 1.7823667879306655, + "grad_norm": 4.8880462646484375, + "learning_rate": 3.0788995993734083e-06, + "loss": 1.6258, + "step": 4165 + }, + { + "epoch": 1.7845067408517012, + "grad_norm": 4.961301803588867, + "learning_rate": 3.019320828325539e-06, + "loss": 1.6463, + "step": 4170 + }, + { + "epoch": 1.786646693772737, + "grad_norm": 4.813554286956787, + "learning_rate": 2.9603061976387736e-06, + "loss": 1.6903, + "step": 4175 + }, + { + "epoch": 1.7887866466937727, + "grad_norm": 4.724278450012207, + "learning_rate": 2.9018564159562224e-06, + "loss": 1.646, + "step": 4180 + }, + { + "epoch": 1.7909265996148085, + "grad_norm": 5.349635601043701, + "learning_rate": 2.8439721851383383e-06, + "loss": 1.6847, + "step": 4185 + }, + { + "epoch": 1.7930665525358442, + "grad_norm": 4.801799297332764, + "learning_rate": 2.786654200254496e-06, + "loss": 1.6871, + "step": 4190 + }, + { + "epoch": 1.79520650545688, + "grad_norm": 4.788339138031006, + "learning_rate": 2.7299031495746252e-06, + "loss": 1.7072, + "step": 4195 + }, + { + "epoch": 1.7973464583779157, + "grad_norm": 4.700235843658447, + "learning_rate": 2.6737197145609404e-06, + "loss": 1.6723, + "step": 4200 + }, + { + "epoch": 1.7994864112989513, + "grad_norm": 4.791804313659668, + "learning_rate": 2.6181045698597972e-06, + "loss": 1.674, + "step": 4205 + }, + { + "epoch": 1.8016263642199872, + "grad_norm": 5.270465850830078, + "learning_rate": 2.563058383293537e-06, + "loss": 1.6434, + "step": 4210 + }, + { + "epoch": 1.8037663171410228, + "grad_norm": 5.29375696182251, + "learning_rate": 2.508581815852523e-06, + "loss": 1.6719, + "step": 4215 + }, + { + "epoch": 1.8059062700620587, + "grad_norm": 4.851658821105957, + "learning_rate": 2.4546755216871496e-06, + "loss": 1.6652, + "step": 4220 + }, + { + "epoch": 1.8080462229830943, + "grad_norm": 5.476074695587158, + "learning_rate": 2.40134014810004e-06, + "loss": 1.7037, + "step": 4225 + }, + { + "epoch": 1.8101861759041302, + "grad_norm": 5.279761791229248, + "learning_rate": 2.3485763355382273e-06, + "loss": 1.6697, + "step": 4230 + }, + { + "epoch": 1.8123261288251657, + "grad_norm": 4.817234039306641, + "learning_rate": 2.296384717585487e-06, + "loss": 1.6609, + "step": 4235 + }, + { + "epoch": 1.8144660817462017, + "grad_norm": 5.009280681610107, + "learning_rate": 2.2447659209547443e-06, + "loss": 1.6959, + "step": 4240 + }, + { + "epoch": 1.8166060346672372, + "grad_norm": 5.085817813873291, + "learning_rate": 2.1937205654805004e-06, + "loss": 1.6254, + "step": 4245 + }, + { + "epoch": 1.8187459875882732, + "grad_norm": 4.714953899383545, + "learning_rate": 2.143249264111441e-06, + "loss": 1.678, + "step": 4250 + }, + { + "epoch": 1.8208859405093087, + "grad_norm": 5.392242908477783, + "learning_rate": 2.09335262290305e-06, + "loss": 1.6443, + "step": 4255 + }, + { + "epoch": 1.8230258934303445, + "grad_norm": 4.728943347930908, + "learning_rate": 2.04403124101033e-06, + "loss": 1.6737, + "step": 4260 + }, + { + "epoch": 1.8251658463513802, + "grad_norm": 4.887874603271484, + "learning_rate": 1.995285710680622e-06, + "loss": 1.6345, + "step": 4265 + }, + { + "epoch": 1.827305799272416, + "grad_norm": 4.753354549407959, + "learning_rate": 1.9471166172464917e-06, + "loss": 1.7165, + "step": 4270 + }, + { + "epoch": 1.8294457521934517, + "grad_norm": 5.000302314758301, + "learning_rate": 1.8995245391186688e-06, + "loss": 1.6339, + "step": 4275 + }, + { + "epoch": 1.8315857051144875, + "grad_norm": 4.688614368438721, + "learning_rate": 1.8525100477791602e-06, + "loss": 1.6857, + "step": 4280 + }, + { + "epoch": 1.8337256580355232, + "grad_norm": 4.900327205657959, + "learning_rate": 1.8060737077743416e-06, + "loss": 1.671, + "step": 4285 + }, + { + "epoch": 1.835865610956559, + "grad_norm": 4.996348857879639, + "learning_rate": 1.7602160767081822e-06, + "loss": 1.6386, + "step": 4290 + }, + { + "epoch": 1.8380055638775947, + "grad_norm": 4.658572673797607, + "learning_rate": 1.7149377052355698e-06, + "loss": 1.6028, + "step": 4295 + }, + { + "epoch": 1.8401455167986305, + "grad_norm": 4.834623336791992, + "learning_rate": 1.6702391370556957e-06, + "loss": 1.6937, + "step": 4300 + }, + { + "epoch": 1.8422854697196662, + "grad_norm": 4.5502028465271, + "learning_rate": 1.6261209089054986e-06, + "loss": 1.6596, + "step": 4305 + }, + { + "epoch": 1.8444254226407018, + "grad_norm": 4.947803497314453, + "learning_rate": 1.5825835505532516e-06, + "loss": 1.6538, + "step": 4310 + }, + { + "epoch": 1.8465653755617377, + "grad_norm": 5.3801774978637695, + "learning_rate": 1.539627584792186e-06, + "loss": 1.661, + "step": 4315 + }, + { + "epoch": 1.8487053284827732, + "grad_norm": 5.104175090789795, + "learning_rate": 1.4972535274342225e-06, + "loss": 1.6419, + "step": 4320 + }, + { + "epoch": 1.8508452814038092, + "grad_norm": 5.043144702911377, + "learning_rate": 1.4554618873037551e-06, + "loss": 1.6123, + "step": 4325 + }, + { + "epoch": 1.8529852343248447, + "grad_norm": 4.843111038208008, + "learning_rate": 1.4142531662315662e-06, + "loss": 1.6482, + "step": 4330 + }, + { + "epoch": 1.8551251872458807, + "grad_norm": 4.794622898101807, + "learning_rate": 1.3736278590487927e-06, + "loss": 1.7072, + "step": 4335 + }, + { + "epoch": 1.8572651401669162, + "grad_norm": 4.912827491760254, + "learning_rate": 1.3335864535809872e-06, + "loss": 1.6744, + "step": 4340 + }, + { + "epoch": 1.8594050930879522, + "grad_norm": 5.024925708770752, + "learning_rate": 1.294129430642238e-06, + "loss": 1.6379, + "step": 4345 + }, + { + "epoch": 1.8615450460089877, + "grad_norm": 4.962955951690674, + "learning_rate": 1.2552572640294247e-06, + "loss": 1.6736, + "step": 4350 + }, + { + "epoch": 1.8636849989300235, + "grad_norm": 4.8275275230407715, + "learning_rate": 1.2169704205165277e-06, + "loss": 1.6271, + "step": 4355 + }, + { + "epoch": 1.8658249518510592, + "grad_norm": 4.821237087249756, + "learning_rate": 1.179269359848989e-06, + "loss": 1.6363, + "step": 4360 + }, + { + "epoch": 1.867964904772095, + "grad_norm": 4.859674453735352, + "learning_rate": 1.1421545347382378e-06, + "loss": 1.6276, + "step": 4365 + }, + { + "epoch": 1.8701048576931307, + "grad_norm": 4.546759605407715, + "learning_rate": 1.105626390856218e-06, + "loss": 1.6892, + "step": 4370 + }, + { + "epoch": 1.8722448106141665, + "grad_norm": 4.87622594833374, + "learning_rate": 1.0696853668300588e-06, + "loss": 1.6633, + "step": 4375 + }, + { + "epoch": 1.8743847635352022, + "grad_norm": 5.357635021209717, + "learning_rate": 1.0343318942367951e-06, + "loss": 1.6551, + "step": 4380 + }, + { + "epoch": 1.876524716456238, + "grad_norm": 4.706364631652832, + "learning_rate": 9.995663975981894e-07, + "loss": 1.6708, + "step": 4385 + }, + { + "epoch": 1.8786646693772737, + "grad_norm": 5.037484645843506, + "learning_rate": 9.653892943756405e-07, + "loss": 1.6359, + "step": 4390 + }, + { + "epoch": 1.8808046222983095, + "grad_norm": 4.868020534515381, + "learning_rate": 9.31800994965154e-07, + "loss": 1.717, + "step": 4395 + }, + { + "epoch": 1.8829445752193452, + "grad_norm": 5.029200077056885, + "learning_rate": 8.98801902692431e-07, + "loss": 1.6388, + "step": 4400 + }, + { + "epoch": 1.885084528140381, + "grad_norm": 5.490426063537598, + "learning_rate": 8.663924138080204e-07, + "loss": 1.6471, + "step": 4405 + }, + { + "epoch": 1.8872244810614167, + "grad_norm": 4.652840614318848, + "learning_rate": 8.345729174825623e-07, + "loss": 1.684, + "step": 4410 + }, + { + "epoch": 1.8893644339824522, + "grad_norm": 4.7486252784729, + "learning_rate": 8.033437958020973e-07, + "loss": 1.701, + "step": 4415 + }, + { + "epoch": 1.8915043869034882, + "grad_norm": 4.830610275268555, + "learning_rate": 7.727054237635146e-07, + "loss": 1.6273, + "step": 4420 + }, + { + "epoch": 1.8936443398245237, + "grad_norm": 4.977087497711182, + "learning_rate": 7.426581692700052e-07, + "loss": 1.6923, + "step": 4425 + }, + { + "epoch": 1.8957842927455597, + "grad_norm": 4.950675010681152, + "learning_rate": 7.132023931266829e-07, + "loss": 1.6456, + "step": 4430 + }, + { + "epoch": 1.8979242456665952, + "grad_norm": 4.89157247543335, + "learning_rate": 6.843384490362259e-07, + "loss": 1.6838, + "step": 4435 + }, + { + "epoch": 1.9000641985876312, + "grad_norm": 4.860950946807861, + "learning_rate": 6.560666835946416e-07, + "loss": 1.6309, + "step": 4440 + }, + { + "epoch": 1.9022041515086667, + "grad_norm": 4.52632999420166, + "learning_rate": 6.283874362871033e-07, + "loss": 1.6225, + "step": 4445 + }, + { + "epoch": 1.9043441044297027, + "grad_norm": 4.779524803161621, + "learning_rate": 6.013010394838702e-07, + "loss": 1.7016, + "step": 4450 + }, + { + "epoch": 1.9064840573507382, + "grad_norm": 4.958658695220947, + "learning_rate": 5.748078184363015e-07, + "loss": 1.6794, + "step": 4455 + }, + { + "epoch": 1.908624010271774, + "grad_norm": 5.345273017883301, + "learning_rate": 5.489080912729428e-07, + "loss": 1.6796, + "step": 4460 + }, + { + "epoch": 1.9107639631928097, + "grad_norm": 5.009010314941406, + "learning_rate": 5.236021689957237e-07, + "loss": 1.6552, + "step": 4465 + }, + { + "epoch": 1.9129039161138455, + "grad_norm": 4.576427936553955, + "learning_rate": 4.988903554761948e-07, + "loss": 1.64, + "step": 4470 + }, + { + "epoch": 1.9150438690348812, + "grad_norm": 4.975499153137207, + "learning_rate": 4.7477294745189584e-07, + "loss": 1.6782, + "step": 4475 + }, + { + "epoch": 1.917183821955917, + "grad_norm": 4.649482250213623, + "learning_rate": 4.512502345228042e-07, + "loss": 1.6728, + "step": 4480 + }, + { + "epoch": 1.9193237748769527, + "grad_norm": 4.831498146057129, + "learning_rate": 4.283224991478374e-07, + "loss": 1.6622, + "step": 4485 + }, + { + "epoch": 1.9214637277979885, + "grad_norm": 4.552825927734375, + "learning_rate": 4.0599001664146097e-07, + "loss": 1.6806, + "step": 4490 + }, + { + "epoch": 1.9236036807190242, + "grad_norm": 4.854389190673828, + "learning_rate": 3.842530551704027e-07, + "loss": 1.5925, + "step": 4495 + }, + { + "epoch": 1.92574363364006, + "grad_norm": 5.051529884338379, + "learning_rate": 3.631118757504159e-07, + "loss": 1.7039, + "step": 4500 + }, + { + "epoch": 1.9278835865610957, + "grad_norm": 4.859341144561768, + "learning_rate": 3.425667322431436e-07, + "loss": 1.6972, + "step": 4505 + }, + { + "epoch": 1.9300235394821312, + "grad_norm": 4.893909454345703, + "learning_rate": 3.226178713530814e-07, + "loss": 1.635, + "step": 4510 + }, + { + "epoch": 1.9321634924031672, + "grad_norm": 4.657041549682617, + "learning_rate": 3.0326553262460255e-07, + "loss": 1.6971, + "step": 4515 + }, + { + "epoch": 1.9343034453242027, + "grad_norm": 4.919985294342041, + "learning_rate": 2.845099484391045e-07, + "loss": 1.6954, + "step": 4520 + }, + { + "epoch": 1.9364433982452387, + "grad_norm": 5.278502464294434, + "learning_rate": 2.663513440121834e-07, + "loss": 1.6607, + "step": 4525 + }, + { + "epoch": 1.9385833511662742, + "grad_norm": 4.634570598602295, + "learning_rate": 2.4878993739095857e-07, + "loss": 1.6459, + "step": 4530 + }, + { + "epoch": 1.9407233040873102, + "grad_norm": 4.695330619812012, + "learning_rate": 2.3182593945144105e-07, + "loss": 1.6391, + "step": 4535 + }, + { + "epoch": 1.9428632570083457, + "grad_norm": 4.747430324554443, + "learning_rate": 2.1545955389600248e-07, + "loss": 1.6485, + "step": 4540 + }, + { + "epoch": 1.9450032099293817, + "grad_norm": 5.009528636932373, + "learning_rate": 1.9969097725094366e-07, + "loss": 1.686, + "step": 4545 + }, + { + "epoch": 1.9471431628504172, + "grad_norm": 4.869572639465332, + "learning_rate": 1.8452039886410199e-07, + "loss": 1.6684, + "step": 4550 + }, + { + "epoch": 1.9492831157714532, + "grad_norm": 4.405606269836426, + "learning_rate": 1.6994800090261997e-07, + "loss": 1.6607, + "step": 4555 + }, + { + "epoch": 1.9514230686924887, + "grad_norm": 4.575326919555664, + "learning_rate": 1.5597395835071915e-07, + "loss": 1.6839, + "step": 4560 + }, + { + "epoch": 1.9535630216135245, + "grad_norm": 4.905466079711914, + "learning_rate": 1.425984390076185e-07, + "loss": 1.6075, + "step": 4565 + }, + { + "epoch": 1.9557029745345602, + "grad_norm": 4.696757793426514, + "learning_rate": 1.298216034855304e-07, + "loss": 1.5987, + "step": 4570 + }, + { + "epoch": 1.957842927455596, + "grad_norm": 5.084411144256592, + "learning_rate": 1.1764360520769568e-07, + "loss": 1.6828, + "step": 4575 + }, + { + "epoch": 1.9599828803766317, + "grad_norm": 4.676747798919678, + "learning_rate": 1.0606459040657935e-07, + "loss": 1.7003, + "step": 4580 + }, + { + "epoch": 1.9621228332976675, + "grad_norm": 4.811861991882324, + "learning_rate": 9.508469812209986e-08, + "loss": 1.6344, + "step": 4585 + }, + { + "epoch": 1.9642627862187032, + "grad_norm": 4.689334392547607, + "learning_rate": 8.470406019994714e-08, + "loss": 1.673, + "step": 4590 + }, + { + "epoch": 1.966402739139739, + "grad_norm": 4.811148166656494, + "learning_rate": 7.492280129002271e-08, + "loss": 1.6674, + "step": 4595 + }, + { + "epoch": 1.9685426920607747, + "grad_norm": 4.551917552947998, + "learning_rate": 6.574103884492422e-08, + "loss": 1.6488, + "step": 4600 + }, + { + "epoch": 1.9706826449818104, + "grad_norm": 4.65984582901001, + "learning_rate": 5.715888311855211e-08, + "loss": 1.6826, + "step": 4605 + }, + { + "epoch": 1.9728225979028462, + "grad_norm": 4.9454240798950195, + "learning_rate": 4.9176437164760726e-08, + "loss": 1.6969, + "step": 4610 + }, + { + "epoch": 1.9749625508238817, + "grad_norm": 4.670926570892334, + "learning_rate": 4.1793796836142596e-08, + "loss": 1.6635, + "step": 4615 + }, + { + "epoch": 1.9771025037449177, + "grad_norm": 4.9202399253845215, + "learning_rate": 3.5011050782879364e-08, + "loss": 1.6676, + "step": 4620 + }, + { + "epoch": 1.9792424566659532, + "grad_norm": 4.523751258850098, + "learning_rate": 2.8828280451653755e-08, + "loss": 1.6235, + "step": 4625 + }, + { + "epoch": 1.9813824095869892, + "grad_norm": 4.875797748565674, + "learning_rate": 2.3245560084700357e-08, + "loss": 1.7377, + "step": 4630 + }, + { + "epoch": 1.9835223625080247, + "grad_norm": 4.714092254638672, + "learning_rate": 1.8262956718884117e-08, + "loss": 1.6771, + "step": 4635 + }, + { + "epoch": 1.9856623154290607, + "grad_norm": 4.643798351287842, + "learning_rate": 1.3880530184934293e-08, + "loss": 1.6668, + "step": 4640 + }, + { + "epoch": 1.9878022683500962, + "grad_norm": 4.464611530303955, + "learning_rate": 1.0098333106672852e-08, + "loss": 1.6362, + "step": 4645 + }, + { + "epoch": 1.9899422212711322, + "grad_norm": 4.848086357116699, + "learning_rate": 6.9164109004427046e-09, + "loss": 1.6405, + "step": 4650 + }, + { + "epoch": 1.9920821741921677, + "grad_norm": 4.318144798278809, + "learning_rate": 4.33480177451373e-09, + "loss": 1.6798, + "step": 4655 + }, + { + "epoch": 1.9942221271132035, + "grad_norm": 5.152481555938721, + "learning_rate": 2.3535367286497966e-09, + "loss": 1.6778, + "step": 4660 + }, + { + "epoch": 1.9963620800342392, + "grad_norm": 4.668585777282715, + "learning_rate": 9.726395537312806e-10, + "loss": 1.6754, + "step": 4665 + }, + { + "epoch": 1.998502032955275, + "grad_norm": 4.8391313552856445, + "learning_rate": 1.921268314608593e-10, + "loss": 1.6921, + "step": 4670 + }, + { + "epoch": 2.0, + "step": 4674, + "total_flos": 7.23540271133465e+18, + "train_loss": 2.1290212889537883, + "train_runtime": 7155.5853, + "train_samples_per_second": 104.472, + "train_steps_per_second": 0.653 + } + ], + "logging_steps": 5, + "max_steps": 4674, + "num_input_tokens_seen": 0, + "num_train_epochs": 2, + "save_steps": 468, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 7.23540271133465e+18, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}