lesso's picture
Training in progress, step 150, checkpoint
b034e98 verified
raw
history blame
27.8 kB
{
"best_metric": 0.6001591086387634,
"best_model_checkpoint": "miner_id_24/checkpoint-150",
"epoch": 0.15128593040847202,
"eval_steps": 50,
"global_step": 150,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0010085728693898135,
"grad_norm": 1.311654806137085,
"learning_rate": 1.0018000000000001e-05,
"loss": 0.9276,
"step": 1
},
{
"epoch": 0.0010085728693898135,
"eval_loss": 1.104080319404602,
"eval_runtime": 47.2809,
"eval_samples_per_second": 8.841,
"eval_steps_per_second": 2.221,
"step": 1
},
{
"epoch": 0.002017145738779627,
"grad_norm": 1.4768002033233643,
"learning_rate": 2.0036000000000003e-05,
"loss": 1.1293,
"step": 2
},
{
"epoch": 0.0030257186081694403,
"grad_norm": 1.863478422164917,
"learning_rate": 3.0054e-05,
"loss": 1.095,
"step": 3
},
{
"epoch": 0.004034291477559254,
"grad_norm": 1.4832903146743774,
"learning_rate": 4.0072000000000005e-05,
"loss": 1.0675,
"step": 4
},
{
"epoch": 0.005042864346949067,
"grad_norm": 1.516528606414795,
"learning_rate": 5.009e-05,
"loss": 1.1571,
"step": 5
},
{
"epoch": 0.006051437216338881,
"grad_norm": 1.0894883871078491,
"learning_rate": 6.0108e-05,
"loss": 0.8794,
"step": 6
},
{
"epoch": 0.0070600100857286935,
"grad_norm": 1.1469849348068237,
"learning_rate": 7.0126e-05,
"loss": 0.7944,
"step": 7
},
{
"epoch": 0.008068582955118508,
"grad_norm": 0.9738489389419556,
"learning_rate": 8.014400000000001e-05,
"loss": 0.7899,
"step": 8
},
{
"epoch": 0.009077155824508321,
"grad_norm": 1.1924914121627808,
"learning_rate": 9.016200000000001e-05,
"loss": 0.8915,
"step": 9
},
{
"epoch": 0.010085728693898134,
"grad_norm": 0.986112117767334,
"learning_rate": 0.00010018,
"loss": 0.735,
"step": 10
},
{
"epoch": 0.011094301563287948,
"grad_norm": 0.9808599948883057,
"learning_rate": 9.965273684210526e-05,
"loss": 0.7869,
"step": 11
},
{
"epoch": 0.012102874432677761,
"grad_norm": 0.8853185176849365,
"learning_rate": 9.912547368421053e-05,
"loss": 0.7262,
"step": 12
},
{
"epoch": 0.013111447302067574,
"grad_norm": 0.7452360987663269,
"learning_rate": 9.859821052631579e-05,
"loss": 0.679,
"step": 13
},
{
"epoch": 0.014120020171457387,
"grad_norm": 0.7864087224006653,
"learning_rate": 9.807094736842106e-05,
"loss": 0.6494,
"step": 14
},
{
"epoch": 0.015128593040847202,
"grad_norm": 0.7922771573066711,
"learning_rate": 9.754368421052633e-05,
"loss": 0.6502,
"step": 15
},
{
"epoch": 0.016137165910237016,
"grad_norm": 0.7772244215011597,
"learning_rate": 9.701642105263158e-05,
"loss": 0.5793,
"step": 16
},
{
"epoch": 0.01714573877962683,
"grad_norm": 0.7786775827407837,
"learning_rate": 9.648915789473685e-05,
"loss": 0.5646,
"step": 17
},
{
"epoch": 0.018154311649016642,
"grad_norm": 0.9989523887634277,
"learning_rate": 9.596189473684211e-05,
"loss": 0.8072,
"step": 18
},
{
"epoch": 0.019162884518406455,
"grad_norm": 0.8197798728942871,
"learning_rate": 9.543463157894737e-05,
"loss": 0.549,
"step": 19
},
{
"epoch": 0.020171457387796268,
"grad_norm": 0.727142333984375,
"learning_rate": 9.490736842105264e-05,
"loss": 0.6405,
"step": 20
},
{
"epoch": 0.02118003025718608,
"grad_norm": 0.7259921431541443,
"learning_rate": 9.43801052631579e-05,
"loss": 0.6329,
"step": 21
},
{
"epoch": 0.022188603126575897,
"grad_norm": 0.7431148886680603,
"learning_rate": 9.385284210526316e-05,
"loss": 0.522,
"step": 22
},
{
"epoch": 0.02319717599596571,
"grad_norm": 0.8169230818748474,
"learning_rate": 9.332557894736843e-05,
"loss": 0.7403,
"step": 23
},
{
"epoch": 0.024205748865355523,
"grad_norm": 0.9084967970848083,
"learning_rate": 9.279831578947369e-05,
"loss": 0.6538,
"step": 24
},
{
"epoch": 0.025214321734745335,
"grad_norm": 0.8007498383522034,
"learning_rate": 9.227105263157896e-05,
"loss": 0.6519,
"step": 25
},
{
"epoch": 0.026222894604135148,
"grad_norm": 0.7854844927787781,
"learning_rate": 9.174378947368421e-05,
"loss": 0.6472,
"step": 26
},
{
"epoch": 0.02723146747352496,
"grad_norm": 0.8193365335464478,
"learning_rate": 9.121652631578948e-05,
"loss": 0.5204,
"step": 27
},
{
"epoch": 0.028240040342914774,
"grad_norm": 0.8219320178031921,
"learning_rate": 9.068926315789475e-05,
"loss": 0.6833,
"step": 28
},
{
"epoch": 0.02924861321230459,
"grad_norm": 0.8295514583587646,
"learning_rate": 9.016200000000001e-05,
"loss": 0.6379,
"step": 29
},
{
"epoch": 0.030257186081694403,
"grad_norm": 0.9452921748161316,
"learning_rate": 8.963473684210526e-05,
"loss": 0.7919,
"step": 30
},
{
"epoch": 0.031265758951084216,
"grad_norm": 0.893158495426178,
"learning_rate": 8.910747368421053e-05,
"loss": 0.8002,
"step": 31
},
{
"epoch": 0.03227433182047403,
"grad_norm": 0.7392338514328003,
"learning_rate": 8.858021052631579e-05,
"loss": 0.6503,
"step": 32
},
{
"epoch": 0.03328290468986384,
"grad_norm": 0.8364588618278503,
"learning_rate": 8.805294736842106e-05,
"loss": 0.7389,
"step": 33
},
{
"epoch": 0.03429147755925366,
"grad_norm": 0.8333988189697266,
"learning_rate": 8.752568421052633e-05,
"loss": 0.718,
"step": 34
},
{
"epoch": 0.03530005042864347,
"grad_norm": 0.8555812835693359,
"learning_rate": 8.699842105263159e-05,
"loss": 0.8115,
"step": 35
},
{
"epoch": 0.036308623298033284,
"grad_norm": 0.909781277179718,
"learning_rate": 8.647115789473686e-05,
"loss": 0.7834,
"step": 36
},
{
"epoch": 0.03731719616742309,
"grad_norm": 0.8471423387527466,
"learning_rate": 8.594389473684211e-05,
"loss": 0.7363,
"step": 37
},
{
"epoch": 0.03832576903681291,
"grad_norm": 0.7747591137886047,
"learning_rate": 8.541663157894737e-05,
"loss": 0.7223,
"step": 38
},
{
"epoch": 0.039334341906202726,
"grad_norm": 0.7503566145896912,
"learning_rate": 8.488936842105264e-05,
"loss": 0.6464,
"step": 39
},
{
"epoch": 0.040342914775592535,
"grad_norm": 0.7485514283180237,
"learning_rate": 8.43621052631579e-05,
"loss": 0.624,
"step": 40
},
{
"epoch": 0.04135148764498235,
"grad_norm": 0.8754268884658813,
"learning_rate": 8.383484210526316e-05,
"loss": 0.7319,
"step": 41
},
{
"epoch": 0.04236006051437216,
"grad_norm": 0.9374569058418274,
"learning_rate": 8.330757894736843e-05,
"loss": 0.8252,
"step": 42
},
{
"epoch": 0.04336863338376198,
"grad_norm": 1.0073909759521484,
"learning_rate": 8.278031578947369e-05,
"loss": 0.8029,
"step": 43
},
{
"epoch": 0.044377206253151794,
"grad_norm": 0.9302268624305725,
"learning_rate": 8.225305263157896e-05,
"loss": 0.7655,
"step": 44
},
{
"epoch": 0.0453857791225416,
"grad_norm": 0.9419411420822144,
"learning_rate": 8.172578947368422e-05,
"loss": 0.7218,
"step": 45
},
{
"epoch": 0.04639435199193142,
"grad_norm": 0.936607301235199,
"learning_rate": 8.119852631578947e-05,
"loss": 0.7549,
"step": 46
},
{
"epoch": 0.04740292486132123,
"grad_norm": 0.8890344500541687,
"learning_rate": 8.067126315789474e-05,
"loss": 0.7809,
"step": 47
},
{
"epoch": 0.048411497730711045,
"grad_norm": 1.0612900257110596,
"learning_rate": 8.014400000000001e-05,
"loss": 0.7987,
"step": 48
},
{
"epoch": 0.049420070600100854,
"grad_norm": 0.9946021437644958,
"learning_rate": 7.961673684210527e-05,
"loss": 0.8006,
"step": 49
},
{
"epoch": 0.05042864346949067,
"grad_norm": 1.1615980863571167,
"learning_rate": 7.908947368421054e-05,
"loss": 0.8517,
"step": 50
},
{
"epoch": 0.05042864346949067,
"eval_loss": 0.6663289666175842,
"eval_runtime": 47.2665,
"eval_samples_per_second": 8.843,
"eval_steps_per_second": 2.221,
"step": 50
},
{
"epoch": 0.05143721633888049,
"grad_norm": 0.8349344730377197,
"learning_rate": 7.856221052631579e-05,
"loss": 0.4587,
"step": 51
},
{
"epoch": 0.052445789208270296,
"grad_norm": 0.6953268647193909,
"learning_rate": 7.803494736842106e-05,
"loss": 0.666,
"step": 52
},
{
"epoch": 0.05345436207766011,
"grad_norm": 0.7224613428115845,
"learning_rate": 7.750768421052632e-05,
"loss": 0.7536,
"step": 53
},
{
"epoch": 0.05446293494704992,
"grad_norm": 0.6757810115814209,
"learning_rate": 7.698042105263157e-05,
"loss": 0.6169,
"step": 54
},
{
"epoch": 0.05547150781643974,
"grad_norm": 0.7635074257850647,
"learning_rate": 7.645315789473686e-05,
"loss": 0.6307,
"step": 55
},
{
"epoch": 0.05648008068582955,
"grad_norm": 0.6975710988044739,
"learning_rate": 7.592589473684211e-05,
"loss": 0.6233,
"step": 56
},
{
"epoch": 0.057488653555219364,
"grad_norm": 0.9252186417579651,
"learning_rate": 7.539863157894737e-05,
"loss": 0.6551,
"step": 57
},
{
"epoch": 0.05849722642460918,
"grad_norm": 0.8248884677886963,
"learning_rate": 7.487136842105264e-05,
"loss": 0.5789,
"step": 58
},
{
"epoch": 0.05950579929399899,
"grad_norm": 0.6074532866477966,
"learning_rate": 7.43441052631579e-05,
"loss": 0.5004,
"step": 59
},
{
"epoch": 0.060514372163388806,
"grad_norm": 0.6752936244010925,
"learning_rate": 7.381684210526315e-05,
"loss": 0.4884,
"step": 60
},
{
"epoch": 0.061522945032778616,
"grad_norm": 0.7720556259155273,
"learning_rate": 7.328957894736844e-05,
"loss": 0.7758,
"step": 61
},
{
"epoch": 0.06253151790216843,
"grad_norm": 0.699116051197052,
"learning_rate": 7.276231578947369e-05,
"loss": 0.5716,
"step": 62
},
{
"epoch": 0.06354009077155824,
"grad_norm": 0.7762227058410645,
"learning_rate": 7.223505263157895e-05,
"loss": 0.6935,
"step": 63
},
{
"epoch": 0.06454866364094806,
"grad_norm": 0.6913344264030457,
"learning_rate": 7.170778947368422e-05,
"loss": 0.6068,
"step": 64
},
{
"epoch": 0.06555723651033787,
"grad_norm": 0.7144531011581421,
"learning_rate": 7.118052631578947e-05,
"loss": 0.6431,
"step": 65
},
{
"epoch": 0.06656580937972768,
"grad_norm": 0.6667282581329346,
"learning_rate": 7.065326315789474e-05,
"loss": 0.5421,
"step": 66
},
{
"epoch": 0.06757438224911749,
"grad_norm": 0.6969847083091736,
"learning_rate": 7.0126e-05,
"loss": 0.6785,
"step": 67
},
{
"epoch": 0.06858295511850732,
"grad_norm": 0.6973515748977661,
"learning_rate": 6.959873684210527e-05,
"loss": 0.5455,
"step": 68
},
{
"epoch": 0.06959152798789713,
"grad_norm": 0.7713499069213867,
"learning_rate": 6.907147368421054e-05,
"loss": 0.7372,
"step": 69
},
{
"epoch": 0.07060010085728693,
"grad_norm": 0.5957585573196411,
"learning_rate": 6.85442105263158e-05,
"loss": 0.4954,
"step": 70
},
{
"epoch": 0.07160867372667676,
"grad_norm": 0.7025538086891174,
"learning_rate": 6.801694736842105e-05,
"loss": 0.5496,
"step": 71
},
{
"epoch": 0.07261724659606657,
"grad_norm": 0.7217448353767395,
"learning_rate": 6.748968421052632e-05,
"loss": 0.5567,
"step": 72
},
{
"epoch": 0.07362581946545638,
"grad_norm": 0.7021524906158447,
"learning_rate": 6.696242105263158e-05,
"loss": 0.5291,
"step": 73
},
{
"epoch": 0.07463439233484619,
"grad_norm": 0.7426285743713379,
"learning_rate": 6.643515789473685e-05,
"loss": 0.6815,
"step": 74
},
{
"epoch": 0.07564296520423601,
"grad_norm": 0.7391223311424255,
"learning_rate": 6.590789473684212e-05,
"loss": 0.6689,
"step": 75
},
{
"epoch": 0.07665153807362582,
"grad_norm": 0.6848605871200562,
"learning_rate": 6.538063157894737e-05,
"loss": 0.5782,
"step": 76
},
{
"epoch": 0.07766011094301563,
"grad_norm": 0.711951732635498,
"learning_rate": 6.485336842105264e-05,
"loss": 0.5801,
"step": 77
},
{
"epoch": 0.07866868381240545,
"grad_norm": 0.5852221846580505,
"learning_rate": 6.43261052631579e-05,
"loss": 0.4995,
"step": 78
},
{
"epoch": 0.07967725668179526,
"grad_norm": 0.9040377736091614,
"learning_rate": 6.379884210526315e-05,
"loss": 0.5318,
"step": 79
},
{
"epoch": 0.08068582955118507,
"grad_norm": 0.756641149520874,
"learning_rate": 6.327157894736842e-05,
"loss": 0.6616,
"step": 80
},
{
"epoch": 0.08169440242057488,
"grad_norm": 0.7158970832824707,
"learning_rate": 6.274431578947368e-05,
"loss": 0.6113,
"step": 81
},
{
"epoch": 0.0827029752899647,
"grad_norm": 0.6931173205375671,
"learning_rate": 6.221705263157895e-05,
"loss": 0.5882,
"step": 82
},
{
"epoch": 0.08371154815935451,
"grad_norm": 0.7512569427490234,
"learning_rate": 6.168978947368422e-05,
"loss": 0.6451,
"step": 83
},
{
"epoch": 0.08472012102874432,
"grad_norm": 0.7793223261833191,
"learning_rate": 6.116252631578948e-05,
"loss": 0.6186,
"step": 84
},
{
"epoch": 0.08572869389813415,
"grad_norm": 0.7280353307723999,
"learning_rate": 6.063526315789474e-05,
"loss": 0.6657,
"step": 85
},
{
"epoch": 0.08673726676752395,
"grad_norm": 0.7213184833526611,
"learning_rate": 6.0108e-05,
"loss": 0.6312,
"step": 86
},
{
"epoch": 0.08774583963691376,
"grad_norm": 0.7809893488883972,
"learning_rate": 5.9580736842105264e-05,
"loss": 0.6928,
"step": 87
},
{
"epoch": 0.08875441250630359,
"grad_norm": 0.8184586763381958,
"learning_rate": 5.905347368421053e-05,
"loss": 0.6551,
"step": 88
},
{
"epoch": 0.0897629853756934,
"grad_norm": 0.8529835939407349,
"learning_rate": 5.85262105263158e-05,
"loss": 0.7481,
"step": 89
},
{
"epoch": 0.0907715582450832,
"grad_norm": 0.8931265473365784,
"learning_rate": 5.799894736842106e-05,
"loss": 0.7928,
"step": 90
},
{
"epoch": 0.09178013111447302,
"grad_norm": 0.7958138585090637,
"learning_rate": 5.747168421052632e-05,
"loss": 0.6151,
"step": 91
},
{
"epoch": 0.09278870398386284,
"grad_norm": 0.7931579351425171,
"learning_rate": 5.694442105263158e-05,
"loss": 0.7141,
"step": 92
},
{
"epoch": 0.09379727685325265,
"grad_norm": 0.8281200528144836,
"learning_rate": 5.641715789473684e-05,
"loss": 0.6538,
"step": 93
},
{
"epoch": 0.09480584972264246,
"grad_norm": 0.8042008280754089,
"learning_rate": 5.5889894736842104e-05,
"loss": 0.6661,
"step": 94
},
{
"epoch": 0.09581442259203228,
"grad_norm": 0.8040047287940979,
"learning_rate": 5.5362631578947374e-05,
"loss": 0.5659,
"step": 95
},
{
"epoch": 0.09682299546142209,
"grad_norm": 1.044537901878357,
"learning_rate": 5.483536842105264e-05,
"loss": 0.835,
"step": 96
},
{
"epoch": 0.0978315683308119,
"grad_norm": 0.9413779377937317,
"learning_rate": 5.43081052631579e-05,
"loss": 0.7605,
"step": 97
},
{
"epoch": 0.09884014120020171,
"grad_norm": 0.9656868577003479,
"learning_rate": 5.378084210526316e-05,
"loss": 0.7144,
"step": 98
},
{
"epoch": 0.09984871406959153,
"grad_norm": 1.0694280862808228,
"learning_rate": 5.3253578947368426e-05,
"loss": 0.8615,
"step": 99
},
{
"epoch": 0.10085728693898134,
"grad_norm": 1.2606571912765503,
"learning_rate": 5.272631578947368e-05,
"loss": 0.8549,
"step": 100
},
{
"epoch": 0.10085728693898134,
"eval_loss": 0.635071337223053,
"eval_runtime": 47.0624,
"eval_samples_per_second": 8.882,
"eval_steps_per_second": 2.231,
"step": 100
},
{
"epoch": 0.10186585980837115,
"grad_norm": 0.7055741548538208,
"learning_rate": 5.2199052631578945e-05,
"loss": 0.6355,
"step": 101
},
{
"epoch": 0.10287443267776097,
"grad_norm": 0.7601715922355652,
"learning_rate": 5.167178947368422e-05,
"loss": 0.6535,
"step": 102
},
{
"epoch": 0.10388300554715078,
"grad_norm": 0.6638849973678589,
"learning_rate": 5.114452631578948e-05,
"loss": 0.6594,
"step": 103
},
{
"epoch": 0.10489157841654059,
"grad_norm": 0.5764726996421814,
"learning_rate": 5.061726315789474e-05,
"loss": 0.5227,
"step": 104
},
{
"epoch": 0.1059001512859304,
"grad_norm": 0.7011322975158691,
"learning_rate": 5.009e-05,
"loss": 0.6485,
"step": 105
},
{
"epoch": 0.10690872415532023,
"grad_norm": 0.6385596394538879,
"learning_rate": 4.9562736842105266e-05,
"loss": 0.617,
"step": 106
},
{
"epoch": 0.10791729702471003,
"grad_norm": 0.613409161567688,
"learning_rate": 4.903547368421053e-05,
"loss": 0.5565,
"step": 107
},
{
"epoch": 0.10892586989409984,
"grad_norm": 0.5934216976165771,
"learning_rate": 4.850821052631579e-05,
"loss": 0.5369,
"step": 108
},
{
"epoch": 0.10993444276348967,
"grad_norm": 0.5698984265327454,
"learning_rate": 4.7980947368421055e-05,
"loss": 0.5095,
"step": 109
},
{
"epoch": 0.11094301563287948,
"grad_norm": 0.6158843636512756,
"learning_rate": 4.745368421052632e-05,
"loss": 0.5607,
"step": 110
},
{
"epoch": 0.11195158850226929,
"grad_norm": 0.6479724645614624,
"learning_rate": 4.692642105263158e-05,
"loss": 0.4708,
"step": 111
},
{
"epoch": 0.1129601613716591,
"grad_norm": 0.7158260345458984,
"learning_rate": 4.6399157894736844e-05,
"loss": 0.6306,
"step": 112
},
{
"epoch": 0.11396873424104892,
"grad_norm": 0.6772667169570923,
"learning_rate": 4.5871894736842107e-05,
"loss": 0.5333,
"step": 113
},
{
"epoch": 0.11497730711043873,
"grad_norm": 0.6242232322692871,
"learning_rate": 4.5344631578947376e-05,
"loss": 0.5704,
"step": 114
},
{
"epoch": 0.11598587997982854,
"grad_norm": 0.6591203212738037,
"learning_rate": 4.481736842105263e-05,
"loss": 0.665,
"step": 115
},
{
"epoch": 0.11699445284921836,
"grad_norm": 0.6875463128089905,
"learning_rate": 4.4290105263157895e-05,
"loss": 0.546,
"step": 116
},
{
"epoch": 0.11800302571860817,
"grad_norm": 0.7117153406143188,
"learning_rate": 4.3762842105263165e-05,
"loss": 0.6204,
"step": 117
},
{
"epoch": 0.11901159858799798,
"grad_norm": 0.6039959192276001,
"learning_rate": 4.323557894736843e-05,
"loss": 0.635,
"step": 118
},
{
"epoch": 0.1200201714573878,
"grad_norm": 0.7009987235069275,
"learning_rate": 4.2708315789473684e-05,
"loss": 0.608,
"step": 119
},
{
"epoch": 0.12102874432677761,
"grad_norm": 0.6250441074371338,
"learning_rate": 4.218105263157895e-05,
"loss": 0.5882,
"step": 120
},
{
"epoch": 0.12203731719616742,
"grad_norm": 0.7398062944412231,
"learning_rate": 4.1653789473684217e-05,
"loss": 0.6349,
"step": 121
},
{
"epoch": 0.12304589006555723,
"grad_norm": 0.7030249834060669,
"learning_rate": 4.112652631578948e-05,
"loss": 0.6066,
"step": 122
},
{
"epoch": 0.12405446293494705,
"grad_norm": 0.7045819163322449,
"learning_rate": 4.0599263157894736e-05,
"loss": 0.6708,
"step": 123
},
{
"epoch": 0.12506303580433686,
"grad_norm": 0.6833453178405762,
"learning_rate": 4.0072000000000005e-05,
"loss": 0.6002,
"step": 124
},
{
"epoch": 0.1260716086737267,
"grad_norm": 0.8382267951965332,
"learning_rate": 3.954473684210527e-05,
"loss": 0.6854,
"step": 125
},
{
"epoch": 0.12708018154311648,
"grad_norm": 0.8998156189918518,
"learning_rate": 3.901747368421053e-05,
"loss": 0.5236,
"step": 126
},
{
"epoch": 0.1280887544125063,
"grad_norm": 0.6268869638442993,
"learning_rate": 3.849021052631579e-05,
"loss": 0.5159,
"step": 127
},
{
"epoch": 0.12909732728189613,
"grad_norm": 0.7306240797042847,
"learning_rate": 3.796294736842106e-05,
"loss": 0.6468,
"step": 128
},
{
"epoch": 0.13010590015128592,
"grad_norm": 0.696013331413269,
"learning_rate": 3.743568421052632e-05,
"loss": 0.5668,
"step": 129
},
{
"epoch": 0.13111447302067575,
"grad_norm": 0.6369208693504333,
"learning_rate": 3.6908421052631576e-05,
"loss": 0.5423,
"step": 130
},
{
"epoch": 0.13212304589006554,
"grad_norm": 0.711659848690033,
"learning_rate": 3.6381157894736846e-05,
"loss": 0.5699,
"step": 131
},
{
"epoch": 0.13313161875945537,
"grad_norm": 0.6578086614608765,
"learning_rate": 3.585389473684211e-05,
"loss": 0.4694,
"step": 132
},
{
"epoch": 0.1341401916288452,
"grad_norm": 0.7941672205924988,
"learning_rate": 3.532663157894737e-05,
"loss": 0.6442,
"step": 133
},
{
"epoch": 0.13514876449823499,
"grad_norm": 0.7573974132537842,
"learning_rate": 3.4799368421052634e-05,
"loss": 0.7441,
"step": 134
},
{
"epoch": 0.1361573373676248,
"grad_norm": 0.7425283789634705,
"learning_rate": 3.42721052631579e-05,
"loss": 0.6675,
"step": 135
},
{
"epoch": 0.13716591023701463,
"grad_norm": 0.6818085312843323,
"learning_rate": 3.374484210526316e-05,
"loss": 0.5508,
"step": 136
},
{
"epoch": 0.13817448310640443,
"grad_norm": 0.729345440864563,
"learning_rate": 3.321757894736842e-05,
"loss": 0.5426,
"step": 137
},
{
"epoch": 0.13918305597579425,
"grad_norm": 0.6872043013572693,
"learning_rate": 3.2690315789473686e-05,
"loss": 0.6318,
"step": 138
},
{
"epoch": 0.14019162884518407,
"grad_norm": 0.8204219937324524,
"learning_rate": 3.216305263157895e-05,
"loss": 0.6653,
"step": 139
},
{
"epoch": 0.14120020171457387,
"grad_norm": 0.788364827632904,
"learning_rate": 3.163578947368421e-05,
"loss": 0.6192,
"step": 140
},
{
"epoch": 0.1422087745839637,
"grad_norm": 0.7906739115715027,
"learning_rate": 3.1108526315789475e-05,
"loss": 0.6494,
"step": 141
},
{
"epoch": 0.14321734745335352,
"grad_norm": 0.8248524069786072,
"learning_rate": 3.058126315789474e-05,
"loss": 0.6254,
"step": 142
},
{
"epoch": 0.1442259203227433,
"grad_norm": 0.7894427180290222,
"learning_rate": 3.0054e-05,
"loss": 0.6097,
"step": 143
},
{
"epoch": 0.14523449319213314,
"grad_norm": 0.8506288528442383,
"learning_rate": 2.9526736842105263e-05,
"loss": 0.6809,
"step": 144
},
{
"epoch": 0.14624306606152296,
"grad_norm": 1.004377841949463,
"learning_rate": 2.899947368421053e-05,
"loss": 0.7994,
"step": 145
},
{
"epoch": 0.14725163893091275,
"grad_norm": 0.9447079300880432,
"learning_rate": 2.847221052631579e-05,
"loss": 0.7159,
"step": 146
},
{
"epoch": 0.14826021180030258,
"grad_norm": 0.9489012360572815,
"learning_rate": 2.7944947368421052e-05,
"loss": 0.7263,
"step": 147
},
{
"epoch": 0.14926878466969237,
"grad_norm": 0.9614557027816772,
"learning_rate": 2.741768421052632e-05,
"loss": 0.6547,
"step": 148
},
{
"epoch": 0.1502773575390822,
"grad_norm": 0.98894202709198,
"learning_rate": 2.689042105263158e-05,
"loss": 0.7373,
"step": 149
},
{
"epoch": 0.15128593040847202,
"grad_norm": 1.164485216140747,
"learning_rate": 2.636315789473684e-05,
"loss": 0.7158,
"step": 150
},
{
"epoch": 0.15128593040847202,
"eval_loss": 0.6001591086387634,
"eval_runtime": 47.5174,
"eval_samples_per_second": 8.797,
"eval_steps_per_second": 2.21,
"step": 150
}
],
"logging_steps": 1,
"max_steps": 200,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 50,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 5,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.2972871159853875e+17,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}