Yhhhhhhhhh's picture
Upload folder using huggingface_hub
7072302 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.0,
"eval_steps": 50000,
"global_step": 928,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.04310344827586207,
"grad_norm": 5.7193193435668945,
"learning_rate": 1.7857142857142859e-06,
"loss": 0.7126,
"step": 10
},
{
"epoch": 0.08620689655172414,
"grad_norm": 2.6962673664093018,
"learning_rate": 3.5714285714285718e-06,
"loss": 0.6662,
"step": 20
},
{
"epoch": 0.12931034482758622,
"grad_norm": 1.6143770217895508,
"learning_rate": 4.999939076763487e-06,
"loss": 0.5303,
"step": 30
},
{
"epoch": 0.1724137931034483,
"grad_norm": 1.3705445528030396,
"learning_rate": 4.997807075247147e-06,
"loss": 0.453,
"step": 40
},
{
"epoch": 0.21551724137931033,
"grad_norm": 1.129184603691101,
"learning_rate": 4.992631880567301e-06,
"loss": 0.4192,
"step": 50
},
{
"epoch": 0.25862068965517243,
"grad_norm": 1.3925697803497314,
"learning_rate": 4.984419797901491e-06,
"loss": 0.418,
"step": 60
},
{
"epoch": 0.3017241379310345,
"grad_norm": 1.4394880533218384,
"learning_rate": 4.973180832407471e-06,
"loss": 0.4118,
"step": 70
},
{
"epoch": 0.3448275862068966,
"grad_norm": 1.3788193464279175,
"learning_rate": 4.958928677033465e-06,
"loss": 0.4065,
"step": 80
},
{
"epoch": 0.3879310344827586,
"grad_norm": 1.1382204294204712,
"learning_rate": 4.9416806958354206e-06,
"loss": 0.3819,
"step": 90
},
{
"epoch": 0.43103448275862066,
"grad_norm": 1.081235647201538,
"learning_rate": 4.921457902821578e-06,
"loss": 0.399,
"step": 100
},
{
"epoch": 0.47413793103448276,
"grad_norm": 1.21293044090271,
"learning_rate": 4.898284936350144e-06,
"loss": 0.3962,
"step": 110
},
{
"epoch": 0.5172413793103449,
"grad_norm": 1.3453164100646973,
"learning_rate": 4.8721900291112415e-06,
"loss": 0.3818,
"step": 120
},
{
"epoch": 0.5603448275862069,
"grad_norm": 1.2487531900405884,
"learning_rate": 4.84320497372973e-06,
"loss": 0.3901,
"step": 130
},
{
"epoch": 0.603448275862069,
"grad_norm": 1.1610265970230103,
"learning_rate": 4.811365084030784e-06,
"loss": 0.3867,
"step": 140
},
{
"epoch": 0.646551724137931,
"grad_norm": 1.0656601190567017,
"learning_rate": 4.776709152015443e-06,
"loss": 0.3882,
"step": 150
},
{
"epoch": 0.6896551724137931,
"grad_norm": 1.2680625915527344,
"learning_rate": 4.7392794005985324e-06,
"loss": 0.3739,
"step": 160
},
{
"epoch": 0.7327586206896551,
"grad_norm": 1.3623387813568115,
"learning_rate": 4.699121432166542e-06,
"loss": 0.3733,
"step": 170
},
{
"epoch": 0.7758620689655172,
"grad_norm": 1.096368432044983,
"learning_rate": 4.656284173018144e-06,
"loss": 0.3625,
"step": 180
},
{
"epoch": 0.8189655172413793,
"grad_norm": 0.9958175420761108,
"learning_rate": 4.610819813755038e-06,
"loss": 0.3698,
"step": 190
},
{
"epoch": 0.8620689655172413,
"grad_norm": 1.023909091949463,
"learning_rate": 4.562783745695738e-06,
"loss": 0.3471,
"step": 200
},
{
"epoch": 0.9051724137931034,
"grad_norm": 1.061679482460022,
"learning_rate": 4.512234493389785e-06,
"loss": 0.3712,
"step": 210
},
{
"epoch": 0.9482758620689655,
"grad_norm": 1.3929497003555298,
"learning_rate": 4.4592336433146e-06,
"loss": 0.3651,
"step": 220
},
{
"epoch": 0.9913793103448276,
"grad_norm": 0.8806857466697693,
"learning_rate": 4.403845768841842e-06,
"loss": 0.3641,
"step": 230
},
{
"epoch": 1.0344827586206897,
"grad_norm": 0.9649218916893005,
"learning_rate": 4.346138351564711e-06,
"loss": 0.3057,
"step": 240
},
{
"epoch": 1.0775862068965518,
"grad_norm": 1.1328972578048706,
"learning_rate": 4.286181699082008e-06,
"loss": 0.3045,
"step": 250
},
{
"epoch": 1.1206896551724137,
"grad_norm": 1.1146079301834106,
"learning_rate": 4.224048859339175e-06,
"loss": 0.301,
"step": 260
},
{
"epoch": 1.1637931034482758,
"grad_norm": 1.1736148595809937,
"learning_rate": 4.159815531630604e-06,
"loss": 0.3236,
"step": 270
},
{
"epoch": 1.206896551724138,
"grad_norm": 1.1570653915405273,
"learning_rate": 4.093559974371725e-06,
"loss": 0.3078,
"step": 280
},
{
"epoch": 1.25,
"grad_norm": 1.335072636604309,
"learning_rate": 4.02536290975317e-06,
"loss": 0.3145,
"step": 290
},
{
"epoch": 1.293103448275862,
"grad_norm": 1.0565122365951538,
"learning_rate": 3.955307425393224e-06,
"loss": 0.3018,
"step": 300
},
{
"epoch": 1.3362068965517242,
"grad_norm": 1.3002219200134277,
"learning_rate": 3.88347887310836e-06,
"loss": 0.3215,
"step": 310
},
{
"epoch": 1.3793103448275863,
"grad_norm": 1.0199170112609863,
"learning_rate": 3.8099647649251984e-06,
"loss": 0.2973,
"step": 320
},
{
"epoch": 1.4224137931034484,
"grad_norm": 1.118922233581543,
"learning_rate": 3.7348546664605777e-06,
"loss": 0.3095,
"step": 330
},
{
"epoch": 1.4655172413793103,
"grad_norm": 1.080239176750183,
"learning_rate": 3.658240087799655e-06,
"loss": 0.3172,
"step": 340
},
{
"epoch": 1.5086206896551724,
"grad_norm": 0.9963559508323669,
"learning_rate": 3.5802143720049565e-06,
"loss": 0.2923,
"step": 350
},
{
"epoch": 1.5517241379310345,
"grad_norm": 1.1155935525894165,
"learning_rate": 3.5008725813922383e-06,
"loss": 0.3032,
"step": 360
},
{
"epoch": 1.5948275862068966,
"grad_norm": 1.4892568588256836,
"learning_rate": 3.4203113817116955e-06,
"loss": 0.3113,
"step": 370
},
{
"epoch": 1.6379310344827587,
"grad_norm": 1.1873013973236084,
"learning_rate": 3.338628924375638e-06,
"loss": 0.3122,
"step": 380
},
{
"epoch": 1.6810344827586206,
"grad_norm": 1.215104103088379,
"learning_rate": 3.2559247268761117e-06,
"loss": 0.2986,
"step": 390
},
{
"epoch": 1.7241379310344827,
"grad_norm": 1.2356464862823486,
"learning_rate": 3.1722995515381644e-06,
"loss": 0.2969,
"step": 400
},
{
"epoch": 1.7672413793103448,
"grad_norm": 1.217281699180603,
"learning_rate": 3.087855282756475e-06,
"loss": 0.294,
"step": 410
},
{
"epoch": 1.8103448275862069,
"grad_norm": 1.3647332191467285,
"learning_rate": 3.002694802864912e-06,
"loss": 0.2987,
"step": 420
},
{
"epoch": 1.853448275862069,
"grad_norm": 1.295482873916626,
"learning_rate": 2.9169218667902562e-06,
"loss": 0.3087,
"step": 430
},
{
"epoch": 1.896551724137931,
"grad_norm": 1.1709827184677124,
"learning_rate": 2.8306409756428067e-06,
"loss": 0.2945,
"step": 440
},
{
"epoch": 1.9396551724137931,
"grad_norm": 1.2571437358856201,
"learning_rate": 2.743957249397874e-06,
"loss": 0.3036,
"step": 450
},
{
"epoch": 1.9827586206896552,
"grad_norm": 1.7302639484405518,
"learning_rate": 2.6569762988232838e-06,
"loss": 0.3119,
"step": 460
},
{
"epoch": 2.0258620689655173,
"grad_norm": 0.9986317157745361,
"learning_rate": 2.569804096808923e-06,
"loss": 0.2789,
"step": 470
},
{
"epoch": 2.0689655172413794,
"grad_norm": 1.130508303642273,
"learning_rate": 2.482546849255096e-06,
"loss": 0.2435,
"step": 480
},
{
"epoch": 2.1120689655172415,
"grad_norm": 1.446877121925354,
"learning_rate": 2.3953108656770018e-06,
"loss": 0.2475,
"step": 490
},
{
"epoch": 2.1551724137931036,
"grad_norm": 1.60965096950531,
"learning_rate": 2.3082024296829538e-06,
"loss": 0.2384,
"step": 500
},
{
"epoch": 2.1982758620689653,
"grad_norm": 1.383519172668457,
"learning_rate": 2.2213276694841866e-06,
"loss": 0.2356,
"step": 510
},
{
"epoch": 2.2413793103448274,
"grad_norm": 1.3211714029312134,
"learning_rate": 2.134792428593971e-06,
"loss": 0.2466,
"step": 520
},
{
"epoch": 2.2844827586206895,
"grad_norm": 1.302563190460205,
"learning_rate": 2.0487021368736002e-06,
"loss": 0.2456,
"step": 530
},
{
"epoch": 2.3275862068965516,
"grad_norm": 1.30355703830719,
"learning_rate": 1.963161682082342e-06,
"loss": 0.2406,
"step": 540
},
{
"epoch": 2.3706896551724137,
"grad_norm": 1.1259819269180298,
"learning_rate": 1.8782752820878636e-06,
"loss": 0.2225,
"step": 550
},
{
"epoch": 2.413793103448276,
"grad_norm": 1.2397651672363281,
"learning_rate": 1.7941463578928088e-06,
"loss": 0.2312,
"step": 560
},
{
"epoch": 2.456896551724138,
"grad_norm": 1.3761606216430664,
"learning_rate": 1.7108774076322443e-06,
"loss": 0.2415,
"step": 570
},
{
"epoch": 2.5,
"grad_norm": 1.3666378259658813,
"learning_rate": 1.6285698816954626e-06,
"loss": 0.2487,
"step": 580
},
{
"epoch": 2.543103448275862,
"grad_norm": 1.183164119720459,
"learning_rate": 1.547324059124315e-06,
"loss": 0.2378,
"step": 590
},
{
"epoch": 2.586206896551724,
"grad_norm": 1.4499019384384155,
"learning_rate": 1.467238925438646e-06,
"loss": 0.2765,
"step": 600
},
{
"epoch": 2.6293103448275863,
"grad_norm": 1.447636604309082,
"learning_rate": 1.388412052037682e-06,
"loss": 0.2584,
"step": 610
},
{
"epoch": 2.6724137931034484,
"grad_norm": 1.449857473373413,
"learning_rate": 1.3109394773243117e-06,
"loss": 0.2323,
"step": 620
},
{
"epoch": 2.7155172413793105,
"grad_norm": 1.3404827117919922,
"learning_rate": 1.234915589697091e-06,
"loss": 0.2384,
"step": 630
},
{
"epoch": 2.7586206896551726,
"grad_norm": 1.3379237651824951,
"learning_rate": 1.160433012552508e-06,
"loss": 0.2382,
"step": 640
},
{
"epoch": 2.8017241379310347,
"grad_norm": 1.4709217548370361,
"learning_rate": 1.0875824914376555e-06,
"loss": 0.2303,
"step": 650
},
{
"epoch": 2.844827586206897,
"grad_norm": 1.3744317293167114,
"learning_rate": 1.0164527834907468e-06,
"loss": 0.2423,
"step": 660
},
{
"epoch": 2.887931034482759,
"grad_norm": 1.7406655550003052,
"learning_rate": 9.471305493042243e-07,
"loss": 0.2426,
"step": 670
},
{
"epoch": 2.9310344827586206,
"grad_norm": 1.6748679876327515,
"learning_rate": 8.797002473421729e-07,
"loss": 0.2494,
"step": 680
},
{
"epoch": 2.9741379310344827,
"grad_norm": 1.5766329765319824,
"learning_rate": 8.142440310406923e-07,
"loss": 0.2407,
"step": 690
},
{
"epoch": 3.0172413793103448,
"grad_norm": 1.2436530590057373,
"learning_rate": 7.508416487165862e-07,
"loss": 0.2254,
"step": 700
},
{
"epoch": 3.060344827586207,
"grad_norm": 1.3559637069702148,
"learning_rate": 6.895703464063319e-07,
"loss": 0.2015,
"step": 710
},
{
"epoch": 3.103448275862069,
"grad_norm": 1.3423281908035278,
"learning_rate": 6.305047737536707e-07,
"loss": 0.1952,
"step": 720
},
{
"epoch": 3.146551724137931,
"grad_norm": 1.4924376010894775,
"learning_rate": 5.737168930605272e-07,
"loss": 0.1967,
"step": 730
},
{
"epoch": 3.189655172413793,
"grad_norm": 1.4202907085418701,
"learning_rate": 5.192758916120236e-07,
"loss": 0.2128,
"step": 740
},
{
"epoch": 3.2327586206896552,
"grad_norm": 1.252336025238037,
"learning_rate": 4.672480973824312e-07,
"loss": 0.1907,
"step": 750
},
{
"epoch": 3.2758620689655173,
"grad_norm": 1.7021604776382446,
"learning_rate": 4.1769689822475147e-07,
"loss": 0.2046,
"step": 760
},
{
"epoch": 3.3189655172413794,
"grad_norm": 1.637277364730835,
"learning_rate": 3.7068266464238085e-07,
"loss": 0.1973,
"step": 770
},
{
"epoch": 3.3620689655172415,
"grad_norm": 1.2321523427963257,
"learning_rate": 3.262626762369525e-07,
"loss": 0.2107,
"step": 780
},
{
"epoch": 3.405172413793103,
"grad_norm": 1.3565343618392944,
"learning_rate": 2.844910519219632e-07,
"loss": 0.2055,
"step": 790
},
{
"epoch": 3.4482758620689653,
"grad_norm": 1.6926054954528809,
"learning_rate": 2.454186839872158e-07,
"loss": 0.2083,
"step": 800
},
{
"epoch": 3.4913793103448274,
"grad_norm": 1.5500190258026123,
"learning_rate": 2.0909317609440093e-07,
"loss": 0.1973,
"step": 810
},
{
"epoch": 3.5344827586206895,
"grad_norm": 1.4574564695358276,
"learning_rate": 1.7555878527937164e-07,
"loss": 0.2336,
"step": 820
},
{
"epoch": 3.5775862068965516,
"grad_norm": 1.371783971786499,
"learning_rate": 1.4485636803175828e-07,
"loss": 0.2028,
"step": 830
},
{
"epoch": 3.6206896551724137,
"grad_norm": 1.5117493867874146,
"learning_rate": 1.1702333051763271e-07,
"loss": 0.1897,
"step": 840
},
{
"epoch": 3.663793103448276,
"grad_norm": 1.4221725463867188,
"learning_rate": 9.209358300585474e-08,
"loss": 0.2053,
"step": 850
},
{
"epoch": 3.706896551724138,
"grad_norm": 1.3613110780715942,
"learning_rate": 7.009749855363457e-08,
"loss": 0.2019,
"step": 860
},
{
"epoch": 3.75,
"grad_norm": 1.6522403955459595,
"learning_rate": 5.106187600163987e-08,
"loss": 0.2012,
"step": 870
},
{
"epoch": 3.793103448275862,
"grad_norm": 1.1785835027694702,
"learning_rate": 3.5009907323737826e-08,
"loss": 0.2041,
"step": 880
},
{
"epoch": 3.836206896551724,
"grad_norm": 1.3110618591308594,
"learning_rate": 2.1961149371145795e-08,
"loss": 0.2157,
"step": 890
},
{
"epoch": 3.8793103448275863,
"grad_norm": 1.1076109409332275,
"learning_rate": 1.193150004542204e-08,
"loss": 0.2124,
"step": 900
},
{
"epoch": 3.9224137931034484,
"grad_norm": 1.5730844736099243,
"learning_rate": 4.933178929321103e-09,
"loss": 0.2164,
"step": 910
},
{
"epoch": 3.9655172413793105,
"grad_norm": 1.6986498832702637,
"learning_rate": 9.747123991141193e-10,
"loss": 0.209,
"step": 920
},
{
"epoch": 4.0,
"step": 928,
"total_flos": 1.573815416315904e+17,
"train_loss": 0.29322911118125095,
"train_runtime": 2500.7039,
"train_samples_per_second": 5.936,
"train_steps_per_second": 0.371
}
],
"logging_steps": 10,
"max_steps": 928,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.573815416315904e+17,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}