rakhman-llm commited on
Commit
984eb7f
·
verified ·
1 Parent(s): 8a15d77

Training in progress, step 18500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c271689f6760d4359c2a43653eb4d2901456cbdb4b87714c22ec9cdaa2c8b8e
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfa8b32014bbb1116102bb096f351b477c375f02a76f56941cd43b6b2b8c9ae0
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b2573858366c51da080dc52026b5cf706984e6769263fae62f35276ad423d25
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24c06d2ce08030ea9353adcf0c618c2748ba4afafcd7e2ef1fa5088554e20156
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:265f7ddb321b2fa01d3a22602c4d8e894508ea8d4458f230512f39344ac67fcf
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40162c830f5df6b739cc1f24fd1ff7e4f55f1a27766e6deeff7911f41b300f3c
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:944535d8118858ad045305f7f05f43c95ec4aaef08aedf918a1fec60acb21f30
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b23d12d924a2c5ac71a2463338c282c759dd6e8f289f7165dd510cc1f6cd61fa
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.2,
5
  "eval_steps": 500,
6
- "global_step": 18000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1275,6 +1275,41 @@
1275
  "learning_rate": 1.2002222222222225e-05,
1276
  "loss": 0.0591,
1277
  "step": 18000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1278
  }
1279
  ],
1280
  "logging_steps": 100,
@@ -1294,7 +1329,7 @@
1294
  "attributes": {}
1295
  }
1296
  },
1297
- "total_flos": 4.384496812032e+16,
1298
  "train_batch_size": 4,
1299
  "trial_name": null,
1300
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.2333333333333334,
5
  "eval_steps": 500,
6
+ "global_step": 18500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1275
  "learning_rate": 1.2002222222222225e-05,
1276
  "loss": 0.0591,
1277
  "step": 18000
1278
+ },
1279
+ {
1280
+ "epoch": 1.2066666666666666,
1281
+ "grad_norm": 0.1682252287864685,
1282
+ "learning_rate": 1.195777777777778e-05,
1283
+ "loss": 0.0613,
1284
+ "step": 18100
1285
+ },
1286
+ {
1287
+ "epoch": 1.2133333333333334,
1288
+ "grad_norm": 0.11706658452749252,
1289
+ "learning_rate": 1.1913333333333335e-05,
1290
+ "loss": 0.0676,
1291
+ "step": 18200
1292
+ },
1293
+ {
1294
+ "epoch": 1.22,
1295
+ "grad_norm": 0.22997494041919708,
1296
+ "learning_rate": 1.186888888888889e-05,
1297
+ "loss": 0.0576,
1298
+ "step": 18300
1299
+ },
1300
+ {
1301
+ "epoch": 1.2266666666666666,
1302
+ "grad_norm": 0.2240549921989441,
1303
+ "learning_rate": 1.1824444444444445e-05,
1304
+ "loss": 0.0532,
1305
+ "step": 18400
1306
+ },
1307
+ {
1308
+ "epoch": 1.2333333333333334,
1309
+ "grad_norm": 0.10644257068634033,
1310
+ "learning_rate": 1.178e-05,
1311
+ "loss": 0.0657,
1312
+ "step": 18500
1313
  }
1314
  ],
1315
  "logging_steps": 100,
 
1329
  "attributes": {}
1330
  }
1331
  },
1332
+ "total_flos": 4.506288390144e+16,
1333
  "train_batch_size": 4,
1334
  "trial_name": null,
1335
  "trial_params": null