rakhman-llm commited on
Commit
9cb9560
·
verified ·
1 Parent(s): d69119b

Training in progress, step 18000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e12d395d228986d4722c9bfdd3b7c1a0c87813d6c1948380328534f75ad454cd
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c271689f6760d4359c2a43653eb4d2901456cbdb4b87714c22ec9cdaa2c8b8e
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ad3401b5dfde1c4ac0b81be604349a5d0c27404fba8d352372bd54950216bb24
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b2573858366c51da080dc52026b5cf706984e6769263fae62f35276ad423d25
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f87436d1ed788958cb08087900b45d6f1fbde14c0d7db3fa0e7c937f406062a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:265f7ddb321b2fa01d3a22602c4d8e894508ea8d4458f230512f39344ac67fcf
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d166d24f789387f6695cee2a14320335faa00c9fa99f8038d6fcfa9ae735c9d
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:944535d8118858ad045305f7f05f43c95ec4aaef08aedf918a1fec60acb21f30
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.1666666666666667,
5
  "eval_steps": 500,
6
- "global_step": 17500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1240,6 +1240,41 @@
1240
  "learning_rate": 1.2224444444444445e-05,
1241
  "loss": 0.0571,
1242
  "step": 17500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1243
  }
1244
  ],
1245
  "logging_steps": 100,
@@ -1259,7 +1294,7 @@
1259
  "attributes": {}
1260
  }
1261
  },
1262
- "total_flos": 4.26270523392e+16,
1263
  "train_batch_size": 4,
1264
  "trial_name": null,
1265
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.2,
5
  "eval_steps": 500,
6
+ "global_step": 18000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1240
  "learning_rate": 1.2224444444444445e-05,
1241
  "loss": 0.0571,
1242
  "step": 17500
1243
+ },
1244
+ {
1245
+ "epoch": 1.1733333333333333,
1246
+ "grad_norm": 0.16585813462734222,
1247
+ "learning_rate": 1.218e-05,
1248
+ "loss": 0.0646,
1249
+ "step": 17600
1250
+ },
1251
+ {
1252
+ "epoch": 1.18,
1253
+ "grad_norm": 0.204860657453537,
1254
+ "learning_rate": 1.2135555555555556e-05,
1255
+ "loss": 0.0632,
1256
+ "step": 17700
1257
+ },
1258
+ {
1259
+ "epoch": 1.1866666666666668,
1260
+ "grad_norm": 0.25248029828071594,
1261
+ "learning_rate": 1.209111111111111e-05,
1262
+ "loss": 0.0603,
1263
+ "step": 17800
1264
+ },
1265
+ {
1266
+ "epoch": 1.1933333333333334,
1267
+ "grad_norm": 0.1613234281539917,
1268
+ "learning_rate": 1.204666666666667e-05,
1269
+ "loss": 0.0611,
1270
+ "step": 17900
1271
+ },
1272
+ {
1273
+ "epoch": 1.2,
1274
+ "grad_norm": 0.29864856600761414,
1275
+ "learning_rate": 1.2002222222222225e-05,
1276
+ "loss": 0.0591,
1277
+ "step": 18000
1278
  }
1279
  ],
1280
  "logging_steps": 100,
 
1294
  "attributes": {}
1295
  }
1296
  },
1297
+ "total_flos": 4.384496812032e+16,
1298
  "train_batch_size": 4,
1299
  "trial_name": null,
1300
  "trial_params": null