rakhman-llm commited on
Commit
efa7941
·
verified ·
1 Parent(s): 6d46f60

Training in progress, step 17500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:688c3b1d28f6b863bee92004557adc364c4b2e5ee4097d484ed3aaba40d50c14
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e12d395d228986d4722c9bfdd3b7c1a0c87813d6c1948380328534f75ad454cd
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a993357ca1b551b4a094ee2409999ca1a4657c3428fde811fa9954bc1b8b645
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad3401b5dfde1c4ac0b81be604349a5d0c27404fba8d352372bd54950216bb24
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8fa477eb7529f96adee96884288f62ffdf70dbb1416c416ce42fdc1eba5e518
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f87436d1ed788958cb08087900b45d6f1fbde14c0d7db3fa0e7c937f406062a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:57233dc32b29f0dc0010f99ddb975758e3cef5e747558908f115a1d09e80f707
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d166d24f789387f6695cee2a14320335faa00c9fa99f8038d6fcfa9ae735c9d
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.1333333333333333,
5
  "eval_steps": 500,
6
- "global_step": 17000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1205,6 +1205,41 @@
1205
  "learning_rate": 1.2446666666666667e-05,
1206
  "loss": 0.058,
1207
  "step": 17000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1208
  }
1209
  ],
1210
  "logging_steps": 100,
@@ -1224,7 +1259,7 @@
1224
  "attributes": {}
1225
  }
1226
  },
1227
- "total_flos": 4.140913655808e+16,
1228
  "train_batch_size": 4,
1229
  "trial_name": null,
1230
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.1666666666666667,
5
  "eval_steps": 500,
6
+ "global_step": 17500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1205
  "learning_rate": 1.2446666666666667e-05,
1206
  "loss": 0.058,
1207
  "step": 17000
1208
+ },
1209
+ {
1210
+ "epoch": 1.1400000000000001,
1211
+ "grad_norm": 0.26193323731422424,
1212
+ "learning_rate": 1.2402222222222222e-05,
1213
+ "loss": 0.0606,
1214
+ "step": 17100
1215
+ },
1216
+ {
1217
+ "epoch": 1.1466666666666667,
1218
+ "grad_norm": 0.18597596883773804,
1219
+ "learning_rate": 1.235777777777778e-05,
1220
+ "loss": 0.0565,
1221
+ "step": 17200
1222
+ },
1223
+ {
1224
+ "epoch": 1.1533333333333333,
1225
+ "grad_norm": 0.27906695008277893,
1226
+ "learning_rate": 1.2313333333333335e-05,
1227
+ "loss": 0.0616,
1228
+ "step": 17300
1229
+ },
1230
+ {
1231
+ "epoch": 1.16,
1232
+ "grad_norm": 0.20379580557346344,
1233
+ "learning_rate": 1.226888888888889e-05,
1234
+ "loss": 0.0572,
1235
+ "step": 17400
1236
+ },
1237
+ {
1238
+ "epoch": 1.1666666666666667,
1239
+ "grad_norm": 0.2315666228532791,
1240
+ "learning_rate": 1.2224444444444445e-05,
1241
+ "loss": 0.0571,
1242
+ "step": 17500
1243
  }
1244
  ],
1245
  "logging_steps": 100,
 
1259
  "attributes": {}
1260
  }
1261
  },
1262
+ "total_flos": 4.26270523392e+16,
1263
  "train_batch_size": 4,
1264
  "trial_name": null,
1265
  "trial_params": null