rakhman-llm commited on
Commit
306e8db
·
verified ·
1 Parent(s): 99714ec

Training in progress, step 20500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af54bf45b4ea734175c8bd735e147b83bfa4467a6a02927fce71013f50f2e031
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c86189a77f037cbdd789ce6756a983f35d8424562557bec70f1d6f9478782956
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91db2977a4e8b2ac56d4d9528647fa170313a74ed3d096d47e9525e450940a2b
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f3b9b0a558aef18a5c43688fb84bceaf30915a8e0f42a94e913404b25ec6988
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6d5d4da7add6dcbf76add6d041d8a356430f2fcdcb5e01b3a00941303e2c6c7d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1cb24b28fd79d1b2e08e97a1214134b31144058db4370cb477a455701ee578a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92a26e97f6adf463a9a2328a07aa990cecfc61051330c04894ef3e783375a50f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02d6f63efc0f911cd45ec87b983b2cf047da0e86188af05e21eda256bea09b7e
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.3333333333333333,
5
  "eval_steps": 500,
6
- "global_step": 20000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1415,6 +1415,41 @@
1415
  "learning_rate": 1.1113777777777779e-05,
1416
  "loss": 0.0672,
1417
  "step": 20000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1418
  }
1419
  ],
1420
  "logging_steps": 100,
@@ -1434,7 +1469,7 @@
1434
  "attributes": {}
1435
  }
1436
  },
1437
- "total_flos": 4.87166312448e+16,
1438
  "train_batch_size": 4,
1439
  "trial_name": null,
1440
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.3666666666666667,
5
  "eval_steps": 500,
6
+ "global_step": 20500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1415
  "learning_rate": 1.1113777777777779e-05,
1416
  "loss": 0.0672,
1417
  "step": 20000
1418
+ },
1419
+ {
1420
+ "epoch": 1.34,
1421
+ "grad_norm": 0.19268549978733063,
1422
+ "learning_rate": 1.1069333333333334e-05,
1423
+ "loss": 0.0577,
1424
+ "step": 20100
1425
+ },
1426
+ {
1427
+ "epoch": 1.3466666666666667,
1428
+ "grad_norm": 0.24096272885799408,
1429
+ "learning_rate": 1.102488888888889e-05,
1430
+ "loss": 0.0597,
1431
+ "step": 20200
1432
+ },
1433
+ {
1434
+ "epoch": 1.3533333333333333,
1435
+ "grad_norm": 0.3714766204357147,
1436
+ "learning_rate": 1.0980444444444445e-05,
1437
+ "loss": 0.0602,
1438
+ "step": 20300
1439
+ },
1440
+ {
1441
+ "epoch": 1.3599999999999999,
1442
+ "grad_norm": 0.15587754547595978,
1443
+ "learning_rate": 1.0936e-05,
1444
+ "loss": 0.056,
1445
+ "step": 20400
1446
+ },
1447
+ {
1448
+ "epoch": 1.3666666666666667,
1449
+ "grad_norm": 0.21720756590366364,
1450
+ "learning_rate": 1.0891555555555557e-05,
1451
+ "loss": 0.0609,
1452
+ "step": 20500
1453
  }
1454
  ],
1455
  "logging_steps": 100,
 
1469
  "attributes": {}
1470
  }
1471
  },
1472
+ "total_flos": 4.993454702592e+16,
1473
  "train_batch_size": 4,
1474
  "trial_name": null,
1475
  "trial_params": null