rakhman-llm commited on
Commit
94f8476
·
verified ·
1 Parent(s): 3bf8a3e

Training in progress, step 24500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34a5bf7d80afb43e227b324a1b6cda7e106c37feb23b0824b268c18e7b3a3d1a
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3030d4ce44a28d2687854bece3d36dd11a2483143dc2362c64bdaf84a148a35c
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd314d98a4869313e973a437be445c914aa71bb8b03a7af5f6f9a32ad34d63bf
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8336f578c2856a2611344b57de4fde614962159ed2c75eff7de5a881717f1aa8
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:727aa5fd6a222ba79327243b6bf0f36bc88c1b723f0949eac077fe52d4306974
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7a44025f64c6f7ad3ba883de4c1f4995d535cdd50931b810e0a7d00e22e5cdc
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:751ad724592c1d67b441302a9338d9d899df51042c6edaf69d5d08369bc3c9df
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d9ad9a351ae61b83248fa17cb95ac2021f5fe42e6699ef47e0f8bde893c34bb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.6,
5
  "eval_steps": 500,
6
- "global_step": 24000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1695,6 +1695,41 @@
1695
  "learning_rate": 9.33688888888889e-06,
1696
  "loss": 0.0581,
1697
  "step": 24000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1698
  }
1699
  ],
1700
  "logging_steps": 100,
@@ -1714,7 +1749,7 @@
1714
  "attributes": {}
1715
  }
1716
  },
1717
- "total_flos": 5.845995749376e+16,
1718
  "train_batch_size": 4,
1719
  "trial_name": null,
1720
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.6333333333333333,
5
  "eval_steps": 500,
6
+ "global_step": 24500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1695
  "learning_rate": 9.33688888888889e-06,
1696
  "loss": 0.0581,
1697
  "step": 24000
1698
+ },
1699
+ {
1700
+ "epoch": 1.6066666666666667,
1701
+ "grad_norm": 0.19345501065254211,
1702
+ "learning_rate": 9.292444444444445e-06,
1703
+ "loss": 0.0587,
1704
+ "step": 24100
1705
+ },
1706
+ {
1707
+ "epoch": 1.6133333333333333,
1708
+ "grad_norm": 0.11753817647695541,
1709
+ "learning_rate": 9.248e-06,
1710
+ "loss": 0.0553,
1711
+ "step": 24200
1712
+ },
1713
+ {
1714
+ "epoch": 1.62,
1715
+ "grad_norm": 0.3054827153682709,
1716
+ "learning_rate": 9.203555555555557e-06,
1717
+ "loss": 0.0552,
1718
+ "step": 24300
1719
+ },
1720
+ {
1721
+ "epoch": 1.6266666666666667,
1722
+ "grad_norm": 0.2914266884326935,
1723
+ "learning_rate": 9.159111111111112e-06,
1724
+ "loss": 0.0534,
1725
+ "step": 24400
1726
+ },
1727
+ {
1728
+ "epoch": 1.6333333333333333,
1729
+ "grad_norm": 0.21610131859779358,
1730
+ "learning_rate": 9.114666666666668e-06,
1731
+ "loss": 0.0613,
1732
+ "step": 24500
1733
  }
1734
  ],
1735
  "logging_steps": 100,
 
1749
  "attributes": {}
1750
  }
1751
  },
1752
+ "total_flos": 5.967787327488e+16,
1753
  "train_batch_size": 4,
1754
  "trial_name": null,
1755
  "trial_params": null