rakhman-llm commited on
Commit
7bfd3d8
·
verified ·
1 Parent(s): 60c6390

Training in progress, step 25000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3030d4ce44a28d2687854bece3d36dd11a2483143dc2362c64bdaf84a148a35c
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8cc3b5dc11d4b2e43886ec06772cc1df8fb96be6f1f9a0270f9f6e971f6d0e2
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8336f578c2856a2611344b57de4fde614962159ed2c75eff7de5a881717f1aa8
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3972de8e85ec92a8ee5bab9415d6a3d3f55feb08f92a26b7359e3c1d72344bc5
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7a44025f64c6f7ad3ba883de4c1f4995d535cdd50931b810e0a7d00e22e5cdc
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fbd5b30263f6cc98f2bb34e98264f7dd554a2e53f93ad44d0666e3bd6bfb80c
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6d9ad9a351ae61b83248fa17cb95ac2021f5fe42e6699ef47e0f8bde893c34bb
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7471c9521805c24a8ce63003bd5efeaf8bf27814fa393d151be750f3265d98e
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.6333333333333333,
5
  "eval_steps": 500,
6
- "global_step": 24500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1730,6 +1730,41 @@
1730
  "learning_rate": 9.114666666666668e-06,
1731
  "loss": 0.0613,
1732
  "step": 24500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1733
  }
1734
  ],
1735
  "logging_steps": 100,
@@ -1749,7 +1784,7 @@
1749
  "attributes": {}
1750
  }
1751
  },
1752
- "total_flos": 5.967787327488e+16,
1753
  "train_batch_size": 4,
1754
  "trial_name": null,
1755
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.6666666666666665,
5
  "eval_steps": 500,
6
+ "global_step": 25000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1730
  "learning_rate": 9.114666666666668e-06,
1731
  "loss": 0.0613,
1732
  "step": 24500
1733
+ },
1734
+ {
1735
+ "epoch": 1.6400000000000001,
1736
+ "grad_norm": 0.25385352969169617,
1737
+ "learning_rate": 9.070222222222223e-06,
1738
+ "loss": 0.0599,
1739
+ "step": 24600
1740
+ },
1741
+ {
1742
+ "epoch": 1.6466666666666665,
1743
+ "grad_norm": 0.1330012083053589,
1744
+ "learning_rate": 9.025777777777778e-06,
1745
+ "loss": 0.0601,
1746
+ "step": 24700
1747
+ },
1748
+ {
1749
+ "epoch": 1.6533333333333333,
1750
+ "grad_norm": 0.40342622995376587,
1751
+ "learning_rate": 8.981333333333333e-06,
1752
+ "loss": 0.0599,
1753
+ "step": 24800
1754
+ },
1755
+ {
1756
+ "epoch": 1.6600000000000001,
1757
+ "grad_norm": 0.21778804063796997,
1758
+ "learning_rate": 8.93688888888889e-06,
1759
+ "loss": 0.0537,
1760
+ "step": 24900
1761
+ },
1762
+ {
1763
+ "epoch": 1.6666666666666665,
1764
+ "grad_norm": 0.17448577284812927,
1765
+ "learning_rate": 8.892444444444445e-06,
1766
+ "loss": 0.0566,
1767
+ "step": 25000
1768
  }
1769
  ],
1770
  "logging_steps": 100,
 
1784
  "attributes": {}
1785
  }
1786
  },
1787
+ "total_flos": 6.0895789056e+16,
1788
  "train_batch_size": 4,
1789
  "trial_name": null,
1790
  "trial_params": null