rakhman-llm commited on
Commit
dd5c577
·
verified ·
1 Parent(s): 734d263

Training in progress, step 25500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8cc3b5dc11d4b2e43886ec06772cc1df8fb96be6f1f9a0270f9f6e971f6d0e2
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77f8bbdcef942ca91b5ec7bbbd735231f5e4d26584c5cdee0e481501fe23e48b
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3972de8e85ec92a8ee5bab9415d6a3d3f55feb08f92a26b7359e3c1d72344bc5
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f30063e10b99a88358cb2b15c6445572352ad0fdb9ca22aac6f16a615cc3216
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6fbd5b30263f6cc98f2bb34e98264f7dd554a2e53f93ad44d0666e3bd6bfb80c
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbcd414e2ebe010dcb7c52553aab3fd4fa6d365b6f63593edeab1858ed2ed198
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c7471c9521805c24a8ce63003bd5efeaf8bf27814fa393d151be750f3265d98e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:797fe0fbbc0deae3aec7fbd325c4b3b26250213d429ef253b6bc3ac068bea992
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.6666666666666665,
5
  "eval_steps": 500,
6
- "global_step": 25000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1765,6 +1765,41 @@
1765
  "learning_rate": 8.892444444444445e-06,
1766
  "loss": 0.0566,
1767
  "step": 25000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1768
  }
1769
  ],
1770
  "logging_steps": 100,
@@ -1784,7 +1819,7 @@
1784
  "attributes": {}
1785
  }
1786
  },
1787
- "total_flos": 6.0895789056e+16,
1788
  "train_batch_size": 4,
1789
  "trial_name": null,
1790
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.7,
5
  "eval_steps": 500,
6
+ "global_step": 25500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1765
  "learning_rate": 8.892444444444445e-06,
1766
  "loss": 0.0566,
1767
  "step": 25000
1768
+ },
1769
+ {
1770
+ "epoch": 1.6733333333333333,
1771
+ "grad_norm": 0.14110605418682098,
1772
+ "learning_rate": 8.848e-06,
1773
+ "loss": 0.0554,
1774
+ "step": 25100
1775
+ },
1776
+ {
1777
+ "epoch": 1.6800000000000002,
1778
+ "grad_norm": 0.17642982304096222,
1779
+ "learning_rate": 8.803555555555556e-06,
1780
+ "loss": 0.0581,
1781
+ "step": 25200
1782
+ },
1783
+ {
1784
+ "epoch": 1.6866666666666665,
1785
+ "grad_norm": 0.19941475987434387,
1786
+ "learning_rate": 8.759111111111111e-06,
1787
+ "loss": 0.0598,
1788
+ "step": 25300
1789
+ },
1790
+ {
1791
+ "epoch": 1.6933333333333334,
1792
+ "grad_norm": 0.15960603952407837,
1793
+ "learning_rate": 8.714666666666666e-06,
1794
+ "loss": 0.0531,
1795
+ "step": 25400
1796
+ },
1797
+ {
1798
+ "epoch": 1.7,
1799
+ "grad_norm": 0.17013077437877655,
1800
+ "learning_rate": 8.670222222222223e-06,
1801
+ "loss": 0.0596,
1802
+ "step": 25500
1803
  }
1804
  ],
1805
  "logging_steps": 100,
 
1819
  "attributes": {}
1820
  }
1821
  },
1822
+ "total_flos": 6.211370483712e+16,
1823
  "train_batch_size": 4,
1824
  "trial_name": null,
1825
  "trial_params": null