rakhman-llm commited on
Commit
b35b62e
·
verified ·
1 Parent(s): 4157a91

Training in progress, step 26000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77f8bbdcef942ca91b5ec7bbbd735231f5e4d26584c5cdee0e481501fe23e48b
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2eb2e99816ec43430cbc064844c5e9639d5777774741cf5128e00d44cc22f84
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f30063e10b99a88358cb2b15c6445572352ad0fdb9ca22aac6f16a615cc3216
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffff2b04f3ac17f12a961b31f194814499e7fd990021cd2b459a7cd25f837f54
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fbcd414e2ebe010dcb7c52553aab3fd4fa6d365b6f63593edeab1858ed2ed198
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f52f95da495bb27845cab982f10d0992545e1a1f3379ad63ed77c13e8413544
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:797fe0fbbc0deae3aec7fbd325c4b3b26250213d429ef253b6bc3ac068bea992
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34f762ade12141f21b1db11036bc2604582ab540ca56d64f7190f6e92cb5741c
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.7,
5
  "eval_steps": 500,
6
- "global_step": 25500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1800,6 +1800,41 @@
1800
  "learning_rate": 8.670222222222223e-06,
1801
  "loss": 0.0596,
1802
  "step": 25500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1803
  }
1804
  ],
1805
  "logging_steps": 100,
@@ -1819,7 +1854,7 @@
1819
  "attributes": {}
1820
  }
1821
  },
1822
- "total_flos": 6.211370483712e+16,
1823
  "train_batch_size": 4,
1824
  "trial_name": null,
1825
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.7333333333333334,
5
  "eval_steps": 500,
6
+ "global_step": 26000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1800
  "learning_rate": 8.670222222222223e-06,
1801
  "loss": 0.0596,
1802
  "step": 25500
1803
+ },
1804
+ {
1805
+ "epoch": 1.7066666666666666,
1806
+ "grad_norm": 0.20626091957092285,
1807
+ "learning_rate": 8.625777777777779e-06,
1808
+ "loss": 0.0523,
1809
+ "step": 25600
1810
+ },
1811
+ {
1812
+ "epoch": 1.7133333333333334,
1813
+ "grad_norm": 0.1952398121356964,
1814
+ "learning_rate": 8.581333333333334e-06,
1815
+ "loss": 0.0508,
1816
+ "step": 25700
1817
+ },
1818
+ {
1819
+ "epoch": 1.72,
1820
+ "grad_norm": 0.15694676339626312,
1821
+ "learning_rate": 8.53688888888889e-06,
1822
+ "loss": 0.052,
1823
+ "step": 25800
1824
+ },
1825
+ {
1826
+ "epoch": 1.7266666666666666,
1827
+ "grad_norm": 0.07521438598632812,
1828
+ "learning_rate": 8.492444444444446e-06,
1829
+ "loss": 0.0597,
1830
+ "step": 25900
1831
+ },
1832
+ {
1833
+ "epoch": 1.7333333333333334,
1834
+ "grad_norm": 0.07147266715765,
1835
+ "learning_rate": 8.448000000000001e-06,
1836
+ "loss": 0.0555,
1837
+ "step": 26000
1838
  }
1839
  ],
1840
  "logging_steps": 100,
 
1854
  "attributes": {}
1855
  }
1856
  },
1857
+ "total_flos": 6.333162061824e+16,
1858
  "train_batch_size": 4,
1859
  "trial_name": null,
1860
  "trial_params": null