rakhman-llm commited on
Commit
7110c3b
·
verified ·
1 Parent(s): 0714c03

Training in progress, step 26500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c2eb2e99816ec43430cbc064844c5e9639d5777774741cf5128e00d44cc22f84
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d79b372779eff7f0c6aceb504965064dd1c783ad82104b4d76980d00e49918c
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ffff2b04f3ac17f12a961b31f194814499e7fd990021cd2b459a7cd25f837f54
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c628badd7736af0b2faf987d91b45461eb187ce865defe1587b4f0f12761f4a
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f52f95da495bb27845cab982f10d0992545e1a1f3379ad63ed77c13e8413544
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca75d61d47f3eab98fc406af0fc3456adbe0767939034d58dea95e5d6ff029b5
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34f762ade12141f21b1db11036bc2604582ab540ca56d64f7190f6e92cb5741c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d82350f62694be97e6ec7d5caa84c96c6ffb8c480e6a50519ae958022a2c0417
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.7333333333333334,
5
  "eval_steps": 500,
6
- "global_step": 26000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1835,6 +1835,41 @@
1835
  "learning_rate": 8.448000000000001e-06,
1836
  "loss": 0.0555,
1837
  "step": 26000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1838
  }
1839
  ],
1840
  "logging_steps": 100,
@@ -1854,7 +1889,7 @@
1854
  "attributes": {}
1855
  }
1856
  },
1857
- "total_flos": 6.333162061824e+16,
1858
  "train_batch_size": 4,
1859
  "trial_name": null,
1860
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.7666666666666666,
5
  "eval_steps": 500,
6
+ "global_step": 26500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1835
  "learning_rate": 8.448000000000001e-06,
1836
  "loss": 0.0555,
1837
  "step": 26000
1838
+ },
1839
+ {
1840
+ "epoch": 1.74,
1841
+ "grad_norm": 0.25770798325538635,
1842
+ "learning_rate": 8.403555555555556e-06,
1843
+ "loss": 0.0513,
1844
+ "step": 26100
1845
+ },
1846
+ {
1847
+ "epoch": 1.7466666666666666,
1848
+ "grad_norm": 0.2838607430458069,
1849
+ "learning_rate": 8.359111111111112e-06,
1850
+ "loss": 0.0605,
1851
+ "step": 26200
1852
+ },
1853
+ {
1854
+ "epoch": 1.7533333333333334,
1855
+ "grad_norm": 0.12263695150613785,
1856
+ "learning_rate": 8.314666666666667e-06,
1857
+ "loss": 0.0574,
1858
+ "step": 26300
1859
+ },
1860
+ {
1861
+ "epoch": 1.76,
1862
+ "grad_norm": 0.17367112636566162,
1863
+ "learning_rate": 8.270666666666667e-06,
1864
+ "loss": 0.0581,
1865
+ "step": 26400
1866
+ },
1867
+ {
1868
+ "epoch": 1.7666666666666666,
1869
+ "grad_norm": 0.26262718439102173,
1870
+ "learning_rate": 8.226222222222224e-06,
1871
+ "loss": 0.0541,
1872
+ "step": 26500
1873
  }
1874
  ],
1875
  "logging_steps": 100,
 
1889
  "attributes": {}
1890
  }
1891
  },
1892
+ "total_flos": 6.454953639936e+16,
1893
  "train_batch_size": 4,
1894
  "trial_name": null,
1895
  "trial_params": null