rakhman-llm commited on
Commit
db35be4
·
verified ·
1 Parent(s): 66111fe

Training in progress, step 27000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1d79b372779eff7f0c6aceb504965064dd1c783ad82104b4d76980d00e49918c
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5aee5c58e4fc85927dc4013d0fe750e56e604920994d516f324190a1a8d7a05f
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c628badd7736af0b2faf987d91b45461eb187ce865defe1587b4f0f12761f4a
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d58926cc53cc96e7baec8d580220f0a290e47d6bc25d45f7fd7e61bf35d05d9a
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca75d61d47f3eab98fc406af0fc3456adbe0767939034d58dea95e5d6ff029b5
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8469fa4abc9740f9af7a5dacb10ea8c5107fb6c8cbc91f6fcad8c8b05132d7d0
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d82350f62694be97e6ec7d5caa84c96c6ffb8c480e6a50519ae958022a2c0417
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8dc2ab4847205143252d3ebb3c6929a744853487e6aadfb5f1c78d5dcdea141c
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.7666666666666666,
5
  "eval_steps": 500,
6
- "global_step": 26500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1870,6 +1870,41 @@
1870
  "learning_rate": 8.226222222222224e-06,
1871
  "loss": 0.0541,
1872
  "step": 26500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1873
  }
1874
  ],
1875
  "logging_steps": 100,
@@ -1889,7 +1924,7 @@
1889
  "attributes": {}
1890
  }
1891
  },
1892
- "total_flos": 6.454953639936e+16,
1893
  "train_batch_size": 4,
1894
  "trial_name": null,
1895
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.8,
5
  "eval_steps": 500,
6
+ "global_step": 27000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1870
  "learning_rate": 8.226222222222224e-06,
1871
  "loss": 0.0541,
1872
  "step": 26500
1873
+ },
1874
+ {
1875
+ "epoch": 1.7733333333333334,
1876
+ "grad_norm": 0.3220883309841156,
1877
+ "learning_rate": 8.181777777777779e-06,
1878
+ "loss": 0.0543,
1879
+ "step": 26600
1880
+ },
1881
+ {
1882
+ "epoch": 1.78,
1883
+ "grad_norm": 0.18442386388778687,
1884
+ "learning_rate": 8.137333333333334e-06,
1885
+ "loss": 0.0633,
1886
+ "step": 26700
1887
+ },
1888
+ {
1889
+ "epoch": 1.7866666666666666,
1890
+ "grad_norm": 0.25864550471305847,
1891
+ "learning_rate": 8.09288888888889e-06,
1892
+ "loss": 0.0572,
1893
+ "step": 26800
1894
+ },
1895
+ {
1896
+ "epoch": 1.7933333333333334,
1897
+ "grad_norm": 0.23834674060344696,
1898
+ "learning_rate": 8.048444444444445e-06,
1899
+ "loss": 0.0537,
1900
+ "step": 26900
1901
+ },
1902
+ {
1903
+ "epoch": 1.8,
1904
+ "grad_norm": 0.1399671733379364,
1905
+ "learning_rate": 8.004e-06,
1906
+ "loss": 0.0572,
1907
+ "step": 27000
1908
  }
1909
  ],
1910
  "logging_steps": 100,
 
1924
  "attributes": {}
1925
  }
1926
  },
1927
+ "total_flos": 6.576745218048e+16,
1928
  "train_batch_size": 4,
1929
  "trial_name": null,
1930
  "trial_params": null