rakhman-llm commited on
Commit
b6e54a3
·
verified ·
1 Parent(s): 963ae3c

Training in progress, step 13000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:35af312258ec2ecc580cd7ccdb601acf0843f55378f8ced88aa3b6791062f0ec
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:890ad2873a2bbd9defa42e9be53c396630c1fa479c804ecbcda9c1bef5c9a5b7
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc0410b7f00647d202243d6032612ef86b9b1ad740cc13589fc9a20d6719723c
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7df779224de4aac7dbc2c481b18f226ef75df55b33434fd2a24aaa00199ae182
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f42d5f8190e167df10f51b86e202ce65c6f8b2e34f31e34e57377eb8f3a2ccc
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:392a7eb84bc6c0f11073a38c2f0980cc2fcea2ab711f0ee9d31a3b2c47437b70
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72aa0a4db9a1d4fcb9226563e6d35e410e13273eff2f98d7fe41149f20f9d2f4
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8f863bfc80aea1aeb546487ead9ef6092bc37fe14387401d70bc8b324d152f2
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.8333333333333334,
5
  "eval_steps": 500,
6
- "global_step": 12500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -882,6 +882,41 @@
882
  "learning_rate": 1.4445777777777779e-05,
883
  "loss": 0.0666,
884
  "step": 12500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
885
  }
886
  ],
887
  "logging_steps": 100,
@@ -901,7 +936,7 @@
901
  "attributes": {}
902
  }
903
  },
904
- "total_flos": 3.0447894528e+16,
905
  "train_batch_size": 4,
906
  "trial_name": null,
907
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.8666666666666667,
5
  "eval_steps": 500,
6
+ "global_step": 13000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
882
  "learning_rate": 1.4445777777777779e-05,
883
  "loss": 0.0666,
884
  "step": 12500
885
+ },
886
+ {
887
+ "epoch": 0.84,
888
+ "grad_norm": 0.12742094695568085,
889
+ "learning_rate": 1.4401333333333334e-05,
890
+ "loss": 0.0641,
891
+ "step": 12600
892
+ },
893
+ {
894
+ "epoch": 0.8466666666666667,
895
+ "grad_norm": 0.12476657330989838,
896
+ "learning_rate": 1.4356888888888889e-05,
897
+ "loss": 0.0708,
898
+ "step": 12700
899
+ },
900
+ {
901
+ "epoch": 0.8533333333333334,
902
+ "grad_norm": 0.18345853686332703,
903
+ "learning_rate": 1.4312444444444446e-05,
904
+ "loss": 0.0666,
905
+ "step": 12800
906
+ },
907
+ {
908
+ "epoch": 0.86,
909
+ "grad_norm": 0.24870257079601288,
910
+ "learning_rate": 1.4268000000000001e-05,
911
+ "loss": 0.0643,
912
+ "step": 12900
913
+ },
914
+ {
915
+ "epoch": 0.8666666666666667,
916
+ "grad_norm": 0.41013413667678833,
917
+ "learning_rate": 1.4223555555555557e-05,
918
+ "loss": 0.0723,
919
+ "step": 13000
920
  }
921
  ],
922
  "logging_steps": 100,
 
936
  "attributes": {}
937
  }
938
  },
939
+ "total_flos": 3.166581030912e+16,
940
  "train_batch_size": 4,
941
  "trial_name": null,
942
  "trial_params": null