rakhman-llm commited on
Commit
fc687b1
·
verified ·
1 Parent(s): 835fff0

Training in progress, step 12500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f1d368685b466c586ae7754c9d6601a94b88a29fa10089466eaf7691d618758c
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35af312258ec2ecc580cd7ccdb601acf0843f55378f8ced88aa3b6791062f0ec
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be528ec45cd20464a9f579ede6fc32d36b21c545d4649748da7c0e442878da12
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc0410b7f00647d202243d6032612ef86b9b1ad740cc13589fc9a20d6719723c
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca90357941cc8ba75656fcb364086055fdccb9002e991bd59479c3c303e64d7f
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f42d5f8190e167df10f51b86e202ce65c6f8b2e34f31e34e57377eb8f3a2ccc
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c83b3bacd9fafe4707bb796f339853939afe8625c6ed3b03b48e890e264959d8
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72aa0a4db9a1d4fcb9226563e6d35e410e13273eff2f98d7fe41149f20f9d2f4
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.8,
5
  "eval_steps": 500,
6
- "global_step": 12000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -847,6 +847,41 @@
847
  "learning_rate": 1.4668000000000003e-05,
848
  "loss": 0.0601,
849
  "step": 12000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
850
  }
851
  ],
852
  "logging_steps": 100,
@@ -866,7 +901,7 @@
866
  "attributes": {}
867
  }
868
  },
869
- "total_flos": 2.922997874688e+16,
870
  "train_batch_size": 4,
871
  "trial_name": null,
872
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.8333333333333334,
5
  "eval_steps": 500,
6
+ "global_step": 12500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
847
  "learning_rate": 1.4668000000000003e-05,
848
  "loss": 0.0601,
849
  "step": 12000
850
+ },
851
+ {
852
+ "epoch": 0.8066666666666666,
853
+ "grad_norm": 0.12947213649749756,
854
+ "learning_rate": 1.4623555555555558e-05,
855
+ "loss": 0.07,
856
+ "step": 12100
857
+ },
858
+ {
859
+ "epoch": 0.8133333333333334,
860
+ "grad_norm": 0.12053196877241135,
861
+ "learning_rate": 1.4579111111111113e-05,
862
+ "loss": 0.0675,
863
+ "step": 12200
864
+ },
865
+ {
866
+ "epoch": 0.82,
867
+ "grad_norm": 0.15298992395401,
868
+ "learning_rate": 1.4534666666666668e-05,
869
+ "loss": 0.0663,
870
+ "step": 12300
871
+ },
872
+ {
873
+ "epoch": 0.8266666666666667,
874
+ "grad_norm": 0.11816833913326263,
875
+ "learning_rate": 1.4490222222222223e-05,
876
+ "loss": 0.0681,
877
+ "step": 12400
878
+ },
879
+ {
880
+ "epoch": 0.8333333333333334,
881
+ "grad_norm": 0.22705060243606567,
882
+ "learning_rate": 1.4445777777777779e-05,
883
+ "loss": 0.0666,
884
+ "step": 12500
885
  }
886
  ],
887
  "logging_steps": 100,
 
901
  "attributes": {}
902
  }
903
  },
904
+ "total_flos": 3.0447894528e+16,
905
  "train_batch_size": 4,
906
  "trial_name": null,
907
  "trial_params": null