rakhman-llm commited on
Commit
33fad90
·
verified ·
1 Parent(s): ed49df7

Training in progress, step 12000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8dfb1303c43437d0a5d16c718f1a0be0355a6bc6198d78fcee515c503884a9a2
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1d368685b466c586ae7754c9d6601a94b88a29fa10089466eaf7691d618758c
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9169b770e0ac0200417a6d68811804c1fb73c586cfc8a20550b0caa6beef0a3
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be528ec45cd20464a9f579ede6fc32d36b21c545d4649748da7c0e442878da12
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68393a5eefa6936320457e661ed8769751e258badc1e65d137b49f873ca59e29
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca90357941cc8ba75656fcb364086055fdccb9002e991bd59479c3c303e64d7f
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8a1695a275cff99a2c62fc4abe2a84b274a1c4f4a080a28dd6e841bca4417f1
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c83b3bacd9fafe4707bb796f339853939afe8625c6ed3b03b48e890e264959d8
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.7666666666666667,
5
  "eval_steps": 500,
6
- "global_step": 11500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -812,6 +812,41 @@
812
  "learning_rate": 1.4890222222222223e-05,
813
  "loss": 0.0624,
814
  "step": 11500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
815
  }
816
  ],
817
  "logging_steps": 100,
@@ -831,7 +866,7 @@
831
  "attributes": {}
832
  }
833
  },
834
- "total_flos": 2.801206296576e+16,
835
  "train_batch_size": 4,
836
  "trial_name": null,
837
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.8,
5
  "eval_steps": 500,
6
+ "global_step": 12000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
812
  "learning_rate": 1.4890222222222223e-05,
813
  "loss": 0.0624,
814
  "step": 11500
815
+ },
816
+ {
817
+ "epoch": 0.7733333333333333,
818
+ "grad_norm": 0.2603534758090973,
819
+ "learning_rate": 1.4845777777777778e-05,
820
+ "loss": 0.0637,
821
+ "step": 11600
822
+ },
823
+ {
824
+ "epoch": 0.78,
825
+ "grad_norm": 0.2619217038154602,
826
+ "learning_rate": 1.4801333333333333e-05,
827
+ "loss": 0.0593,
828
+ "step": 11700
829
+ },
830
+ {
831
+ "epoch": 0.7866666666666666,
832
+ "grad_norm": 0.06773953139781952,
833
+ "learning_rate": 1.4756888888888889e-05,
834
+ "loss": 0.0689,
835
+ "step": 11800
836
+ },
837
+ {
838
+ "epoch": 0.7933333333333333,
839
+ "grad_norm": 0.09101361781358719,
840
+ "learning_rate": 1.4712444444444444e-05,
841
+ "loss": 0.0595,
842
+ "step": 11900
843
+ },
844
+ {
845
+ "epoch": 0.8,
846
+ "grad_norm": 0.30984750390052795,
847
+ "learning_rate": 1.4668000000000003e-05,
848
+ "loss": 0.0601,
849
+ "step": 12000
850
  }
851
  ],
852
  "logging_steps": 100,
 
866
  "attributes": {}
867
  }
868
  },
869
+ "total_flos": 2.922997874688e+16,
870
  "train_batch_size": 4,
871
  "trial_name": null,
872
  "trial_params": null