rakhman-llm commited on
Commit
b3d7d4f
·
verified ·
1 Parent(s): a6c2c79

Training in progress, step 10500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f50dfec7b348b398b16d49e0dd1b5d02944005adb5fa282d41d880b7743a85f
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2eb707efd45838c999c2a5555ec9f4e3948d1f4bebf0270a24ea3404256d4cc
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:25a47f14c217dde4efbeced6d0ecf1b11f4ee8587e56f4b26d3f7cdb603d5217
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3af0e4724c98927d8a051181e73393d7a65a3c727af1c5a91618fb92dd27408
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82f3f920fee151f842583919a34cd3a487b202dcc03fce7b250a1108c0d5893e
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8305955eac22add53d91f06fc3781aadbba74448cb0b21c58a2f9dadbea85a43
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f157fb66e9f6bdea8c4e5f6e07a539c2ce54fcc9de0e2acd2a8bb7cdbb367f5
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e3714e917b57b234e811d02cff6475eee9087a5c2b589dd59d17c8602540672
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.6666666666666666,
5
  "eval_steps": 500,
6
- "global_step": 10000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -707,6 +707,41 @@
707
  "learning_rate": 1.555688888888889e-05,
708
  "loss": 0.0694,
709
  "step": 10000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
710
  }
711
  ],
712
  "logging_steps": 100,
@@ -726,7 +761,7 @@
726
  "attributes": {}
727
  }
728
  },
729
- "total_flos": 2.43583156224e+16,
730
  "train_batch_size": 4,
731
  "trial_name": null,
732
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.7,
5
  "eval_steps": 500,
6
+ "global_step": 10500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
707
  "learning_rate": 1.555688888888889e-05,
708
  "loss": 0.0694,
709
  "step": 10000
710
+ },
711
+ {
712
+ "epoch": 0.6733333333333333,
713
+ "grad_norm": 0.16695012152194977,
714
+ "learning_rate": 1.5512444444444446e-05,
715
+ "loss": 0.0722,
716
+ "step": 10100
717
+ },
718
+ {
719
+ "epoch": 0.68,
720
+ "grad_norm": 0.07470796257257462,
721
+ "learning_rate": 1.5468e-05,
722
+ "loss": 0.0699,
723
+ "step": 10200
724
+ },
725
+ {
726
+ "epoch": 0.6866666666666666,
727
+ "grad_norm": 0.12706786394119263,
728
+ "learning_rate": 1.5423555555555557e-05,
729
+ "loss": 0.068,
730
+ "step": 10300
731
+ },
732
+ {
733
+ "epoch": 0.6933333333333334,
734
+ "grad_norm": 0.14934158325195312,
735
+ "learning_rate": 1.5379111111111112e-05,
736
+ "loss": 0.0643,
737
+ "step": 10400
738
+ },
739
+ {
740
+ "epoch": 0.7,
741
+ "grad_norm": 0.34316006302833557,
742
+ "learning_rate": 1.5334666666666667e-05,
743
+ "loss": 0.0683,
744
+ "step": 10500
745
  }
746
  ],
747
  "logging_steps": 100,
 
761
  "attributes": {}
762
  }
763
  },
764
+ "total_flos": 2.557623140352e+16,
765
  "train_batch_size": 4,
766
  "trial_name": null,
767
  "trial_params": null