rakhman-llm commited on
Commit
f149cba
·
verified ·
1 Parent(s): bb36aa3

Training in progress, step 41000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9879d5d88e7644494f4c879b1c5e524603249d32b69969a0534f73a0d99cdd8
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a43c2a0afd3692b4457c78e060bd95a6ca8422aa3f2bfb48776d966c67148820
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b43d2d95bf538247d49799d209038264216cbb911042f24a24e56175bd002ec5
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb247ef281f39dbf138924be64b77fed104e544ac60daad36b17f01b15edabf4
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:78de93de67391223b4c448919361f3b2f57691bd51a633014c5f582b70c0d977
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83988fc1574bdb56177858ad2504e5d5e90771242080ec94288ec2934ea05174
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6aeb9cc77150e881276f00e59441fb5620f10d8b951379e0f05e5c2a60199333
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a519ef81b846f8a03a3bd8bed22cf3df79611baad6f40e402df6bfc345f1aa61
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.7,
5
  "eval_steps": 500,
6
- "global_step": 40500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2858,6 +2858,41 @@
2858
  "learning_rate": 2.006666666666667e-06,
2859
  "loss": 0.0529,
2860
  "step": 40500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2861
  }
2862
  ],
2863
  "logging_steps": 100,
@@ -2877,7 +2912,7 @@
2877
  "attributes": {}
2878
  }
2879
  },
2880
- "total_flos": 9.865117827072e+16,
2881
  "train_batch_size": 4,
2882
  "trial_name": null,
2883
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.7333333333333334,
5
  "eval_steps": 500,
6
+ "global_step": 41000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2858
  "learning_rate": 2.006666666666667e-06,
2859
  "loss": 0.0529,
2860
  "step": 40500
2861
+ },
2862
+ {
2863
+ "epoch": 2.7066666666666666,
2864
+ "grad_norm": 0.24057607352733612,
2865
+ "learning_rate": 1.9622222222222224e-06,
2866
+ "loss": 0.0537,
2867
+ "step": 40600
2868
+ },
2869
+ {
2870
+ "epoch": 2.7133333333333334,
2871
+ "grad_norm": 0.22103776037693024,
2872
+ "learning_rate": 1.9182222222222225e-06,
2873
+ "loss": 0.0499,
2874
+ "step": 40700
2875
+ },
2876
+ {
2877
+ "epoch": 2.7199999999999998,
2878
+ "grad_norm": 0.23199285566806793,
2879
+ "learning_rate": 1.873777777777778e-06,
2880
+ "loss": 0.0574,
2881
+ "step": 40800
2882
+ },
2883
+ {
2884
+ "epoch": 2.7266666666666666,
2885
+ "grad_norm": 0.3680346608161926,
2886
+ "learning_rate": 1.8293333333333336e-06,
2887
+ "loss": 0.0455,
2888
+ "step": 40900
2889
+ },
2890
+ {
2891
+ "epoch": 2.7333333333333334,
2892
+ "grad_norm": 0.11471223831176758,
2893
+ "learning_rate": 1.784888888888889e-06,
2894
+ "loss": 0.046,
2895
+ "step": 41000
2896
  }
2897
  ],
2898
  "logging_steps": 100,
 
2912
  "attributes": {}
2913
  }
2914
  },
2915
+ "total_flos": 9.986909405184e+16,
2916
  "train_batch_size": 4,
2917
  "trial_name": null,
2918
  "trial_params": null