rakhman-llm commited on
Commit
389d5c0
·
verified ·
1 Parent(s): ebf3b46

Training in progress, step 40000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f89e86203fc6f340f2bb2ff8e994975dfeb321d8847629162cb09b057f7a04af
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b86dc095c2649030c80dbb3b306b6dc458cfc29054732a394c23118a3921570
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb05543b93a1625973f09dd8a4f47b7ff974fc94669448d693bf400cf4b1820e
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbe16dcb2945b2b310dff07ecbc105e5d8932c93cc840533642e145b021fcd46
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:86347a47866fbcda684d924341c7041bee88821d888a3574e3156decc1c5d213
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:996bcc2fc9a86c5f72875fc2c94242013d036dcc7c497de3b8bae67dc7683e4b
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9ca915c8cfa9f185cabf22307b900bc37acbb465a96ff7179034cb96d5b0895
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bc8f3c97f58f2fb9ce92ef34272eb4a5ca6976677613743a4e1abd3e5adada4
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.6333333333333333,
5
  "eval_steps": 500,
6
- "global_step": 39500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2788,6 +2788,41 @@
2788
  "learning_rate": 2.4506666666666667e-06,
2789
  "loss": 0.053,
2790
  "step": 39500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2791
  }
2792
  ],
2793
  "logging_steps": 100,
@@ -2807,7 +2842,7 @@
2807
  "attributes": {}
2808
  }
2809
  },
2810
- "total_flos": 9.621534670848e+16,
2811
  "train_batch_size": 4,
2812
  "trial_name": null,
2813
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.6666666666666665,
5
  "eval_steps": 500,
6
+ "global_step": 40000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2788
  "learning_rate": 2.4506666666666667e-06,
2789
  "loss": 0.053,
2790
  "step": 39500
2791
+ },
2792
+ {
2793
+ "epoch": 2.64,
2794
+ "grad_norm": 0.10127383470535278,
2795
+ "learning_rate": 2.4062222222222223e-06,
2796
+ "loss": 0.0494,
2797
+ "step": 39600
2798
+ },
2799
+ {
2800
+ "epoch": 2.6466666666666665,
2801
+ "grad_norm": 0.1442176252603531,
2802
+ "learning_rate": 2.3622222222222224e-06,
2803
+ "loss": 0.0489,
2804
+ "step": 39700
2805
+ },
2806
+ {
2807
+ "epoch": 2.6533333333333333,
2808
+ "grad_norm": 0.2334638386964798,
2809
+ "learning_rate": 2.317777777777778e-06,
2810
+ "loss": 0.0493,
2811
+ "step": 39800
2812
+ },
2813
+ {
2814
+ "epoch": 2.66,
2815
+ "grad_norm": 0.2831490635871887,
2816
+ "learning_rate": 2.2733333333333333e-06,
2817
+ "loss": 0.0515,
2818
+ "step": 39900
2819
+ },
2820
+ {
2821
+ "epoch": 2.6666666666666665,
2822
+ "grad_norm": 0.35798129439353943,
2823
+ "learning_rate": 2.228888888888889e-06,
2824
+ "loss": 0.0502,
2825
+ "step": 40000
2826
  }
2827
  ],
2828
  "logging_steps": 100,
 
2842
  "attributes": {}
2843
  }
2844
  },
2845
+ "total_flos": 9.74332624896e+16,
2846
  "train_batch_size": 4,
2847
  "trial_name": null,
2848
  "trial_params": null