rakhman-llm commited on
Commit
a0834a0
·
verified ·
1 Parent(s): 090c316

Training in progress, step 39500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8cd71b89eb4cf5b54368e95fa13ee6cc3864c3091e759091438dd601d6f1af7
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f89e86203fc6f340f2bb2ff8e994975dfeb321d8847629162cb09b057f7a04af
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc0dd7a1f2c9ad84ebf73220aa56cda18faf115b1e79f8dbee1b26db6cddaa7d
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb05543b93a1625973f09dd8a4f47b7ff974fc94669448d693bf400cf4b1820e
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:823ea6f73dcec531968a187071c6ecff33e9bb6c6d030f15a43dbc4f25e7c39e
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86347a47866fbcda684d924341c7041bee88821d888a3574e3156decc1c5d213
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90e54113eca63b2259f253dad8dc263308f86084513f881434f4da417da62d75
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9ca915c8cfa9f185cabf22307b900bc37acbb465a96ff7179034cb96d5b0895
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.6,
5
  "eval_steps": 500,
6
- "global_step": 39000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2753,6 +2753,41 @@
2753
  "learning_rate": 2.6728888888888893e-06,
2754
  "loss": 0.0516,
2755
  "step": 39000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2756
  }
2757
  ],
2758
  "logging_steps": 100,
@@ -2772,7 +2807,7 @@
2772
  "attributes": {}
2773
  }
2774
  },
2775
- "total_flos": 9.499743092736e+16,
2776
  "train_batch_size": 4,
2777
  "trial_name": null,
2778
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.6333333333333333,
5
  "eval_steps": 500,
6
+ "global_step": 39500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2753
  "learning_rate": 2.6728888888888893e-06,
2754
  "loss": 0.0516,
2755
  "step": 39000
2756
+ },
2757
+ {
2758
+ "epoch": 2.6066666666666665,
2759
+ "grad_norm": 0.11047784239053726,
2760
+ "learning_rate": 2.6284444444444445e-06,
2761
+ "loss": 0.0494,
2762
+ "step": 39100
2763
+ },
2764
+ {
2765
+ "epoch": 2.6133333333333333,
2766
+ "grad_norm": 0.27986884117126465,
2767
+ "learning_rate": 2.5840000000000006e-06,
2768
+ "loss": 0.0523,
2769
+ "step": 39200
2770
+ },
2771
+ {
2772
+ "epoch": 2.62,
2773
+ "grad_norm": 0.17573469877243042,
2774
+ "learning_rate": 2.539555555555556e-06,
2775
+ "loss": 0.0549,
2776
+ "step": 39300
2777
+ },
2778
+ {
2779
+ "epoch": 2.626666666666667,
2780
+ "grad_norm": 0.26180362701416016,
2781
+ "learning_rate": 2.495111111111111e-06,
2782
+ "loss": 0.0543,
2783
+ "step": 39400
2784
+ },
2785
+ {
2786
+ "epoch": 2.6333333333333333,
2787
+ "grad_norm": 0.22933009266853333,
2788
+ "learning_rate": 2.4506666666666667e-06,
2789
+ "loss": 0.053,
2790
+ "step": 39500
2791
  }
2792
  ],
2793
  "logging_steps": 100,
 
2807
  "attributes": {}
2808
  }
2809
  },
2810
+ "total_flos": 9.621534670848e+16,
2811
  "train_batch_size": 4,
2812
  "trial_name": null,
2813
  "trial_params": null