rakhman-llm commited on
Commit
efc5c76
·
verified ·
1 Parent(s): a28855e

Training in progress, step 10000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e739525a605197613b8e712741d416769ed5f57619fff2ba4f8511681ce890ed
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f50dfec7b348b398b16d49e0dd1b5d02944005adb5fa282d41d880b7743a85f
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d4c2b2a5410ff424a41e1a6cde39521ad45e7ac7555648b29eaff3b782134b0c
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25a47f14c217dde4efbeced6d0ecf1b11f4ee8587e56f4b26d3f7cdb603d5217
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ebf0d13686eae3e0c0f52410d677d657c1203d01171ca9804ddd94825765dcc7
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82f3f920fee151f842583919a34cd3a487b202dcc03fce7b250a1108c0d5893e
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4e47b661f0982d98bcfee7c8a9db89091e3f5b040309ef2739e290fa07adec1
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f157fb66e9f6bdea8c4e5f6e07a539c2ce54fcc9de0e2acd2a8bb7cdbb367f5
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.6333333333333333,
5
  "eval_steps": 500,
6
- "global_step": 9500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -672,6 +672,41 @@
672
  "learning_rate": 1.577911111111111e-05,
673
  "loss": 0.0635,
674
  "step": 9500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
675
  }
676
  ],
677
  "logging_steps": 100,
@@ -691,7 +726,7 @@
691
  "attributes": {}
692
  }
693
  },
694
- "total_flos": 2.314039984128e+16,
695
  "train_batch_size": 4,
696
  "trial_name": null,
697
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.6666666666666666,
5
  "eval_steps": 500,
6
+ "global_step": 10000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
672
  "learning_rate": 1.577911111111111e-05,
673
  "loss": 0.0635,
674
  "step": 9500
675
+ },
676
+ {
677
+ "epoch": 0.64,
678
+ "grad_norm": 0.1181679293513298,
679
+ "learning_rate": 1.5734666666666667e-05,
680
+ "loss": 0.0656,
681
+ "step": 9600
682
+ },
683
+ {
684
+ "epoch": 0.6466666666666666,
685
+ "grad_norm": 0.16007280349731445,
686
+ "learning_rate": 1.5690222222222226e-05,
687
+ "loss": 0.0678,
688
+ "step": 9700
689
+ },
690
+ {
691
+ "epoch": 0.6533333333333333,
692
+ "grad_norm": 0.34087005257606506,
693
+ "learning_rate": 1.564577777777778e-05,
694
+ "loss": 0.0637,
695
+ "step": 9800
696
+ },
697
+ {
698
+ "epoch": 0.66,
699
+ "grad_norm": 0.27932071685791016,
700
+ "learning_rate": 1.5601333333333336e-05,
701
+ "loss": 0.0715,
702
+ "step": 9900
703
+ },
704
+ {
705
+ "epoch": 0.6666666666666666,
706
+ "grad_norm": 0.21511399745941162,
707
+ "learning_rate": 1.555688888888889e-05,
708
+ "loss": 0.0694,
709
+ "step": 10000
710
  }
711
  ],
712
  "logging_steps": 100,
 
726
  "attributes": {}
727
  }
728
  },
729
+ "total_flos": 2.43583156224e+16,
730
  "train_batch_size": 4,
731
  "trial_name": null,
732
  "trial_params": null