rakhman-llm commited on
Commit
71645f7
·
verified ·
1 Parent(s): 2e23ff4

Training in progress, step 14500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e948c8d9a7ad56ccbd6700926716442cecc28d16ec12db2c3ed5e7e3f659db9b
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44b5263c1decc645c9e701779bd3d240d1bd128257b6edb8ead3471aa6e882ff
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af58a191c850477c58ee5c9fd6d80b4f79d202f3c260b244c0bb72f25e9c46bf
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4662d2a8a819e0d85a2d8fe1130697f20508f25fc829adbfc7041f6ee0ff9440
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aba8071d282c3cb60c40a4d3269fc6a11f1d4d5aa00c59a4ed8e5ce671a0abf0
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcf69dbca5717fd7e2d59b8c55b37845db01c5e5f067d9b29c7651b0c07744c5
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aecb0190c300aa10c453b1ef86e11d1993afb13996c225cb9c0fcb417f89ff36
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ac2ccaa6ddf120173c07745369f72632efad440b0563e86a0141350749aa0ae
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.9333333333333333,
5
  "eval_steps": 500,
6
- "global_step": 14000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -987,6 +987,41 @@
987
  "learning_rate": 1.3779555555555557e-05,
988
  "loss": 0.0666,
989
  "step": 14000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
990
  }
991
  ],
992
  "logging_steps": 100,
@@ -1006,7 +1041,7 @@
1006
  "attributes": {}
1007
  }
1008
  },
1009
- "total_flos": 3.410164187136e+16,
1010
  "train_batch_size": 4,
1011
  "trial_name": null,
1012
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.9666666666666667,
5
  "eval_steps": 500,
6
+ "global_step": 14500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
987
  "learning_rate": 1.3779555555555557e-05,
988
  "loss": 0.0666,
989
  "step": 14000
990
+ },
991
+ {
992
+ "epoch": 0.94,
993
+ "grad_norm": 0.12045079469680786,
994
+ "learning_rate": 1.3735111111111112e-05,
995
+ "loss": 0.0712,
996
+ "step": 14100
997
+ },
998
+ {
999
+ "epoch": 0.9466666666666667,
1000
+ "grad_norm": 0.1625666618347168,
1001
+ "learning_rate": 1.3690666666666667e-05,
1002
+ "loss": 0.0633,
1003
+ "step": 14200
1004
+ },
1005
+ {
1006
+ "epoch": 0.9533333333333334,
1007
+ "grad_norm": 0.17143802344799042,
1008
+ "learning_rate": 1.3646222222222223e-05,
1009
+ "loss": 0.0627,
1010
+ "step": 14300
1011
+ },
1012
+ {
1013
+ "epoch": 0.96,
1014
+ "grad_norm": 0.19708478450775146,
1015
+ "learning_rate": 1.3601777777777778e-05,
1016
+ "loss": 0.067,
1017
+ "step": 14400
1018
+ },
1019
+ {
1020
+ "epoch": 0.9666666666666667,
1021
+ "grad_norm": 0.1801692545413971,
1022
+ "learning_rate": 1.3557333333333333e-05,
1023
+ "loss": 0.0643,
1024
+ "step": 14500
1025
  }
1026
  ],
1027
  "logging_steps": 100,
 
1041
  "attributes": {}
1042
  }
1043
  },
1044
+ "total_flos": 3.531955765248e+16,
1045
  "train_batch_size": 4,
1046
  "trial_name": null,
1047
  "trial_params": null