rakhman-llm commited on
Commit
dd03e64
·
verified ·
1 Parent(s): e8c7392

Training in progress, step 15000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44b5263c1decc645c9e701779bd3d240d1bd128257b6edb8ead3471aa6e882ff
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7454f4d66c270e44df1eacbd6185e4004d87782931028de1f535c0307f116fd
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4662d2a8a819e0d85a2d8fe1130697f20508f25fc829adbfc7041f6ee0ff9440
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ad72731605d1660caf040ca225055f57942f6652ac3f8e3f4a48d6e14eb50fd
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bcf69dbca5717fd7e2d59b8c55b37845db01c5e5f067d9b29c7651b0c07744c5
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94a590caf5b0791267c7c662cc1f8162ae428b45baf48632e99e270be42d5011
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ac2ccaa6ddf120173c07745369f72632efad440b0563e86a0141350749aa0ae
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d713a762ef6487a34237b674f0e37296a124b258ea254d4be9d4a61b4da657a1
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.9666666666666667,
5
  "eval_steps": 500,
6
- "global_step": 14500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1022,6 +1022,41 @@
1022
  "learning_rate": 1.3557333333333333e-05,
1023
  "loss": 0.0643,
1024
  "step": 14500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1025
  }
1026
  ],
1027
  "logging_steps": 100,
@@ -1041,7 +1076,7 @@
1041
  "attributes": {}
1042
  }
1043
  },
1044
- "total_flos": 3.531955765248e+16,
1045
  "train_batch_size": 4,
1046
  "trial_name": null,
1047
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
  "eval_steps": 500,
6
+ "global_step": 15000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1022
  "learning_rate": 1.3557333333333333e-05,
1023
  "loss": 0.0643,
1024
  "step": 14500
1025
+ },
1026
+ {
1027
+ "epoch": 0.9733333333333334,
1028
+ "grad_norm": 0.18082498013973236,
1029
+ "learning_rate": 1.351288888888889e-05,
1030
+ "loss": 0.0646,
1031
+ "step": 14600
1032
+ },
1033
+ {
1034
+ "epoch": 0.98,
1035
+ "grad_norm": 0.14383332431316376,
1036
+ "learning_rate": 1.3468444444444447e-05,
1037
+ "loss": 0.066,
1038
+ "step": 14700
1039
+ },
1040
+ {
1041
+ "epoch": 0.9866666666666667,
1042
+ "grad_norm": 0.20313547551631927,
1043
+ "learning_rate": 1.3424000000000002e-05,
1044
+ "loss": 0.0631,
1045
+ "step": 14800
1046
+ },
1047
+ {
1048
+ "epoch": 0.9933333333333333,
1049
+ "grad_norm": 0.14930222928524017,
1050
+ "learning_rate": 1.3379555555555557e-05,
1051
+ "loss": 0.0593,
1052
+ "step": 14900
1053
+ },
1054
+ {
1055
+ "epoch": 1.0,
1056
+ "grad_norm": 0.30177292227745056,
1057
+ "learning_rate": 1.3335111111111113e-05,
1058
+ "loss": 0.0651,
1059
+ "step": 15000
1060
  }
1061
  ],
1062
  "logging_steps": 100,
 
1076
  "attributes": {}
1077
  }
1078
  },
1079
+ "total_flos": 3.65374734336e+16,
1080
  "train_batch_size": 4,
1081
  "trial_name": null,
1082
  "trial_params": null