rakhman-llm commited on
Commit
dd3483d
·
verified ·
1 Parent(s): 2aef9fb

Training in progress, step 31500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:35881b4a958bf056508d114860b14084addd8682c5b35498eb4cbf8f6a42ffcc
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b61d5a5a0f5b7ace55f8bc3cfcc9327599d34056cf4715c1510e9c33291ed41
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a03972f1e2c45bae275800dd72230f94d0382cbf9d6d293e189b2e288ac62bf
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31972f8c5588889908297685ea316a6c5603a5c6d49476ca372f9024aae59d33
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d5a75fa06d8e0567428171895f2f5be669b2393bf43f572fee60b1f39d5db02
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d03e6c49a145ce2e51d87480bb76687cc7b5c7b56e94d2991de3e2ab67bd210
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4cadac78638c6cb012c738c824827d4d4609dcfa94ea2fed086f7884c5c610aa
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c724d2349acf4c39c5d4d83bc11668bbfaa5c5543e4339570ed104ddf6da9c9e
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.066666666666667,
5
  "eval_steps": 500,
6
- "global_step": 31000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2193,6 +2193,41 @@
2193
  "learning_rate": 6.2271111111111115e-06,
2194
  "loss": 0.046,
2195
  "step": 31000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2196
  }
2197
  ],
2198
  "logging_steps": 100,
@@ -2212,7 +2247,7 @@
2212
  "attributes": {}
2213
  }
2214
  },
2215
- "total_flos": 7.551077842944e+16,
2216
  "train_batch_size": 4,
2217
  "trial_name": null,
2218
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.1,
5
  "eval_steps": 500,
6
+ "global_step": 31500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2193
  "learning_rate": 6.2271111111111115e-06,
2194
  "loss": 0.046,
2195
  "step": 31000
2196
+ },
2197
+ {
2198
+ "epoch": 2.0733333333333333,
2199
+ "grad_norm": 0.2015141099691391,
2200
+ "learning_rate": 6.182666666666667e-06,
2201
+ "loss": 0.0552,
2202
+ "step": 31100
2203
+ },
2204
+ {
2205
+ "epoch": 2.08,
2206
+ "grad_norm": 0.13618551194667816,
2207
+ "learning_rate": 6.138222222222223e-06,
2208
+ "loss": 0.0504,
2209
+ "step": 31200
2210
+ },
2211
+ {
2212
+ "epoch": 2.086666666666667,
2213
+ "grad_norm": 0.11296480149030685,
2214
+ "learning_rate": 6.093777777777779e-06,
2215
+ "loss": 0.0527,
2216
+ "step": 31300
2217
+ },
2218
+ {
2219
+ "epoch": 2.0933333333333333,
2220
+ "grad_norm": 0.16036346554756165,
2221
+ "learning_rate": 6.049333333333334e-06,
2222
+ "loss": 0.0513,
2223
+ "step": 31400
2224
+ },
2225
+ {
2226
+ "epoch": 2.1,
2227
+ "grad_norm": 0.17830543220043182,
2228
+ "learning_rate": 6.004888888888889e-06,
2229
+ "loss": 0.053,
2230
+ "step": 31500
2231
  }
2232
  ],
2233
  "logging_steps": 100,
 
2247
  "attributes": {}
2248
  }
2249
  },
2250
+ "total_flos": 7.672869421056e+16,
2251
  "train_batch_size": 4,
2252
  "trial_name": null,
2253
  "trial_params": null