rakhman-llm commited on
Commit
5158233
·
verified ·
1 Parent(s): 3538c60

Training in progress, step 32000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b61d5a5a0f5b7ace55f8bc3cfcc9327599d34056cf4715c1510e9c33291ed41
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9a244e9cb55adf7cda56ccb82e1a9bee92d342f231471fb5502ab29003b5e6b
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:31972f8c5588889908297685ea316a6c5603a5c6d49476ca372f9024aae59d33
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2fb02fb3109cf50329151281ddbb10cbbd67d7bcacbec12c415183618e7c8c4
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d03e6c49a145ce2e51d87480bb76687cc7b5c7b56e94d2991de3e2ab67bd210
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba3afeb76c9729f93ec4d0e7011fb72d06e1cea76b77b26bfd4a5e76df22f8c2
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c724d2349acf4c39c5d4d83bc11668bbfaa5c5543e4339570ed104ddf6da9c9e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89d99968826ce47bd88d7b05d470f4fb17f0bcc9ac13362244060de99696b592
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.1,
5
  "eval_steps": 500,
6
- "global_step": 31500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2228,6 +2228,41 @@
2228
  "learning_rate": 6.004888888888889e-06,
2229
  "loss": 0.053,
2230
  "step": 31500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2231
  }
2232
  ],
2233
  "logging_steps": 100,
@@ -2247,7 +2282,7 @@
2247
  "attributes": {}
2248
  }
2249
  },
2250
- "total_flos": 7.672869421056e+16,
2251
  "train_batch_size": 4,
2252
  "trial_name": null,
2253
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.1333333333333333,
5
  "eval_steps": 500,
6
+ "global_step": 32000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2228
  "learning_rate": 6.004888888888889e-06,
2229
  "loss": 0.053,
2230
  "step": 31500
2231
+ },
2232
+ {
2233
+ "epoch": 2.1066666666666665,
2234
+ "grad_norm": 0.18772698938846588,
2235
+ "learning_rate": 5.960444444444445e-06,
2236
+ "loss": 0.0483,
2237
+ "step": 31600
2238
+ },
2239
+ {
2240
+ "epoch": 2.1133333333333333,
2241
+ "grad_norm": 0.18564331531524658,
2242
+ "learning_rate": 5.916000000000001e-06,
2243
+ "loss": 0.0493,
2244
+ "step": 31700
2245
+ },
2246
+ {
2247
+ "epoch": 2.12,
2248
+ "grad_norm": 0.14556245505809784,
2249
+ "learning_rate": 5.871555555555556e-06,
2250
+ "loss": 0.049,
2251
+ "step": 31800
2252
+ },
2253
+ {
2254
+ "epoch": 2.1266666666666665,
2255
+ "grad_norm": 0.14260981976985931,
2256
+ "learning_rate": 5.827111111111111e-06,
2257
+ "loss": 0.0509,
2258
+ "step": 31900
2259
+ },
2260
+ {
2261
+ "epoch": 2.1333333333333333,
2262
+ "grad_norm": 0.08872347325086594,
2263
+ "learning_rate": 5.782666666666667e-06,
2264
+ "loss": 0.0466,
2265
+ "step": 32000
2266
  }
2267
  ],
2268
  "logging_steps": 100,
 
2282
  "attributes": {}
2283
  }
2284
  },
2285
+ "total_flos": 7.794660999168e+16,
2286
  "train_batch_size": 4,
2287
  "trial_name": null,
2288
  "trial_params": null